var/home/core/zuul-output/0000755000175000017500000000000015145063214014526 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015145074240015473 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000321025015145074063020256 0ustar corecore3xikubelet.log_o[;r)Br'o-n(!9t%Cs7}g/غIs,r.k9GfB ?KYI_翪|mvşo#oVݏKf+ovpZjC4%_̿f\ϘקjzuQ6/㴻|]=ry+/vWŊ7 .=*EbqZnx.h{nۯSa ׋D*%(Ϗ_϶ݬvGR)$DD D~m{]iX\|U. $ॄKЗ/83Jp ώI8&xėv=E|;FmZl8T*v (6pk**+ Le*gUWi [ӊg*XCF*A(-aD~JwFPO7M$n6iXύO^%26lDt#3{f!f6;WR.!$5 J:1*S%V!F([EbD]娍ԹiE03`Cfw&:ɴ@=yN{f}\{+>2^G) *ɚL}ӄ]C }I4Vv@%٘e#dc0Fn 촂iHSr`岮X7̝4?qKf, # qe䧤 ss]QzH.ad!rJBi`V +|i}}THW{y|*/BP3m3A- ZPmN^iL[NrrݝE)~QGGAj^3}wy/{47[q)&c(޸0"$5ڪҾη*t:%?vEmO5tqÜ3Cyu '~qlN?}|nLFR6f8yWxYd ;K44|CK4UQviYDZh$#*)e\W$IAT;s0Gp}=9ڠedۜ+EaH#QtDV:?7#w4r_۾8ZJ%PgS!][5ߜQZ݇~- MR9z_Z;57xh|_/CWuU%v[_((G yMi@'3Pmz8~Y >hl%}Р`sMC77Aztԝp ,}Nptt%q6& ND lM;ָPZGa(X(2*91n,50/mx'})')SĔv}S%xhRe)a@r AF' ]J)ӨbqMWNjʵ2PK-guZZg !M)a(!H/?R?Q~}% ;]/ľv%T&hoP~(*טj=dߛ_SRzSa™:']*}EXɧM<@:jʨΨrPE%NT&1H>g":ͨ ҄v`tYoTq&OzcP_k(PJ'ήYXFgGہwħkIM*򸆔l=q VJީ#b8&RgX2qBMoN w1ђZGd m 2P/Ɛ!" aGd;0RZ+ 9O5KiPc7CDG.b~?|ђP? -8%JNIt"`HP!]ZrͰ4j8!*(jPcǷ!)'xmv>!0[r_G{j 6JYǹ>zs;tc.mctie:x&"bR4S uV8/0%X8Ua0NET݃jYAT` &AD]Ax95mvXYs"(A+/_+*{b }@UP*5ì"M|܊W7|}N{mL=d]' =MS2[3(/hoj$=Zm Mlh>P>Qwf8*c4˥Ęk(+,«.c%_~&^%80=1Jgͤ39(&ʤdH0Ζ@.!)CGt?}=ˢ>f>\bN<Ⱦtë{{b2hKNh`0=/9Gɺɔ+'Х[)9^iX,N&+1Id0ֶ|}!oѶvhu|8Qz:^S-7;k>U~H><~5i ˿7^0*]h,*aklVIKS7d'qAWEݰLkS :}%J6TIsbFʶ褢sFUC)(k-C"TQ[;4j39_WiZSس:$3w}o$[4x:bl=pd9YfAMpIrv̡}XI{B%ZԎuHvhd`Η|ʣ)-iaE';_j{(8xPA*1bv^JLj&DY3#-1*I+g8a@(*%kX{ Z;#es=oi_)qb㼃{buU?zT u]68 QeC Hl @R SFZuU&uRz[2(A1ZK(O5dc}QQufCdX($0j(HX_$GZaPo|P5q @3ǟ6 mR!c/24مQNֆ^n,hU֝cfT :):[gCa?\&IpW$8!+Uph*/ o/{")qq҈78݇hA sTB*F$6 2C` |ɧJ~iM cO;m#NV?d?TCg5otޔC1s`u.EkB6ga׬9J2&vV,./ӐoQJ*Dw*^sCeyWtɖ9F.[-cʚmD (QMW`zP~n"U'8%kEq*Lr;TY *BCCpJhxUpܺDoGdlaQ&8#v| (~~yZ-VW"T- 0@4ޙ-did˥]5]5᪩QJlyIPEQZȰ<'$VO"dwEр%}5zW]S)ZbF p$^(2JцQImuzhpyXڈ2ͤh}1ieQ*-=hiך_%הrΒ7 _:+$ߗv{wzM$VbήdsOw<}#b[E7imH'Y`;5{$ь'gISzp; AQvDIyHc<槔w w?38v?Lsb s "NDr3\{J K纂P//Q۲6+ktwLzG,87^ 9H\yqū1)\(v8pHA"ΈGVp"c ?Z)hm.2;sl$瓴ӘIe~H|.Y#C^SJĽHǀeTwvy"v܅ ]?22R.lQPa ˆSܛ1z.Tvg.0Mn}#~sz1}z}Vܟ? 8CWA˫[LznbWoѷ 'hG iׄ?:JmE1y<6"`3T7h|ޒe|if]ޒywUnj|_$cݢaȤ금~{죤z[=_d rtc\MU&IzhBYň(&&WV#@..$ 91XMBQϷmQm{mO!_ysW1!n<0!2~.볿j HmE]j `7ruuŨԀ![Z !iHlf[7Ua6BEZd9 NpydqZrS6U A~@Ve Ȇ*dXsl o:d lzzMvYź ^ ٲAPm쪊m\9htwmjQ\c5&,|^C.SS P󂏛o n8Fkb^s/&a[s~W &ɿ^\r\ߺn`Q6eMӁKzFZf;5IV1i[xS 0FPM]gl}>6sUD_忤(-ʏw\ CLJjh7hM=~ ֿ.X;|iDr'# )v3v֠:b(v6& `-K;~(fwnv :9?TlСE,m]XaINE`!6uOhUut/wY5H0Yhu ;ZeY}Qg?lvחzäTC 4ząCLJk7߯@qC cN o~1-b }kAn=)m 3fo˶_ XJ^jm3Kx6BDhvzZn8hSlz}mBЦxDƨ||Yt,zF u _5_WdC<mWu[7-D[9)/*˸P\C[n膗:68DkM\7"Ǻzfbx[ۮC=1ÓOv$sY6eX%]Y{⦁# &SlM'iMJ았 t% ~@ .m}x˫P}@.供A[?~R6*.9t,綨  6DFe^u; +֡X< paan}7ftJ^%0\?m'5k][iWᚡilRї`m] leu]+?T4v\% ;oqF0qR(]Wᚡmi@ zF(n&P;pS俺kCߤ`ZnTV*P,rq<-mOK[[ߢm۽ȑt^, tJbظ&Pg%㢒\QS܁vx]U5g B(, qA9r;$IN&CM(F+ hGI~Q<웰[, qnriY]3_P${,<\V}7T '6Zapto}PhS/b&X0$Ba{a`W%ATyvo\T|A_mnZ7sDbyT'77J6:ѩ> EKud^5+mn(fnc.^xt4gD638L"!}LpInTeD_1,֒J2)p|ݛwu0{֩2ْM4tޖӳM\Qe%*?vQ~W  yr3-2+=Щp!k2 ûQÝBsN&4wx,k%*nD4qL~`|%4Q0q["< HK'f dt(d/ZoQ%_}~Yki7}SWekk̗E\e'Nh􇋲rTG_77:0@Iuʙ?&Ԕ8e,žDG"1lͧQѶGM]}yxZl 0JM"dλ=`Yƚ^"gJT>u8.H#',c@V8 iRX &4ڻ8zݽ.7jhtsϑ`94t1!F PI;gi`.W&&ף \98 y{͝Bx#)u62ݽ[X-w &ld)r;#Q'  yH QveJ=WhwS]֫l"]Јzg6eze;\Mdv!E]?CLC4ʍ@1Ssc;l?ߨG~oB(ъ{zgZJ }z&OF wkߓG9!1u8eBe%5A%!s#JXBz-Bȃ82,߫ ~c a^ 5%Di&hWZ n193T9Щp-NC֤^pY鳡Śk 2` PfjJc%e0nx MSkƊզ OT_jX3&:e}@t;;XAsꍷ=fO6LIR[T[Qp-&u8~a+3w]_,;?.W(91ݧ$uxp/Cq6Un9%ZxðvGL qG $ X:u06 E=oWlzN7st˪C:?*|k.TD.Z[mrlTPzQ[$vJKL .FZ*(1'$7^mSNxC8Ña!^{4uE?ܰ8_r 6.sϸRw# y'uR;$厛p!&*̽dPt/ ̿f3Nd0 ~n~YͤBoK&9<{̻*RmသLΕbDOJx߭&;~+WrVXyԋ4ԙ./_A9B_-Z\PM `iĸ&Tt (6{\٢K 5XGU/m >6JXa5FA@ q}4BooRe&#c5t'B6Ni/~?aX9QR5'%9hb,dsPn2Y??N M<0YaXJ)?ѧ| ;&kEYhjo?BOy)O˧?GϧmI C6HJ{jc kkA ~u?u7<?gd iAe1YB siҷ,vm}S|z(N%Wг5=08`S*՟݃*־%NǸ*kb05 V8[l?W]^@G:{N-i bɵFWǙ*+Ss*iނLޕ6ql?N/e1N2iDEu&ݛȘPˬ-Ő\B`xr`"F'Iٺ*DnA)yzr^!3Ír!S$,.:+d̋BʺJ#SX*8ҁW7~>oOFe-<uJQ|FZEP__gi(`0/ƍcv7go2G$ N%v$^^&Q 4AMbvvɀ1J{ڔhэK'9*W )IYO;E4z⛢79"hK{BFEmBAΛ3>IO j u߿d{=t-n3Pnef9[}=%G*9sX,¬xS&9'E&"/"ncx}"mV5tŘ:wcZ К G)]$mbXE ^ǽ8%>,0FЕ 6vAVKVCjrD25#Lrv?33Iam:xy`|Q'eű^\ơ' .gygSAixپ im41;P^azl5|JE2z=.wcMԧ ax& =`|#HQ*lS<.U׻`>ajϿ '!9MHK:9#s,jV剤C:LIeHJ"M8P,$N;a-zݸJWc :.<sR6 լ$gu4M*B(A ݖΑِ %H;S*ڳJt>$M!^*n3qESfU, Iĭb#UFJPvBgZvn aE5}~2E|=D' ܇q>8[¿yp/9Om/5|k \6xH.Z'OeCD@cq:Y~<1LٖY9# xe8g IKTQ:+Xg:*}.<M{ZH[^>m0G{ ̷hiOO|9Y"mma[sSbb'Rv&{@6; KE.a\}:<]Oyve3h9}E[kMD,5 %sO{킒 8.K?]i/`׎tp NvԻV4|<{H@#*h{Yp/E%dlh\bU:E%h@&SEK [ Ƣ xg{z%ǻViX~鮦w35QE~qp[ʕ@}ZL! Z0!A⼏q)[f &E1K3i+`JG P/EG 4 9LڑKL|`PОnG#|}qOR{Q|2_tH߫%pD?1%(@nfxOrs25rMլf{sk7݇fjӞh2HkeL'Wʿ}Ƞ%>9cSH|cEyQp 'ˢd:,v-us"Iidw>%zM@9IqrGq:&_p3õB!>9'0LL]M[lwWVR9I5YpVgtuZfG{RoZr3ٮr;wW:͋nqCRu1y=㊻Ij z[|W%q0 CJV٨3,ib{eH7 mҝ(3ɏO/̗-=OR\dIoHZ6n`R֑&#.Mv0vԬ]I˟vrK}F9X|FI#g.Gi)%!iK|o}|ֵ7!ېATJKB2Z/"BfB(gdj۸=}'),-iX'|M2roK\e5Pt:*qSH PgƉU'VKξ ,!3`˞tӱ&Jy%١oBbFM=$OQYꐙ^=Zza5a%פG,ϒPV3^KPbGVO'daOU%tt!ƖRG9lhfd#]y=DFT8F}$RD<8 ].v\-v:8F+Mt|ga.!! р#ݴtӫߴ]vWͽ2]Q6Û͘`_}KnK"]p<)Xg '鸽= &Xu=y`g[#ɯO"?5Vg3gR(Җ}f`ӀSqUق0D L?U7_nMBLϸY&0Ro6Qžl+nݷ" 㬙g|ӱFB@qNx^eCSW3\ZSA !c/!b"'9k I S2=bgj쯏W?=`}H0--VV#YmKW^[?R$+ +cU )?wW@!j-gw2ŝl1!iaI%~`{Tռl>~,?5D K\gd(ZH8@x~5w.4\h(`dc)}1Kqi4~'p!;_V>&M!s}FDͳ֧0O*Vr/tdQu!4YhdqT nXeb|Ivż7>! &ĊL:}3*8&6f5 %>~R݄}WgѨ@OĹCtWai4AY!XH _pw騋[b[%/d>. !Df~;)(Oy )r#.<]]i-*ػ-f24qlT1  jL>1qY|\䛧\|r>Ch}Ϊ=jnk?p ^C8"M#Eޑ-5@f,|Ά(Շ*(XCK*"pXR[كrq IH!6=Ocnи%G"|ڔ^kПy׏<:n:!d#[7>^.hd/}ӾP'k2MؤYy/{!ca /^wT j˚ب|MLE7Ee/I lu//j8MoGqdDt^_Y\-8!ד|$@D.ݮl`p48io^.š{_f>O)J=iwwӑ؇n-i3,1׿5'odۆ3(h>1UW蚍R$Wv8FHӜ"D$aǽO8'1lfYuB!aMHVuV_K2k*`cKxuBG&24T}Lai 0Va(7K#ӊ!,ZDxFQO*lם>!4ӥ2 ]8â6 U`V%`!c%؎ʨTzrKh! c.}.D>)d_ 8rcu,wf2?Ǡ*_lDn}rauyFp*ɨ:UiM2r:9ct X1lmĪ o玓,R%!`hGT LYF#g<cm${|Xdu4tmtїUJ\~dc0KcMlf2?mμQ ߉J4WrSHTdp"ӹ'cJq2zPlX̯.0H!ND@UapVoGڧD5>H]f@!=߸2V%Z 0"G4ȇʩ@]>Y$ًF_Mm_Tt)ib+q&EXFu򾬳ǝ/RS>r,C2NfOjpcm{Ll9vQOT>9U;])>6JdbXԠ `Z#_+D[7IIjJɟUh ҙ"`"a ߒ"G̾H`6yiCk(OA/$ ^%K^+(Vr[RR1"u4A.1X0=7f/"(o9/L1X{]q`Ȝ/; 9a>E)XOS K9mUxBa"'4T[Jl /K/9,rlCAj_TiǘP,:4F%_0E5IE'rX-|_W8ʐ/=ӹjhO%>| :S Px„*3_y.g9| ;b`w NtZtc> ײ1KĴ{3Gl& KT1ZWX8?C]~We$9; -.D087?1a@P5B,c}jcGȱ WW/ @a#LA4.ٹ^XڋXٝ:^Izq. ٽƎDn6ٹBc5Lt;3#i3RAٽ9| cbpcTfp> 6L/_x 'ۙz7~w~);qU9GDT! 6]c_:VlnEUdn6UˇKU;V`JUݵޙEO[)ܶCy*8¢/[cչjx&? ՃJȚ9!j[~[' "ssTV2i sLq>z@JM->=@NỲ\쀜*/) rdGP8_S( $?uk| ]bP\vۗ晋cgLz2r~MMp!?~h?ljUc>rw}xxݸǻ*Wu{}M?\GSߋ2ꮺ5w"7U0)lۨB0ח*zW߬V}Z۫ܨJ<]B=\>V7¯8nq~q?A-?T_qOq?5-3 |q|w.dަ?ߐ"U;gm>?Z֒Z6`!2XY]-Zcp˿˘ɲ}MV<в~!?YXV+lx)RRfb-I7p)3XɯEr^,bfbKJ'@hX><[@ ,&,]$*բk-Yv5 '1T9!(*t 0'b@񲱥-kc6VnR0h& 0Z|ђ8 CGV[4xIIWN?Yt>lf@ Vi`D~ڇŁQLLkY <ZPKoma_u` !>Z;3F\dEB n+0Z ?&s{ 6(E|<ޭLk1Yn(F!%sx]>CTl9"و5 |ݹր|/#.w0ޒx"khD?O`-9C| &8֨O8VH5uH)28 Ǿ-R9~ +#e;U6]aD6Xzqd5y n';)VKL]O@b OIAG Lmc 2;\d˽$Mu>WmCEQuabAJ;`uy-u.M>9VsWٔo RS`S#m8k;(WAXq 8@+S@+' 8U˜z+ZU;=eTtX->9U-q .AV/|\ǔ%&$]1YINJ2]:a0OWvI.O6xMY0/M$ *s5x{gsəL3{$)ՆbG(}1wt!wVf;I&Xi43غgR 6 ݩJ$)}Ta@ nS*X#r#v6*;WJ-_@q.+?DK១btMp1 1Gȩ f,M`,Lr6E} m"8_SK$_#O;V 7=xLOu-ȹ2NKLjp*: 'SasyrFrcC0 ѱ LKV:U} -:U8t[=EAV$=i[mhm"roe5jqf$i>;V0eOޞ4ccc2J1TN.7q;"sդSP) 0v3-)-ٕAg"pZ: "ka+n!e߮lɹL V3Os\ဝ+A= 2䣔AzG\ ` \vc"Kj61O Px"3Pc /' PW*3GX liWv-6W&)cX |]O;C%8@*Z1%8Gk@5^NtY"Fbi8D'+_1&1 7U^k6v읨gQ`LRx+I&s5Www` q:cdʰ H`X;"}B=-/M~C>''1R[sdJm RD3Q{)bJatdq>*Ct/GǍ-`2:u)"\**dPdvc& HwMlF@a5`+F>ΰ-q>0*s%Q)L>$ćYV\dsEGز/:ٕycZtO 2ze31cDB/eWy!A/V4cbpWaPBIpqS<(lȣ'3K?e Z?ڠ8VSZM}pnqL f2D?mzq*a[~;DY〩b𻾋-]f8dBմVs6傊zF"daeY(R+q%sor|.v\sfa:TX%;3Xl= \k>kqBbB;t@/Cԍ)Ga[ r=nl-w/38ѮI*/=2!j\FW+[3=`BZWX Zd>t*Uǖ\*Fu6Y3[yBPj|LcwaIuR;uݷ㺾|47ߍeys=.EinE% 1zY\+͕߬VͭW_겼cazyU1wOw)Ǽn@6 |lk'Z|VZpsqL5 څB}>u)^v~,󿴝} 3+m𢛲Pz_Sp2auQAP*tLnIXA6L7 8UgKdT)*7>p{Pgi-b)>U6IXabPde Ӽ8Ģ8GɄnb'G ֤Mcv4?>HC78NE@UMc8>`TvZwm>xTHIԙ\v3۴3EET=8~%ٱӤt uv['7Hb IX:N%Hcqw=Q"CUTK@UH0TbP$[.|P V0~qi;ȋTUiCS1TE)U"*OZ/HaQbZe9osIr< <$*LM|` ^x{Q U)QHw`5˂z܄m>B*31'dдNe2j2TuWLϣA O=¹ BMK2<'r:H'cq^/q)9.q*p_q b(ɮ*xp!%_KNh#E쀄L_`>%!{| }^\yGWyq(q#7AOWG9yq%At`mJHБl ZH0I7J֨ =J.q{ЬXyVTfA^!=U8΋~_ bDY'`˨gəFLSֿ{K$<,wRg¶ξFaȦợxdZվZVUoJZBE|2Sz6.ƎM|5QLE5hA,(ec${o:T.e;@Ƿb Dc=֛7#32Ǧ)DSkP?8[lqP lxIsx E&*ѠLaʾ'D]5XrGՄE4W2TRԈ%,N{T"$IXƍ^碚&*b\f2 K퀡5,㢂P[yG>!9\:2u Rҵ%~Ox*BΠ̒l*EVy G]ix2RRo EHP$i6 y#* u|^>knyDFf*@yㅲT*,EJ#r`JTJg\bUyV' >52]bTըK.ha\u-c-惀ev>f7J}`iCӕM"[h]=D2KHlE>W:* 7l|,'m4-+$g^"T~p?ߗ8S`7>6߽la@s\2quz^qQ"E€Un"S*ͫ6D"Íg߽JQkq `9A7֓J xhտ_>$M ;TɼN.Y!@8ޝvw[ cou_뮣'Y']ߴ_Ev"T(8} x|ŭAytkhEP\xWGzk~p!:Uu,:>DR"E@XM]|DvC/| L[s|*M%5GB|3(<(M`pABXiŲ>>EN*LtQw2)` X"f(e_colk"3߀UV U&n85lviڗNDQ8>‘(҂o>eeˁCM$L/t53QjiTGuvw5Xj&cd+Lò KX?!0ygݹ(fJٗ>1 ˬyܜrEr.r qST%;<yݒoOBi5, ;!~c8@T,TF'p(k<'W5Ea^[aTZU8d-pn=3E[^sy{|!nTɖHKOm'n6x*ZW[ ذj;AjZf04}}%;E ~xH9Nd}&QVJ!nO߾L6.x@K/ؕvIYkZXe+: ׉HqwLp](*DZ{Ύ0XbG5٪0ȫ e&uylJM+mų4+Z*'itΑyaf0j.$wX^ҲE7k,0)[e+JVTOJOM悛]`X5s)ln&뼱V)|@/rg5gsüg 8>."5$ޢk#]]|bQ5%Ck bj]ׇ: upOq_ /Itm Loܕj_eswG\}˾@mu)Vb76) 1؁DsZfrs:\quketN1\o8Dhf6Wl|p@fz_O߆"X/#̣n2^wJT{;RiKKUWnဘJV@3"CT'W0ʱCFdksU usG\|V98'!Ra]q|_pEY^g UX$iIK}H/5{*n#̴9rp]^Ҿ 9cXӶD3iF2=-op+Ghū}\ v0!y}Lm{BsP.gMk^ఙwae7y*)Se*b*pggj0"C-W֕E_Yn{?Q<?+c^Te钣uZEh ~&)7Zk<\d:3f*YgͭUoW&-ͰlvwBig21?HqHX@av`uP8ޏz6 \swO 0vf'տa||~9, +hm_"+NoszҀێ{`BVwm>ba._Mpr2}zp-̹x?6 }1%'>1?h ԠJsL;EN?K9A&wN|0Zsc%Nmr Mj8G9EDM}^)u9ޜ"<g%3w[691s~gW_N99"~_9nX_4b췛&@ADYVZ6m;XmzL|v=s|ډ;z/L"(zi?t8c#Fɓ?8޵6n$ H&g&Fc=N lHAUuT˖M$MWUuUMWgAZR>h2v .&_> ѯPh\ܮ|^~ӉF{.$Wn ojݤ܁pwo=|`pnO>|9Hn"=ib)MO/+!6f|,ʪ%Wɖ`(@[&޾Hq]U>.MUV(:f՟u|I?toals׵~a?fNѪ'L=J ._!~5%`*{ ' eU.SlQ*y `U_~Dq9,)*L/ĪFTN"Y7ﰾIt˙pzSZ= ~G`L /dPMva $ue/彲/'`\"9(:~\^E "rrg9)N;ݰ<#FCb`h8xa~-=5BRV Kc hUC #VO'!?]0۴kGJtGXZ2"/g`E^Gt+804y{st#C'` B<݌|Y ΋3@avH¥mw:x] 0Ӎw# $h] ] Pw00iߍc?sڍI^L~):g:XDǽ\WWB;xD.~NᜈDtťont# Bq_9Ŝ]KNs`}~Iϳb 䍿\z>1ŜKbN!W``jݍ p %$!jnwIҠ6m؁A@ˉ&Z[v Y[˘qvE 0ѦP`LEeK|]a ([P%;bs$ޫ7r80N T=ս5p)n\P+ ; A} 4 :,覨nۂvaߑx v3ȩe@HR7F:c`fj1px"kw, & 7̷;@  ?Mvv!lH&OR^s?6S6y ku L`u(˽ 08}mB :+lN{HVaʵ @n ^LHpRhؘ >-e<`zp?W] _}"X_5_~BD8~Q;}FC GԼ~O&Wݽ%WS?i hGcDgG䆑B<;`c LG}\K(H":G:0=N36S qU\JJ߳rNr"`zC<8:!et:V&Qԋm7q7{C'r{0vS[x Io7IěCyxl'@YPy@]ֱ< s!ǚ#uGtB =\  $6A} P_yR2ɦJ'QIH?NV[Bj-!n'f@{ [{fj}x(?05|QCl)x}74Pp%ӼyyO6|IyžpU^# }vf8 o)l?h(O/4ɋpl!b\=GRfw蘂 hTvN/]bP/CqgrJP'n^%|u|h:o]hg 7W-F%Wp(èJOD ʡapH;(|k[Uo`|Rף~Ϝ@Cq#/Ncr.7C7.;NQzNlLT:@Yϥ"s\ XЋX/ cG Dh D}\zgix\(fD&B()*'kpo/E}mQ/> N@ %/+mg^px$0\=(g#Qβ3u}7?*xiiʸ6'+`u6t3nM#Q8mL 6*74MR v |8djH8ͧ`%1j ^z}?I-E0졓,t`}cnYEō G^7A(Dhzl._օ٦{-TVc}Zm! Av4ʹi誓OנK&'p-ڎCzC-NB{ [~-D"b5(`PwGB-V ۞Pyz;mAJ#ߑP BUB- '4xk]%9QBWZ:oD^BՇG lye«lyh"z0Lde^=am6s_x7w )7FӷHz#o`ʳo vkh2@xQ6}m/h|7iZu3ޟ9ẙLɇl9ɀECEAz<9G=8_>1U:7hvgcy+"m "Y,T6 {V\"e"C2Aiq=M(LE`Wo1 !H9Kc`xz+K3x̾n6a/MJl JlBxN0Ce j44Y|+ƪ$Iej,r O j( PL[)|~rr|{~/+Af9" Ujv[ZV?DcM)u [;,z᫇>Pv90O-t-0UWY*/,˒Ň+n^@~c"Nq*W<5Ynpv)@:E4 xz44-O6<2|͒^c(,/~Gef:, ^Hu4D$់HzטabF4g9@X,Q6Q~B&b`^.G`ĮV(|ycfٮGe[#۳]^9 4ndu}<]rsqzWZC>mGeMs_ >^Ӽa Ny3s{P\uߖ1 KV-1k)W=+.^ ﰾ~q%3͓(Dƒꗥ jBpOj){ \:([|cZ Q8BnSi9Ҫ&vm{(Fء~YejH !]O0|}R$X*"fcDy9IR@[PQiFi\遀ʸp$FE1mfD3haq[Ģ' 5dc9^e$-.^ҌOΨ6[jlm@:И)@@TzUc7D 4I=!"{R2\J@u`'V~ )=0^ p[QLƠ?7JJ9s9rf|kT)Os͠Gn)9188?WɳhFSe9n"R]" B>ܰ=XuNFPF6uZLd'J[l!t,!켣%ou9irTLxd;gwZеSR7g\Og؅nbkVʊ#?d՘ir.C%(Ll1 ^qv*N< ؿajd/qD/),),:d*ύ4~K])I>4-6Zwg'fq lG]NuJVTGb15^ /glE/*@CLs/J/nY&l.]޶%pqET\ETWOLyZkP%4ELz.{mVwYkHn~u׍EE F`Q[TtԶrRCk!O\<1gE:kr^vq0ʒbm2Op*%n=@W h(%&IG>c}nx&gQ串 i>'ik1Wapj$f/CBԘEZ--ڣ&jVhal5ed1Ծ,*UmMsdֵJ{%b6OeYtƤ5Gm ZqZ [瓒,= Y iH$6{+X۝xUF5(#uPɐA5Gw&,!p}l {jk,[^IgO+~5gN  3>{=mR ^}M;wΗ &?o3uuW,`$?_Ac??gFSv_~)^jϥS?e1.^>o]4lv\.e7W㿆 pP.X^_V.2I@}۟Vw3mi2-uŵ 7FIxG.3Bfm41pMW?(+.u:' Fޚ\ZlIro_xpuaF^+?LQnP-TAhsMn6*/i_]VpPJ TBRXS}}:17PD;o;y v$nR- s;xX> ps)♃f΃I3cnt)H"0ؕ KAm0FkE?N9/[x'Hj>F^8+b\cnk=" +[Dٿ(?GW1"Q .p3pW4%E4ŶOiOBSy~W^'Z`?8X/0 NP4;|Pt8RPIh"~ݥ8b(Tsz! `FXTE,&j %㲆MZz w:o'Lw/¹9`'Fϩn#݊v6cc;XYqE)BlInlmC륕7H{>,Ş6?>?8H] J дJz!}_ Ρ!׭F?" 8Jotp*?shX$g[:1Ɣ` qY "AxBÈ!ARQҜU;כMj:J _8j=6ZʗKz-F5/ o%GsY( 1mnMXF{Ƙ!ʜuAh*CEy.EHӻ8>89Yւȅd"3|Jb &n>8|JJxxX,).ht天$$H׈md e [GʠVsrлK4 ).p?Q͡R5YZ58$^?$T%y9&j2Q$q7g54624v>7]ϦMkZZAq9"U;6/L /E̎$%VƱF%C;Vd|lH%>E>Ӛ=Ä[i4a3d&?r|pD`\*ћaZ3m MMI~1 mzgÝ\܊k`:3>49k[IGgrc`)>r>r׿k };PD!6{_`+sj&5sd{=M*Z&tss(Ά)eGG=T~jOGBO{6o_z(QZr}ÑTr}cj{Dġ#p]:^_!t+,*4RȍbS$ z;ؔ0[RVv'ϘgOׂ$rg| &zdæ d Bӗ3aY5R^'D`K|*˛ ^{`G0Ϋh_y_-@3 VHBasDbxvLMI]{0DwJjo%yؗ{`Yaz9&] r8]P|0$D H_u=#cz25*ź**#6&pIa ~Q'N >eK=j]g(NN3`rjG$u$?OʦG5G"U5 <=h۩er%1DwL_MY9W}#W=x˟GgΊ!; ~1wVRZ`/Fw5W 1u}c38ßs!)yMw/t\iܽ4?O ec%ӋIG8noK=p5V6S<^Rčr1Vi0 KLĠ`RfADGLFL1Q$Ax e\Y8@z)ɻO&{-CLFwfpӌ Rz:b_szW_Ȝ/UgƲkH,K*õyꂣ$Wj;|9JF\cHFÑ"Sw\= ,7U(OBWO"x*@cHQ+OndMuC;6TE}\ w  N ƶ*ELgFayqS;lSh-4<j 漖?%I˟$CXf߭՜w=@M2>],N}?xeZɂX%HR0 \I D=uAQ댖-R}2phJQ^SxIµ|-0ŜJ>H ?J-_{' .rJ9k{Pm0L⢚U(\T~qqmwï]\<%02r#+ +Uickԗͥ K#ڽ ?~| "N/D|^o.m\5:.'<N򨒯%c$&sq%b&^bX޵wwҧnU_) _WyٲᦧgbXů$G#.8:uVC?2x\ju۩Q}YKb16%y!E!qtG1(s-/ d) Q)<b"WG^/E԰3򬰛m`*Kƨ߰ ޟՄ,A;=˹ݟsb)D Z j8t$vP$y{G?2}t?lllD&<.k4^8<Ee fSkS<'jE1וs SNκ( 9Ey(zv)4X̴ PAmCkd.> щ\z]D n D緸Vjf`]D}F#|;V銏j' ֤MIG>?w;<vt~fu:n-' 6u³>eQo88'xj,O|l`)r@3r3Yȁa۞-Tk֪&#S$\xs:g_ „yAv|pC|s>__ms%J ZQ)Z<@׼bER[N-dIGl .̙{QaP Q3m8?BlY:5qPvZ-KOӯp7 pO2FjV *FdK\mDtpש*}nc%=zA)Gξ?vqjS{zgޣMv>uqpboX>e,@1[*9x f n52p*{۶ !n%!(8 703Ø^W{-J#ԃ]IlJ399s&s/;k6]3O><9f}Yr '2K_WaO9>vqJpx;(LpK qg t N%fLu55FeW>QGݭΕ\C[eLOcީBJyBMmۃ5? ZNa\-شu:p:իrrvU|HȁJӆk4d];3Gqv-M,>2f42s_ŷőo=jz:wRjV{QlvyVrNfS!f:H&Qh}; /J;=uesЎ"Kщ6A"OB^ld_q<?v;\!WyU4j¦#?33|aX">Ol* .]aeh`ob"'"w-)~b/W9*oj#X{u|q) +@'h |Q2?{6l9f.$\7Yӏύ^o7##Ejq-/f4Cx` %oEӠF|ޛL'\?UbJy'FYJSrar~5IOAs~.> N1f)v'.GA+Z8ڠ%-Ȣŗ{|d (7`h_FJ}a!Ck\eK9ѧtq<-F[˕@隨@\Mf~y{ ut6 3 ^ͧ/S5޿1zЧc3Ete>3 O/ ~Nܛ0|Og#\AU6'}9 Q̯Pc?>9F]u4[D?/bdlY6<; a!8ktNUXʡ`YPƃrцq|(s`C˹A9 #r-,j(:`"{^^f5`.o+bcML(\p|:/dhOfaj|VV5]V2; ~( \4bYQ*O,QUoe_73gneyՎ>0p*qD|}6σw;|:ι_KLG{cHW_e(eյ("nb\i=VJ޸ș)ȭlvi~ Ռ.̡[ڶ4c"W6.*>B5u(W& 㮉 ļ,\]F2eM 41n&+G>ͧ Tj8WFI+r-5%VԎ}/;g4"FK -,{wYhir phrd蛒WXfIHri?+W^T}SFl?/G?\\:]y+;Wo:id{wݺl籱9g~|9<+mxr/lǻ?ʽK0`yC0rL+}J$ߩW$d.cSYiB?]X28ԕKڟ7 ɷ4oȭH dj"3t~IQ!M<3- >AUdM8طhP@]*;*HUH%iwTBJti;mδw[˴hʴՌa ǺވV/`.m쎀wF-nTtdA)Fqcgjѯx` Tb@. 1e[1 X1-#Jwi9:1Khlg[bRNCɺyҽܠ>RM4B$y)zOrޡ=C Wב j`4&yfS:EsJY]'4+uȞB?,8߇RrqQ֨2hF=иA CRc}AK; i Te(ȥݻPR`g}ﺪ4\}n{sXWl5Ig,ֈIM(^ Tx.mwq`_]I ;̐›7joQOZJap̬VJ{z㢲"FsE X2GSkF h3|Dw헩oFl/ Bj2{zs-<ƔŠ2+ϨC"Rr+5鯆\3kUeav'`$'[t龗2fW@#uu/1()Q9V9e!VB:lu8^_XZo/xxu:k#TıXG0_JRUbšaaU4n cEmW^[nDGc: ķiq/ho&mkx[.GM {x5d[1V[1̨Ҷ1iw\:SD4ȽϮm:0g|ky{\F6(zIb# (;il#!`l l+߷v911#Y/Ɂ`N9\Lyk JH*T ŜJm ŠvQˁ].?Ia~ oݬ-sPY%v`KCJ-?yNZiP.Ew";pmIbo i=pDKƨ;u >P+D+s|W@wya^ kcܷjg{˛b /g~bp(~g>n޶\p|ckk=" gHɰJ3-#U‹+FzFI:q7(6(6\ ߖO 0,-9Jo&ͥ;u'0KLU2w٪ZtjT9[$ bսJ1EZX+p! &-F >9g#B@|ă(^EV|+>IȽIہ-C՛^Q:N} J`ͥB )XaCSwEq޸wx?1Dl6H[lMHybNg7o~zҌ˽$ZKXjw'_\Gco]xra4M~n(ajd6ے#_Ӭb#:T6@tۤT߾P%Zp{\:ےlul H?[ܐR C!1~.p$Hݵ."i5J'"{xRLO@r1OGj|]_E&aӾB@̧M{Pld&\+7 ara』Uj*./ώ`[HWr=لq&P;]cEbsL rs3bd  fbGL,41Pr*F`:1Wb( RX&cen/'1o(GNxtB9Ϙ3H[0p{$W 1kИ"b΅V7Ďus*M٭6^PpŽ3r茋TA5HݠOT ~])8`XqL]w +¨P0HKݽg%@:)a IiS0f**͢Mc3€2Pe%7(sYKÚ.Fd8#Rax'N6xʤ4I7jA>_c}  }$G? ;1#/ Qdx0uYQ3F ߫I*xuJUO(QK#e%{h^x&z;*x g(^dTbZp-?$/ oSwv݋1q5" v~%IP#bP!foaqS#ß`pZIi " x?g8 H p+({Z4"{8H},ntO0]5+;v]f{g2e1jQ@C ̈́ Q#&:514hYfﶞ?[ѩ* Ŏ&t'i~G3(RNl _I@ᢻ_J>,2+ + =ҰP ӽ%7(e~qOܠ57JD!p@QI.O8jE \ QChXpq Zq3p*uԎ4#`RӟLs)-*YM)NQPe4)3ku+HE4Ì>]2:ΆpNS,2  &i3fF v3J $"H 긇@)_L3 X%MH%c3K_ mWn f`:.kRjI P E*iw8N(a(.k7ÚqT[͇%|jg1Be,c)`T:gqAiXgr4jFȐ s]jJ'Wd^`RNٞB)+.HIWlLIzS*Yf;U/JD{6;/G`NmXqDc-Yol; FQQFN#3!$\H`crDED5G~ԄbbD |ID(64 ;:mY#xgkP:|?AMoKfpVG _{L 0L=Z[1՛ ՛@%^q1O@1&}n)\&= 0;y,be(ޣEYtnױhbhp ap4SuxFHܦgFRP|mNTGh`R.{M6uIC=_OAq$9$F eqG +z"\nX<bS1lF>d*{-5MDbRwKh7رMtcq7WqQS:h㥄ֿ$_Ų2Wvq]H|]kE!biQGقivkYc0{W/Ni/c|aW ^Hqk)'RlDut%B#R΃ w STlVQk8~uEX^qrok Bh "u8?nt-iSHӽ] Fp5vo, )X>#.B`-D`vlQFX(Ly'E6]5ۓf&K],|:O#HMɋ [V.fH>_iَ f_9Lt2ENBEQw;%~yĚ=SF2麛DĀ&X(Iao+G求DzJ'/ʩ[Mгh>u}2sك(y[Ro{AM`+X#E=$D4ؓ1ET_p%N\+%xY8+9 ymU0B-hP:ˮ_;jXTySˬa7`EdT`FormϒbKj)̟yqi7k:m@U˳VaCs qB$,fomVXL;%#sb%osaU2XڼwB'3ZUԎv5Gc;^)`15Ʌ|O٢_b}@V"v΂8>n}80{l w'.l9t5ǾB&P\؎]zSiJomU5w|2ZD2ޕV\ z2mt%l4_.V ~hGMD./dmtjA`re%nx?3)bUTq\&f8+ [L.VJ:51ܬHs+&I&'t|1jn)(_׹C_fs0OD< ;^l>nZۤ%cfmr4;Y*1~,<<{|hTY*[dVQ`|99ꀝ v|{ۥw=i3KBtD0ƕy)$P&yn!MbvAW>k0ח8wa=95aDžX"`slP&j&]{5Ovr6$JA-#7U g+&M˸:̲ :˜R@'s'%܀#4ƋWxʲoxv $tʝ 8$2 xgya* -R΃ <` ħ˷@TH.9玣b> {tKfbd4^^-Z|#-e N`}H|h3JܘxR0#~=Kr;ٗ]O12bKkS!Ld6wDf@;lj\q0z3#*f("1a-hQc^=ϓx>Ov -j.ґF7:4&ĐViWi*kj&O7ty=[M6 _R]b#[ׁzw<ē#62=`?sq$0;&yr ,L dߏȃ[{d]@=E %8rZSQ VX#cVIP<+QV/w$v 4~c)Hdlؐ9ʹjéW9y"Hiş0w;c$өSAROj( Q!ymIy\q/*$׼-9?PEE'?OLB`oe!y2ηVHVHm,g%@1 fW\="BJwws*D׼-QP>|*haAE N $ ҩEZXc!1yVX*,j5oIrQ.;i~j`Ip"8dd7!40['FU'V$G43 >uw)xB'tTܬf΃8%AbkHwؑu%O'Syꓜ'S!EiFp)KcC:}TH^yKK|rfDkOTJ/hKN"JYp?uO8r(Tev'o U`sD_\ %*a"; frTXxr7Bx @VL"FE,s`eHByW,xc_ձ3*aЖ[څ.jxۃ#{kw.6ӫ%uĚQpDELDeͨx$4ҡrݔs1a2=ˆV ":ʘ>nk,Bγֵ&A9`+BDבt8r@;9^.ηH$ )6( ~s`*DGϭcL$-M\;`Z=~gcqcY8O~tߜभ8w@|/qCk&Qۣ,zOX1% Qttd-4G; 4q?zĎepkJbGX JR0bױFA#p4 | E^l*8փy+!UtVr8p:v p_PCY.S,[zr%G-7pbuԀByͣ>)GhrPOƯPm,W*IR-&g0wMm)cF9AhgNPCYnTII)?~*%7qs>-8g؈F)η9,80gOܪ*N۸q `tFfiR35V/˻Subw2GM==j/랊Q- G>;z/wV '~noxaRjD-jE:6p#4c+i#>~"[|'…)ZEU޾V ,"5|psrQe o%7d`64&=5|1!LnUP >ɒ j%rY} F `$TF,my][]Jn2I3a\8M6nsh;t]17һ;y 2.f5 hO"Z|Sm- F5Tf@8ږ |b *:t}yudFn㘚kc}YDB (w` ZSr@S8@.Ae6iPGtԌ<`#ޓrP39jAn!p[(, +nQwt@P#8u=46 {ERF:a=Z悮1R |v-54N(q6>嶜b?'t鿗A[xޢZ>/pi瓤u%[+rb?liVNM䌏ؒ&Pc)n/x(Q3~͇wW_ &,kcD.WIJ}@BU$W3A+%4mQa+U.K4h?6)5rNT2vQYyɠSLIP*∴dvX序߂AyFK6f`oXe"9}R-J%[ZO(ޣz1|raYnTKth G[ϠN7-T7TFɩxC'yأo::!RqI>䃮Zh49v>嬨Z(TXC7\"Jy 2&BY>>B6qI,.XC ,jrY m:yLc2N8=@GdcsyI@ḿrYh<u?X`$Y \\<6PC:b;r7?vcZi qN#TN2Cn_SrQ3T4+%߃ˢ_l >ltl6~=[/N |(sG/R3R'}ZR)*\EIY+xA;Q-H7w8xdqG]E8+Z%{n`UuԖw:1WwZۋr&J:Y4lٽ nzl (n_/ncݣcIPMZ`pG| 40g=E<ӎtĮSlĆ< {*tZCVvFʫ69bZ/(dŴt(cP9$\ b]qD+b3տrDUu(X>KW&uI,'YqzCfľ.޶rEW6sqi+`, ;o[F}OVnXZ6I~Iju߫"59[n;a r91v(rXdyVБVJ 2 -+%cLqc&ʖx96jHTW_ll3Kv3H\]Ed'Xm@ڶQC%2e#}tp O4Dz,|N<s5b`~b5/GfpI`ϐqIɽƣOb9tɦ¼W%h=u(9$du廖[ysQq+ ;Sشse;f5ɭgzL LKZEmBY$wx&65cWKH:] {.4JNH>ݞ娐ucRb9Qv 9!t05c)tI7]za^$0;h-0Ŧc95 4p{oI >mG:{8lP ZF]jqύXz$a1%G1zŅWX||r'!&C:+mō@ѶQC=΢nެSbW=q|^=ׂV6T(9jzތx=>|3%*dcޠ%>HC( +k|D-)R|1AyLWm-huܗnTn:uc-hN Մ|wR`ld.P8tBBc#$pBj7_)6AV܍P]j<䬞pf; sl*pL]ȥl}4N3{_n2Cۉѕ$6`<1)IÜ)I7SI[9Nɽ_3v͍l"1L6mwb2y}-G}0E|z!^~ÛRk?Mt|'݄R_/fkF,,\q6=܂;'cj`E*ʧ*֕$w1uPG 0ǍڮvUg~MA$HMܽgR!TMy5h޲Sʽ/R۝.R p{oΒL+U ,KCV35OpSVOGoMby"PW?I&TŶ=Sh`'O/z鿹 ՃQ@E%zh~=<~GPq P^l'!'OOa ㄒ$`YƎVZKi*"w-=d? zs?BI51|LӆJY;O4C6Y)_!GVAoBj04F(%@99M+$xM (BP)<^Y hz~ckl@b)q8*+i^*Oc ":}RG~y[d_J;#<=ϊe[-l8kRf#AS-afXDauAE킊 K ·¤A.o57aCVMg ΫBueCY4`G3j<'LĩwY#:Tđ80Β+s3ѭNnK,:MY{'.4v69 MpH`&;Ca x9 ϶6 1 AI]Pr7RE,sTyeY ةSU"~*Vʫ =[Q칟'C{吺A8jh~!Io϶vVKL[M j~7 vigԕ!heyzC> ; 2 ;8 I?3Ü5ESwx YԿe0 ²5xAp*B"8\'S63S)PdA*_K*RB2Qfc] KI5:dV 9nH$n=rb^g!qSQF,A4ęuE$TqluG~DyZܟ>j9˶:Y?W(qhL|4!w"w빗n<_,Z e:!Q1 #.#N!:´pi}2n=a s"R|¦szA]MobBugѭSTx%Rwn(JRX#4A8sX;xþPzfs|1: #eF;2HHSqĒ[G^_2mGkB֜4MDN\=m I1uc3Fci$0ttL}1I_V2!UD4/zY}]ov1ewG^.:8T%p\ǑjmLzaOWK1|;WR*ptN%O J)BCx5? g.>˴ك2Exdž96>~f}~ǗyqʺJ?g)ہ0|`]f;7aDZvTޓm2D*3)q q;&` HvClcK^~\,>_TP>ӕn@Z'/CSD:-ݚVѥvr<įM[2Mάr_r% p%I'd jef */_FNn>ME\r?R~p*b*1NN/&Zf0 N/M#֪_) "ruU,[Wί18q'LA'S-BJpG^)7cJ10qdru|J^u82\꫉kk&i!Pd__,OY'uz a2UQ;iLKR"c zFo¨!YIzV?(Ofƀ( .!~٫l;IkL] 8q8B\q645KtƾD##g%Y'1%N@JTEˀ~kN˝KnQJ=//KpE :x0'o26[ؽ Ny}l W14q+\eԿ;ĠO;z.<]`  0QY O#LÞék V@ow :lE-uEQٽ&*9o~Ǖ7fЁ_HBG(ĉ<ϧENɛۭuLy<?G,{T6֑r"&N$iF[2ViIKF WУ7l[W\3$I.'IzR cN‡s܂noBG px8nEDNg8 gs3Hu;Om5 c׬wGA ֪S /)/Av^y@ d]4Vc);ۖ (Wϟ}sqRA Ws 5Nia8npTT R $x1hF{@}w*}νNXl8V{ ƈ̨1b GALp82<2 ?- [}Έ  C%p51$BNUdv iQk;Esm'Im;?awSC^ :4Fv821A;[Ǝ|ZeF;ܗzYkF큌p|BI둔G=H^ xU#OcGy@EFrUeM=pV7l ) zQ `pB98*0e(~ zkܰ`(GiQtGOnGxE%8Y"1"$SEQ ܑHXl?y\As>9!v.zw}V+'3Ǖb1y^U5`&Gkv;wYpqw嗲jrP)?@=ZH/3d+cW~_Kmik?,̏V|_ߎCM ݀j6lMH9`׏ `n^?hϯ|ʋ+5_>La&af~|jTzr[y3i 4y^w._I'(j$%иgϡ5Uꗟl{ٔQ?7/A37W}[KBcQ4aR4fQ,\`nbC4X,b4N,R$qf)'ݷtMmlvyf&E1v2[$(3y}I qq?LvD.e{繏w<7??Af/>ޚԛ\qmՒ ZWOqpMWvQV]az%+T(|c=݋rRo^hg˧ϫ|bl?y[-һ)YlMvxM벫LvktUZ*ۜ>2twӯBn"uS_Ys?YPu5I9$Q(!'?,·*e~w=vG2 'W@ѓ"]NAF;ÝH2TpO ʌdFT)z)F3b&o.:\6WfNk&k>IƂ+J<AIc& <>Cx!@(~_apJ:e7oX| ziQE *׀j|h(_Q04Fhi8 H # % sӇi^` αgc А5lF9a(22ts$&| u_*Ti"TL&A ^B51PL`I 9{:sh(&{qմ3QĈ# *i$S__("dftQ'9!]kPV U\,l]qi98uwOyyk WEyilG:7W=932@ufxBoxMns5W刓"T&,_8JX!X[QgB\a Fۨb5໊F$͌T V ZR"}R[s(] 2~/0(< CYUnnxЀ@I5LuI3Q8Dqk$X),8͋S PxfLz< O(9TOz0RBx8Y+/KQE h518piN^)4F 1vJa~6*??9ZAooC9ZMc=*q~-/LFOg8`,P^Oj <94F45>-]zFّ,B9k83-Lqb`#+荬kd-#i#. _{- -Ӆ |, Y%*" Ԉ;N*=f:-"ޘ| jpT. e J%7T*zE d, 6/9 Pj~,+q0` eI?|^ڹqKfoSW%b`Iڝ#*[N\ YMHa#EngtȈ|:D&co KLl1$W~1EU@:kxѡkbA[@/]QLe % #(* GA-V1 ~IU⫬ :3*c4Q;UHSR0Յ[O麭A5 jB5AV4.c6≝t2n"#2Tz+ Va#DȔѸ$Rq?Rm)W~2M%!%.4pF.Xg2&THiM!Ɗs,r5gDFXp],b*-Ɖ~UQ*RC3 L qz؍"h?,2%,4A680ӭXr;EQju'_I!NhݍA4sb4uTrX=`lE#$|3z6qH3"& Ȼx,AAaK͞Ն:at+˷RpC2HLh/uu9i7תߖhͿ=(s4:Ke0O@NƤ RQ׻4Ok5OkSSCA`CEI"via-b9Jj:"['%phRpbTeM3Sů頩5uS1r9Mf WwEe7ԕ4xekt^+QK~*+&Vg\~[g5p Bb0KHlYzKX\וHiI!rZs5%; t{UiIiSȵ)cS;o[t lw`GEI.” |uThS˘B<#">rjSzW;Ɂz}H!N  Ci5ݼ <ݥ̞=ۅ!6RE*>Iw|>E:6IMn .nE{;BL%ޏn)2+%\oB{Γ0G|2K/ovYܮ_?s U<}+aya$ZgC:p ?,&EYܴoftϹna**a>Xvj6]1hxQs?Р?jGɠ_??NV|W#;],ˏq=I7~-v?-{06)6D\ylb5ލ~ڞD7p1ٷFj_k2HҕomfGB{CfOoc'#u `K!*1%'\đX(s J*JOՙ|QeYO#|>a0U}O X2[VO'66'r-]MfEd4z uIx<\M6?N`aT}Mm?GZěF'ZY=ޕq>[ԋOixRv\ hZTqmT vۃ9:h(Y{m:Vl\ ZD}!N'ǵbo9zw!f"+j;ienW͊yt=7#?` P!xu7_dbи/#d/%U>P~[! ߍ!NA_.p 2*;Sr(1xa%8c>,I|1 S§J >GJ)uf32$B{`*6 h:6&6S]?qC/_CJUEsz}&kNTU72ۯu GUm<=ca:[,agLh>-a)?8k}7{fmh_|NJqy1pb|FGɭIt{M&M6OnCqvbFŻO]͎/: JED̳eK.)Nh| * +E3FsX++GNjdYI Kt(&Uf^^k1_T 7Qm,8Gdqe5-J W0dӘK)<)6Ld +=',B6ĝAJA+vycXZ{A~M4֞nNAaFl0ހ]z '9 Ue[h9c*YSc)C% q=ZS+TշNuJiT'T'IRxSBm:>q*s]$"aNreJc\ Q}(6Ɩ+?^$)MtH4nf"I[$ՐS녺DM}helәLgDa>!zz{┷ WoST cdIݙtq'.HlE3i`ccSbH셂kyu 7ZHpݍ"\qή FICfOUӮM('Lfdu<&:zZ=zpƲIz w%AFeN)zowa0&`r ^@?'!8#(I-ޕT SSgHwceJX!IϧpgB`f eN%Wh恕';.#Yjf_M!>hueOJ^Qq>3A+ %z@ ]Ge .њ&iBJ}QU8\(_Jn q|Red3} ,MUgAd_h3FY-# У\!p#V' krǣE]8\mlAO!>-Cx#TӪ(cfbHK{Lhlu_zUQ}~BUuc*WBima} [@}&⨷:{`7Ш yױ˜vce%=ygX]sNt|͉rVd>=Ahg=] fLkB ̝2L,) SpxgE7N@/ށ3> 4[\hvX-(#@X2\]k[}9v ~Sh$_&E| C? ~s\-d %:x|^zI'A1F ;nqv2ޥ}ͣ: Gv4 ϣ4 ;&;lZ&7ϦħzN~8#[q:_NV6{y;(At~5sld=$m[m["Y)bŴ`A~Z">,m.7S>t3% Ja=DG?\&:[.x,}pv.&ұb.bt{&H ?(P@Hi?BMb~kτchggi7{~䘖rK7@<pbѳ|>=r".qa&˵ĥ5=Z9]'Sy;E%xf#xQ0 2*w6+= G\.M)2R*Y~eH~=Fô6bFL+nsW8QY Szψp^akAR{9/K6s'سVVcfσ7J2C҅rQ>W0Qd2 ,"8BB!e8>iuDHVِdY10y/&F" 3Vʼ_7̞9(%]H-àg޹J>qR{#/3[Uk҃E)ӻ@TN$%z1(4],IPx+qzFp:b W`DC>Px,qӀ[8`c=fRpfD?+˱VO('(= ΅ˋ_<ի .%ύ3{~V+Ò[kYz![r6B'xCig5HL@@/?`ý8nŋ3no^Ww=Wf{ ػm,ŎONw1n[tWڹ%Q:JJ23(َ eSrg6MdOǶ_ps888 &90S $dfpm,(S4Lc})Mn|`8I#VgiW;sda*L"CBZb*>T(#8Y~ZaRTd?/6yZyca3=\ʤ8dUZ3b !yJ֧(T9~d) w?0(rr숷rL $e{fO8H8y^BI 'Be{ǟ28Zx6Odl0862>M5~ڊh|XDd>rǧGvU X`VAQp@;W;u[ۧpQkc%n`Dzntږ. 7i(gI61G q=(\OQ{d ܃NrqGN:2%nmMFހI lWd@wUd6*q=權,E 8>)<|Vdj wfOC&f+IrYb`NinZ=r~nw澦|#bLH,YS @2'10éKpN>0 blF awSVÂ}lEY:bQ DsF"ёV)(H J(AuWi"Deb+FEQD%8­Fö [8M`x s fCCqd=cz 2}b'#"YR^# 4UYĤhSqB1#E0ڛpksQ`^':SI6,M>m{d ;MUD:X鑉eEK)$`} Y(͸50ZCU*bV2,Fq R3cI)FyVE:zf("ѳxC#9h"<q|V-*8]Gc=2S\+@1;7[yoWH4=0}o5{,ZEλj 9"GܴtI!1mVHG_/&}PGMݷ=2gP@=2GnMNɚ1DdäRef24pd́ 0`̎aU jmցY:dL7:xm.Lfn@]ieJ)P8IiCSAӻ~%أ8v=2ӬYG<JK_U5YZ-qk 5R::Ğ( W?bP A,/BK5:,Ct$0 [G8(WFWCFAUXUE ]1E0(]aYȥrI4#c`ppqgBדmW>,%``d 64^lUdJ؀`06lYOzY{d [d\+I0#&fI3ScEbEv`fauTg"G+j3jU inx]50T&a1]d`[!ʏ!V+Sz`?r_fոȋ~T4K0;>#=&`CNGaBLaENc9z@Q%U =*f`m9-[iL.fh3 P 680#vi,#Y)Bo@#TID T#Jϼ$5 bVe7 ,"X~,Ym:p_kUݕA}MksV s;I~5iv%ui>˂3^Xy,Oo.tgς yL1F&F57 mq w,gы.̏ge}UMO}7u"w+;ߒ@W|%gj)旳U[^~Wn9% hIsMKٵ|Ժf᡾2]aF~nj>JgU5_C'z<T6{ ؝BΛwTJuaF8DR]\߆zao­n Q7eeX$yQu.Xf޽YɆYog gO~(.'yإA+>y6וeg6~QA핞³0ӷ&YG!”M'y,#\Q2I113}3Ȅ% qV.35TM7k}Juw͋ai/7|}R4 Lgfe]2h>3ܩhlxv[K,ɮ] 3zWƳl\v`B: FanD~%l\o}`t/]_;OM2b%  6osqͶq,aZY61 4t4zrq0?45f7sզ zФ`։N/@JfJ[?ҕrnqZ5w׮4(e \ʑ//}ֳ*!0"aSͧo$ sqH( %$L9OOKL.Rwl>+`|=j/98%JB F$ɄByhN0&Rf,Y$$ix}$ҳϜ}AA3 Mj/kS]F7xSgfv K݄3S4'N2XfH4 Y8ʳ=2hNEZ}e=˓ O$AD[(m:>3N>z"ŦxNC?>YЏ|ь>iЏ'A>C?>g$BR7x.7 4~@7 4~h 4~7~ͨB,Js}!"!)ϹLCJ/1WTR aɚB #yžR>H1ocϞYǞ֏?"s, ̔_lWn5(XUu>S[:IK7>,,`['`3gb{$ WRfЊN!nz_Ķ|1I[> r-Fy3'{sZhgQ4ELɿA1/-Ā;VbY 4H%rd*Jb@,$*JD%8r4Jȗl+넸x)E8Jct&&jsb9!4Q$9Nbx~Xc[>scxgQ{vTse^M8~:XO4v$ci`Wߧt~Garq6!,U%+SU &Vwo@ fvQT_ssL_sI 栯9kD)c_sS6_s5}A_s5O$GڿOoA#.5"o{"9fk;h]WcoLI5[H8 ~ŸfY!~kV߮/Hb*kF3*0 5J"QK5G2'X:CI"pJg,IOm^[4'ѧIM0dF-&Agн\l1@5ieͽЖoq.pe$597XftQ_r1lߦ!<$KB($T:KUEp%;EOEeʹM$45<^W *r 6!C89 %0P"yS}/&ַOTDbgwؽ _faG6g`h`>r1DֽdS}%#cw/޾0/uؾ25GN~6)I[]iEWiy?2w7:HgDfz_ua7Rsgw}vȀSa3ۻ?-3ə!U`n(c1 KTHJ$$y8/Mh#M V pet~M0,VRn*|p%Rk\w_ TV,^ [EPti,4xjm]QG"PڄP 5}2LY~qvA=kqk8!,ʀشȀp~cˀ*O n~OcKsg]da4lFq> =׵[(ܲ S1P+A/5٭PfMq=Po4ÂR/zb k鋳{r5*ڙaOV+_R 'Ujő$oy2X?Eӗ.QU$ٕv[,on5]Cؾqڲ5r^}jtH&0peEL` 1JPX,(B^Ouؼ':DQa>g9MBh8)u^6N+g!Jr4ɭŹ9v n@hz B޽WQ@:k|ݓJ2[ Z [x/?&i8Mʷ_?]t£ɇ-r}I&U!]>cp)>EǕ Lz`L6܀a8XTFd8׌9$PG 4bt4y1tErA;P)-^J52|> !SrV)ӀĽ%;9/zE)@hZB9%`@M p]II? é顦>lq^z2Ꜭ% P /2[3,q㝓^A$3ϙWtӺXҵo@67o_%%/umǗݟzZ]%UzZJq M+;^}kN]ڰsYq>q{%uPNӋ;G p(A sTZ'pN ʏ`",lszG,?i5^mDs3vAH͛ŵ#Z].݆2}jAy<_[rZ/3o,oH^<]PB9q<߂*/˿GP%TO_tB`[4tZ=^Gb'/>x}>|po c%xJSy8.ᇏ] wնkKUd|qjQn˩&6tj֍-8mb|tuG =sx:ߡH[kZ/V6]arEZdk\q̎`mp{[j&(`6w#`8zɛ→}{q'~?@ ;m"pKɏ^껬LGJz=ռ[>UzWSx7qz3źIzq29=0uCj\9y.bv܇A\/7]B<(=,Z#m+Ӌ[v֟hLQN&Is/,JIaVe<ںT2[ ^R(硣,šߟrפ q}ڽu,Y A䞻2j/rJ&|H2yb YPKL'P SLڌ2Mvefy~~6ћĵ[ۭ̚hb{[%ݛY^ }lՅ;+"GEzHG62 xc~q/ş"TSLa~q/18_J/m(j27IMj.oa]7E[-4@Pq!hQ*tEK"欵 F* 6 nWsdnlTXz}|uciql8l3NXyȐ")")")е")p@C13qQ!!!!YC94C9C9C9C9C9C9C9(ygwegygYԙ]e>q-,RcH)$٨ݜ@b̉4g$ZUv\zW6Ǖg_ƍG?ădsĩ6[$;I#ѓc YFs*Eijn!x"hi\DNS!/)r) œŪ{͐V9v APof&FW'[B`E%A"o-"o-"o->H ykv!o-"o-"o-~c_Ü,:Yxy#^$bOD&*)9ȋ$$SOm@=因 @`9;< .ʊ ʖ˔i7T %aj|޸T$ђr*A@Ԋ'#rz3qvP r6 %%:MPh(L"ޣlJ,bf=#FG+.SxPxmt<tMLXL rRk$Z%)βdh13!7 4ـPo^꘩WƆlޟ BSYmG4D-x_wR{=ְȂT2KjP5:inD2 u Kbqh-+} 4wWݟ({޾v+_E hkucqbm\"|@ R-, U%RREB sUqy濴>}@}`E)&Zww‹ [딂l:32|75vZi7u'@nja1g H F ȗtdv9q9ꥂ(yj(8~SRշ7LB6&gN^7/]X^k"OG0w?O #)g?Ż}$T$>G3+`)L><&qoG6C&t2η hrE厵jz?,s]7'O'cݧϟG{FRj0`ht^ Ԛg hDr,џ tr:-[zqh9+M^Rk#lSKhd>f$M/yt^ ejj/yhͶ^{`_uT^v逭:vyd2K7zK9n씋/,Gpg#N<B}¹L+{:Hy cώ\Gt6g{5f M=6g_ޝ%쀰r z/!x?H=Rp s!En9![eThkם$~By b_y1/żxv b&y1/żb_y1/żb_Qż+y1/}78tr|y}gy<_4)J tqH翽y+i"u* A`X%/Kyhk"0O͆H% JM]2C%P,c$ƥ#!)G\]B ZKH4JY "3<7Ve=Jh<<͕7V&FK-k*'إ(/(զLX%tОpE#qu̢BgiQw+ߧ&C`ZPL 2jEіx'L{ByLZx,i P=1bxkCj){#dݒ`b1R)h"(Iܫ rWN3O*"S7סo 팧F_Y.-S)llCR5i]vu]BD] AC%P1-[+]IMEKJ JuR;u@yIj%NRA$d1äDy|Rk#0!o@xa.\(y2;|O -flW{JsH I!girK_1)\K\($K03Q`k$@&AXr]UIj黕SB 嵐9iAoߤѣdz%RG MwyC&āBBB3Y ԽC%PƓW|KyztPkc}݃H& ~+E> FQ²RY*<ɯ)^T띐mL:z0J8.l"`u+'b̄E@l%Lޱߧ;xs孀LL{{ UL+=zKhf\4!Pk^qȜ-qp"ٻ7r$W ?M<,xؗreUY,OPGJLU1U|h-A2.~ #ӌ` 5/"CEU)/paMQa΄LH 5P=ma[ .Lo.Av-GdP;$l̬莌\ 5ϩ]A+H-X9i&E餎pR U E'Sy23HEs׃>j,^v$6PΩ:5=} 5 d 1'*,'B&I ҥY}(X`ԡDGV-%<[ 5⨥@Jqx#܇BţǫB9j21 *&UN dPxS%}I1k=A݅Rcl5Da}(TY<*~)XCX,J:pDx<-BPcj`S ڮ!Jl KA&Ʋ#ևB*y$9:e$;m (gNJ E7>wC♦zqz7 7#a~Cf_oK?ThHQep< ]4% 0Ti v%n9t(1Rsݢ~SwY?߅%>H} |/tO'rקQ1أ%uJ1̸CoM)z@>eLc92 gWsurQį+Nޒ{u {QIPmK_uk!lD)b*'ِJk.0&mlꋘmc]f_k{(Jvu U+wJvSd_d1?廕xb6th&fJL0%0n=.E38ÝcLb5 e&qOYA{ePR#|i>.J_Y_ʐ}HOo<&x~ud Hy>YΉn</.3)#u)nk(js ktry{ml~nPn֣; k/94}fg.{D/b,絘 ٱX+18VoEH8θL a݂9zmE:';D_8%6T@b?ic$)3gKY[1F#qFyS}4UgwYކOiUvτM}XtjنN<1is-;w^0i~1.3^sӻg5qΎ3N/wfV<~"(^l:%1.Tz^h[{:ݍ2($;LPpb01KH64VR]T)'"-og҂|^ݞm@/HL^3#v[%1F(`Y-x(8f:5yR%v%g@Zmdž˕%pH|48gxԀXJS@Bd_e!W0ެHg%S㗧CFR 74ve`+ OfpRkRb(@@xcxgBȤb‰R=jr4Be[6BRގ "蜘r!G RR~3#;QӬg2:6|ҼRl{ mC;X# :ds uGf2~D});Wj{Gedy+; mvV 1oƂ,0 PV M943Q3wBrdhq>` b7dc%--*l[mm!VB[+kRjH0h*J+~jTɷnyR#ZgBwי\ p]%p$DE8_)8ּO8cJL M)y9%/唼S-B5Bf>BfSl M!)d6̦qbܲp.z} BDgTnV#rfۓ' h$VK de .y\7'_UX7pY'.;г %V,ymnkmNG 4#lXn,WEշpMXNrؑܛ]VńJmŀY͂{c͂/XWxu2eс\Fd=ء}؍2kQSBstQddg7V-n}6r;ΗW _qmτ>շg/淌1B9Rk}^|A8%N6\ bG|4,+|V݇G!#>'`z.5*\Cc} LvNM< 1+ZYçMM^96y 7 >}2!rx_dш@qIC;zG/^4,̗jȰcqZvX3~o83=j `H o,+TTx(11@ 1)bO`cvmq6YlnrVb-' 񁮕O}6ャf6s:Xnkv5) )3-)X>q knsii <EV3= H>60]ijHXRqt;њW\`K-aT*[ٯP56]la9V+0ǗŲTT꯳+ſ^|KI 50[&| {huf"GpsY\U+} ?ǺdsEC{]"%{# (M>fznv }&}{Z,zOԞֈ9JPdpJRnZޔ$yuoףl,ٛ1E|5覼}{wÚԋޖxxJ{䏏Tsv ۗRd^S.ԁJm雫ãa񀬤}aAXR::mò'~>ޗ ?lN=s3%ů~68Li H2h˯~WECy n녃?޽ߥohqdzzﻖrj97o3ʵe@B&kSԜY@uxQ40Aw҃ |p>$G{FJc_ l,x309;BdcٹSl],+jԶ $ݭ&TUXxVB?!f6"*O.O힚aM7#+jY{b<_NH \邰.E:kk3ʘ F-2^Ӌk|4WvSEʓGůFCDq}[Q4b>-o֣?~? |k2-&'Z/kU[@."+.@&psJUrDa| MNEuxJ k4L셂  h ";(lLEmQegA/ Dr, H)CɇT{ȿ3J72Z7Nu-'dCU-7FoA^E]I( L~b&`I' *4r#R sT$~nvQd`l%{uBQ{7x VyCvy[ ydddɸ+H&Mc]["b)L.qM 4 <3N=NZJ `s)2 |25M d*Ӊ!ZyVd -xO"Tw)g*&l~G٦ثhXĪkNG]QdA"YNI'O2Nϓ%Kqy$}Qt=o.NԬiȤsbR-ڈ!@ 5JП&~$*H s|rZ;m"/V3 VDJowun-G𐋠sH/ά\Z24c}<3 `Fz덳;fE/=ZK 8gQ&;B&bw#9ZV U~ܱt[k[ثk7痓ɵ4C.nJf_!ғQhih"WP<}ҴMPVN~e]Dr--nzͥqXR\ KA`RXIyk,4'-CEsnsI5]m_Jbá.R-!Vi)r@m02X4YDrdEj{ѹcP(g4n\}2Vzgj,ɂRZ(Ek {ecUEg_&MO54Ő`ǠR4ȲR!@(gLFD# yK8kI<$ғ H[^t>qB6Cɘ($ JnHwvBQtgo5{=7.uG Q?SUد;Q?6'nI>f1;^<0F.?hh(] 59+sxR{lN4 m8iKIs9ƒwb;/u`ڤ5ժ mAi2O%z?$uISS9s+辶-£ËӳC AU_=īG;O_}udK07kϴ_|tcKk~īˋŁ$Yu}Mxd.[adj&ϗ]5J4k`oK(t mF̷̺%i2IHde)U G1Ro!Z{V\:+Oii#aIy(f62?߾ᦊt&M:@}u?~՛y}W4@L]Nt૦7vl[6M*z% ߡ]Ň|v6;eBQX  Iht٘K)kW&&.?B&3?f4k]٫RU8 1]3[~|@޹_Nbw2Ӷ@")Z_ jiJБ9#2 U+Rv,g'jQ]]X>&Ȟ L9JQ| S,0Ĩy!ZcRdVf..vTbtZsӭ5k`vY,8YP̧/XL .x貒` $ qߪwL|;q 20ehu:[]$I\0qV֎3u=/{0V5j>IsAו1퓛51uSMsŧGwɇӤg[@`ݯ]_f.j;}H'Af9^;?}[Jdxw+ێ Gޮ)ݑם&+TɷdY{R9qNJ!ږG͚*7DMwyIe('(Lr;.6{8zgIf?w49ѕ%U|jLŽAfFo19 $Қ )!P%!LB{R8.}= fkc\2&fey &cGNQ@+|σڷ28T뺣u]ێt.mn:mg_f>+ƣE-FXj >ș ^P:( 6Lh9ڄhelQ,v H0S@A$M Ecm*edm aa_XUl`Li$$ YiQ#Ő|`RIɄ@pNRp0$-UK-b{Q tN;u'޽Oyo^A\'޷Mk+>:~MQPRݏ!%̢,Tu,B䬬lZpQ .3ڃ!kuKS{v8D'efv758@j(c|uS#ܿ3b y%Z˨ pO9Q1E QJ;1qt ֝AצCF2J6 UO2܊a'B1XH"fuv^P]sΧE7OJkM+QG4S~`6&_zR֋' ͈ SH\F%[܀ikUT". XaG2KtSV"Sxl\(4o;v#ߣMyDjM 5Ak zH~W_5:_wn:MY֍Vx"O@QK#Zt̷y޲L^眛|2M%zCiIcb]E9H9*+΅xYH^ZcىEoܰ9J+ ވ"=_BK|Ţ9ρBdYJʘFeFIМ Z,Wufb1qju6mZ [ub%={ѹ㜎y5h~< ~qgaܡE>oeʐpn, E},epS yɿA*K`h(E*xH2A(@WESDzԻ M=enz6Q U8[Fi\Nmrvm"ZN>z~sKuv Ϗ~rDHFN׋sw+0ZoOH jiw ĭ/pw2Wښo?PœwvE0{:@r鯃4nn Rm)Hl/5C d iTTXT)3%u8ߟ|~Çh0ݲ8N*bUP 2KxC2o}d-i80-]x_?>{y9hSg|.j D<=dO@r>p h ,Z4gMzp#+>nRDU~=`~~<+mi9+3"ɕOd}>|p~h"[Û[% L=u0lt6iH*on+."h@Ph9Iǎ6,07Pbx iꬍXǖY c%Q cVXYcK4t69}+ hh' v=K;?S>+째nWuNj6/x 0m *xuIpͫ^G]"> Kr8kܼ:`  8dv&Vlyǵ0+2 wa\XqGBF ,4R͝"Lq¾3J?lu];_PlHtUe5,ѵigZVX<G0H^К-&a.˞2v{yAz$wQ0 _ +W2<M#Myh.>>7yUtMM8[H*s}oGL~w=hy y>#cݙ&|~ >.d 9x<=\> Ʃiu{/[.+@A:p>%*< 2 నmAc:)Aq<:kF -aB{ҔdG6*]d+|z4pZ3Z&v}LmdGVELȪ(Ђ|ԠxoYӆQAbG RV'MJL5k j쓦 ?"M֋IIMk}쓦 C/ݸi*DGL8dZE6 }XC]HodN7'}ߙv]Q|N,c_V]RGtypfA^ܣO83"㛆ɱ)A?jF!ow{<)Scg gۈq,>iSN!7S>)S#.J!2}")^ |!MPNN":bEFїvQir7g^#*`2ǷV_[N:m\}š]uy8 $ mBߦ(mBߦ(,)mҎP||(IAq :&$˷:Z^.kA"FaD00P,c BQFP8%uXZ$N;kQUZe"[Aa6~G> ;. lz[SjA~&4CTc.17 $VSX;ʴt-JbD"pQ Bǝqgt&*|v ~*,xbwYY#afJp攻&Fk+~F^jĔ+e!Ȱ8a$垹;A$sp%E kEaxn欇:9QSJ<gCx/ WD(Cǩ":T0B*qr;{s=h9W8xB$93Lckt+̘$A*`Xp#VNEY" #:yݦ5قG]^c@?ݳ+Ձ13W;kC'|LBYInZq.;G$L~Mֺ>ooۇ/.ltΎfix;6\މd˻_'=rOVk*֫;;W}zK%k.xn3Ri.rPxKXtNr$N2HQ&EzQr3J6TigQy,Tp$b Rı_""hBp ;H{ g) {k޳>Ss6|"bR< Q"‰N[Jbdc%u%ADDX]V[b:祝#* %.$f`:콑0i]ez/QSGZ5-ݞ_;˯dvT^6n;UEPԝS |gYnK]E]m[P6pW j]y]P[aԮs=l\t nCl1UCDʷ+$jjnA?o O vźECWyeW.%p ut4.3F@@}B/l;sob71ۛJEM5/k:!q˾@{~vfKM3B[B}ũCP< [H{Q3oG&@TP1LYJ}%(Q\]9u3X׃|su=9c.yӬyf>EHJ?@~P#q_TD9Fő@(A c#q-*&1U,^V:AQl( sd)8Fc 2:•$P8 =tөYt쬭|QiUCgX)y?eB&yUz 7R$?=t^r|ӯ ڈzw9䢉6sRa0R5d#FLĒR"*p:JU6$Ahl ; ub`*5`9OÄ$fT%$i"S+9CAQr[VZ׊$NሐI8$,z+"dPX[8 L}ϸky/Cj:t,TG1FD($$8QYAcL CG SDb#6 5+HFT/}' LFALd(Q ,M0'9q` E;Z)9a8xc9ugM Av]̗?W6dz7 ,~̆c>}1F=/+C*WΆYR`΂]:@O.Jo`=)< QĮGd 3_AGUnDD)HSOgÇ^6 psCΟ-rW%S\xU1m"ZN>: \{rt޻j 5`K'Ltod{||:;OG~w+^z @.4FzT VKsg'nA}fT_X\kkzCuOged$#E:4Hpps[nKA gӼCfyڍ$;)o4C*pF0ef.^ `4nYx'P u*Ukm%̗t 9;㇞e.'6qQ''FmԷ X 6HǁLnśwokLOחA{US )~!xAvAh`9krs׀&\(psgv`*7 J}3Ȍ?p1M0.,LWZ<7v.1r03"Ϗg1| 'b>RmQׇ~A#BTخkxu\# {īl#>T}ڽ}J;,L,?P?Dzr"y<0lNajz-woN= ʪ_`%ʱB8;㇋`(mb@Op! ُ Okٻ6r$4eoq1ŗ ,0X.ǞWlGjْjM+Xh'? z[ k\LQ3+sv\\V,)ช?~8@pbhlR&:,s> ` l|EرRPiVm%mg)|i᤽N2s#k$ hH%#<Pwl6Z8H6+EOT7d4[^Z!?*1gVdz*Jo{.ޭs=R#M֞'N'/toskk;Ɏk΋ ^v0߮WCv—ЯMA?zMO *`m@}uAky8SC6﫠+NnM0[WgSMVcus>:͜4 vfY6}u8$ߤȎþ5ޥ'#h*+qbF&'jh2aD)&!6DbWN v?|V֒R4ZjnnYGr9=ONqo-9M]qp$"6rd jmSt;xonm^| c*9#/7y-lK`ɍNv?>Jʍ'YyZwX (P{yxT2!iJfs 5.RM yLGCGH p,tPv6J|Ko!Swg$rsyx±{n՜<:.e|Oǵ{8ܻk;Tiی66@X#Ec@@8ȳM ]U%a\6?\AQ3L8˵8jYU{D>(.B%@!U`8/gnq1Ʒ˕v [qN- QYw#Z2oVS" Pc−MZ3AQw@rR(q.5M8W>Qe6μy3r;P *| ;&e} T EJ7pl@aHG%k2>9.h-D<b\pJ'(}|Ռ5vK[ڏ~D弻`RBr^uH2Bd 6X[J؄R؉KWK麗b5o]laӻn&Ai\9'fh#9Y \*/IEMb`"|L+݅Xs{=[W$aIHeӹ>hL) <*,2%ZBD mTDXzeɘєjIRQ5D%UG4t!c[6-{I%cJYK<gyfv'/sO D_~H9USeN9USe3T?@?OW%r}ܞeyy_?UkH*aҗ\fg7} S1?aw~ܗ0Z_F n0lnwAUfr{mf '(o_%5Vأ—Ahp=3-'% GSYS,ּK\c%IIO wLt9BPƹl $ 7 L$xZĜS)O?veNgȧ>$Bo =ƉYm} ~Atq\L(rueP}s"T(~+ƣeC&Fx!' C?U0`2&--w~fz7YHwu=UaPT hhqQ}1U @#Em+OE)8:M~/(>||7Vm+//\/ nuj>FUN ncW[>{"]}D[0P&t mz{lK5Srkd%pJ^kfE4 8hAA]{dIP&sJFƆTgj ]2hTŭ2Wg\Ͽſ?ϗEX]9\3]Q(.li4$&EѾ/bCO~3O8bc0F%:dU0E!Pd`A5"ϑh2:jhdP% 6yA[Et@RD jcf8K]EW6/$[^ѭRv- +U߿B8B$H4G=yI3& 4H&Iz-") G=9fO&;"Cva䥃Iz/KrIxlґx+T2D5aU\՚  4SBKHelrBYT8kru.&uKO=2.bb{l%F.c!;Vt+R LH!QLQi) AP"ЈU4j!MԨ׀8DBx,F.3&ߛ8)LGzp:$ĒyH8#p^$dL^$߆pL)KY*HL xE JJA$aх6*< >8Ip~/6\ F z*=QDd eUPܣ!$Ib^f桤<"/~{YÎa)qM4IY,R$")(*%V*/PTcQt(Xm:[c&x yn/ʤ Z2tp9+ @M 9!L*VbNz\Q*#;UJs뙰h8[aϸ& I5.z&IDB9QPpD"^ oс}'UɋTG^Nq" n@aQOAC Q$8F4QT׺ȶ l"ۺȶ.l"zNّDJ lKZEum]d[Euu-ZC{4pq!pojO&v_ѳiZ$/Fb8x8,;[GIi9"ieR &N &eF2 ,Z#ՁHOr -LѲ 5CԲ[ji7ˁn;حeO6 {?)}`nji̧PоFFC4R9XV[nI{Sn^r H~e"v9sA Tl9N:X /m 5ʎdqIS7fѵK\:%^`ӼmK4t;X}iRѾQpTb^H&ijb (w[ %QM5IlRQcpͧt^Sm2z($ֳuvviJPZzC\?' R[UN)ɍ]8ɽ+69v|7sbZzs8JeͳYV%sݫ4Oj~,5/,ټotwy4 <ʷT\.+YQ2lXlgEom9X5o8# S>r~}ڶtu(M˭͏dCv i6}K>oE/PQT, kl'}ǝ`m`"~}9~5B(QhxOFK*D*Rtz3e!u[5%k ( 0 Ef4-9%C/}8N{1X׼y "vQo`fEV{ŭQ 1 j} "#^S'p+X5[< \m18(8Nj(Cd(T.(CS0da+sƺ\OAsO@Z|OOz&6#G۫. 75eF{w^Mo6%$o\T'B& QNsJ`⊡PwV ɼ59j4G('jHntexzW .wWg-IRp?>H̞vF"gޙ;OtnyۅMN~R |DOn/CyvmKj[fXm3f6\p,f->f=~l{םl\hjmlnu6ȺVƊ˲V"0(>ʼN^6 gI†J%L+;;Eq3.G_~/?>|/~2>~8DBu齻Ͻ:ݛ^Xlۛ5"Cu5;+YCi)| !vs[ϝ|臟c?>9Iy&ZLr06nƷ;W_](>烘|TnOL\?ekz}O0;z,H8%!X$%tNU VHZjJ&L 1Y $%[#wfk^QS_{&_x<[SK@)"9A$hqwhulu9pOezFCc2@vf` NAxv~bhZAg5_:HRT^|ސĞ)XF4ֳdt׽_rXor~B׽yC׽/ΓDPgDUb(Y@QHf*zxL@' 2y bD!Ǯ;s⠅цI_cNfpx $u$h[@(GKI{IIR3VDDk 5H9ܫ> '4F RniJFS4 ISE>[PW GbS{R*Tƀ ?".VdX 6@uF4|1Z?ݢξ;ԁH8dDI038c<(HA Z[Fqzvﲱc`mwoہp _xЋMsbػɃ0M*o-"IMH!)E"1<0nh\T*uY9h)at\JJMd3hFtJT@A!TAh) IT. wSVY]NZŤc[?߯r nSUMb4|KDP :Yz `H>=a>+)tƟ>E;/)8 Sd{edϩ4'KT<SGbrClIhm+?g͏1ތr.Vh>/g,r/v,Fv.qx8e݈<=sO7RBE텮;X9XşEtI*o QLm^ٍdMLDyMG4ǟ?h>CV 5P2EoBkYo' Oe“Zs$oKKC4߬^{-THƱ`c)U1H*TMFv)ʡE׎ZJ2ϛ4}cW3O@a{xҮȚ'fh#9Y^T^3A%dDD09ZW"^&ր]Z-{AKB*2TL!/YdKԵ1 օ$l}bahJT5 $( X@ry QFI{ ϾM`fӲEZiDY#h0#YlyYM‘| ,>JB AQj0?}* XhkL޺k"7M8Rs #uC"O7E m{+d,?^_MVf%G%Ɵ|<9F#sI>?jSǃ@$SB ÁD ,Z#ՁHO)VY;3u]^Qw8<͑+-=hams`w?Nociqs >P1@#638%Τ Qrh9K$ *@oj<i}>FowKܤ`>-6 f\c_ހ_R=nJ1R2d ML5BB4/@Gge)K% B*&:8rws\G'3<_z ks1hߘ*& ASe(ϑ H0JGi/EђM F[K.*hJeoz15]u߶q7`׫B;*) !n&++"@*hNrE2i 8DЂ!VڨSѵrEg07` &f H\kJ=!4q0;c$:'-NBdi0Gj;;,HBko.WJΒf^ߪe&9둣-Zز##hu,>p^⨀^_6=D Ȃi :Vߝāi)e[H!2΅Egc IW}IψO9_LIExzvo ڙAW@'|HHooޑ@^ڛ>N160d#kQ_Bࡘ&v r?4]Xdp ^kfE4 heP,gښ؍_[jk3.uNr*%>j$)R):~H!) )yTz_=8M<~̯ⷷmFV k}Tnoєv0qR.Ҥ5N2ktqsS70 LǣZHz}u,cY[YfH _WFmǴĴ4VgZ %ƻʕ#Ѱ돯YjkM#PN6o+0utkO®T`g[r@U T;,8G$Cp9yA2㜡rAb Ra4P5moI,ΣSSx<=ߘ#gW}k"Tn ݏH Áp.S<>cK.k:j>YU)zr2KLU 򤸨䌡Ӊ rL&Ztٓm6'HʁaeWg;7h[S﮽Z|?^M8]mpO1+?WtqL2GHDx+j9zuٚ> A-]ϭ]+~'Ff)Ƌt%"=;>!1;6g]$-Y늩wf:nn7MHyf}(_{S ~'Oc߃-}|tűئ_4ݾOӤN"B绿>[5}soO!!~7$5㡾ۘ$ 4ja9b)Cwy.A.䦧 ɣs!c9ЛXNK f*{{vlinol ud=/O/':zz_|tsdG"rmHOB.6nEhU{쬖eN6I7GGńxƣh03V;4,<)[0 Eɉ̀`WN(+ź9A*+1SB(āhz"I$3d}4X,8;bQd5I!:%K׶vaW\^Sr?5.wXLq ɔ!1b1֫୔fii(#S9dKw=戒KN/4ap=.y[VEHB]:rJ 3g9 d3tpn ֋zz=S"=K"Q2 ]ԟOfǵYťJF{)I j \m01NaLK]$ecRREK23!( S\lڡQ0t@ 7FPKH)Z`eAHNLSdHVD9QtP !Q[Y=MbSw6gecRv5/1z{3o*w>; OiOwO[T5*EewwHԛgV NUNOug@ފ#Yd{p-9ecc\2qrNΌtq4p,!V@XȫjZ/Fߊoh<Ai8)s??FS!.tӻ"2m1o9w{iINgNNhEn ĉ4Lݭ]d'|w4ܛw{ensvד Ca,rnags+.moדHn'wkG:#] #Z0@p4]h,+>n=џl9NZGl󨋛l^iYVvq9H FWt}=NQe㔟hvt9lTNkY8 sR?r?χ}OO4J6}|9<~}CxIEwZAͻOjQ[}SgnM@R~v0?͇e2=)NwV<~!\e^\&_L=r]L۪TڭT"N+>:PHx߶'K9kI 7?zN~> sb')Д}Dx4ʖ 5DV%[m]NwMX(;'wk^3Ǹve7ae*FG\*$#$X 11Dgmuۚ6 tFyA,e|pZ;Q9G;G4?QnZZWuK_"Yy-Wh *H@+0641Zl f 2(ȃ˟ryDRp!xDJ7"O4 g@2hԍ ,Y`FRah\ C,ǔp(HtlI8I½ H~M(4{V6,{NQ6FsArfC6:.]bJ :^C !x2`rC&Q:xJVh݌*Lv㘍N!LOF<{l C]ݪ*c}6c({R^)4΁^b>Ef51;"fAT}ɸxtp Qщ`hd4s9!gYv;"]ؑ2ŝ6ְ]Av Nx4HjTK?lJL"5g>Q.!' 7Cn 0ה,ȫ IKS=TW_-THQΦ UDLG$#0s Bg̱IU&C{娊^mҽ/Cݐ1yRⵈY7K-wTPThἥzKB/=1dԮQ-4 6[hF[B]c.]s 1 ,Sl]y f^,o"@]CY}&7$ \e )Kn}L3; Bg.hF LFj+׫& }7o0Wɢ/tM{uQ^8BcZQR"'.#"(I Bs))2Q7wYtGn5m n %| 75r?oixp92 fsųڂ bcrS]oZTzWUSMGڴ H ?BJֶW0=o ЌGKt1r+[JGwZK{1 ߛB+bsM3:~8 3&-<߳s|ĽA9Yjwj6o~B.[o3 ,8}S]"Qk mg~W`2#z!M"()EdPљ1/ZC(RK?-KUQʽlݔ:h:I Z#+Y@BKM%'@D ,"ZϤ& r!wClpYX@(|h0Ť2T%VZT<B4.Dv2$djB[B:}`/{N4,=.kYQX--.y2ICPAtT A'u~ |!^$+_vW_aniFjFQ?0$oCq 9 N=~]]eS.笋avl*MCJe (B&D l6 d+^qt8v3ᝂnZ~ߎ71P"H"2@R-@Ť]03e<K[w;"<3jj_g,rWÎo~(xd"J>S)Ey穀bNHS%zAe4HrA^<۩q]Mq5)lM"Lӯ&v2HM'؄W'K9ٟb)T48a\#y<۔_Dw@[l[j( ( R\J ^!SbN`r(4lȍ w+:%]˖u?]({A/_\6>~ݱ8]"Y$f'P }A͏FsPJ% v"3d]jZgbFr C+w+}Ql̬8 u,Bul#l^o''oo~ۋ}ͱO`x:l3n<ýi7/w{/7r "61DՈߣI"?zrxkzI6Q-&v4Xx&XHVR# YXD莐dB oBB;[XQQ 7$t4Az+$1 @.0Q:$"! H(%Þ Z,2HE/ Jeb0YےP'!V%%̘"uV3|`^T`mY#/gJْ`> axm`Zbl0eMs+PɛYs =cU% 0QGE:ǼDY,cvظߢ؟jSv%0(XْyoIh…h1Pz0 A?X+˳qҜ| 0U23~ ±wZ*W R lEjE%gri.; ^Hü!F$.W_LSfO/6$0]i P]c-;.$^zsFyeo"_j֏A[/zGL=Vw/OuhZ>x2Vc#b56>5Z1VX6AM$ɠfO1s=Lj5`RjfKI-HŲ^ɘ"C$ANIH4}60g׬E3]x+lBz&{WحHWԆ\$ݚ}XSe^erYp{ [\iMGAʽЭKxnCm y(HdDw @ktH^[EXU1,}DQz\**F_t}:tnD׵.A;A[ e,ҟ%> 瑠w(o3\eNW(ZHd45*)u6{M> N_CJӓ!['YH4Bj~_JxZd1(LNT#1 ݆9 h2IDL8|\p%LJ{+zY}t6,]J?lª+^s6u!()!'==#/N(X*2{-C|&+BPpvɉ=1R,PZl$,γ杕ҲDL 2٢ݕ܉}< Ō۞Wz4a}􋕛x𪐗F"\@-AgkD1d5BI ){_Ro4m"H]Xo~`r|z刻6JH~ڨ9+4^.>8dAg/a4Ͼ\YӬ`ڼn1V?O~it>9}j};,9Q9~8%m66߂G4zBʑP!_X/o?ׯ7oK޿W[0&(쯚r[O[M7C͇^[͇f`h]zF 5%[B,Ɲ/Ln+@i_ǩ^"iO~ӥdxXWUף0v5:li9HgpkkJ> *8|蝤mL+?; rw0es>eΊKي@dHD){?Pgv$]v|.k1L7>~Q◴t`佅na0bxqm \gM䪈uq3)Ʀ(R ȍ  `mwlلJAlNwEѣѳx `ooCZNΏr# 4rHUkmW9V}ի>NV:65͑0\tV-qU9Խ`%(6}Q/+7֚R=zHr&U Z_DMC.MͻJN 8yiWOVue(Vzu tOGGgO t-i=Ҳ;GZ!_yt|2jSpK`clb%=4!#nÅCԻ}w(vOW43HWH$B򅢊h RtH1cU`P*B6%TuA* ]̬K [LPQp@eap!;>l:ŐӳalNφgvbo5BWsL܄Rj̳p,7"%'_ JM4But Vh!,AX+V3!V G^/rV) DU;|#zr|vr2)"Ϟ|xA,xbS§h5~EY>gid׽gy%?x-+Yא۠PNFF_Q=s=s1;&\GbɕP"׭ɚ0uwZ"E9ᢓ(Gs 1zHR1 %PB`LllTct69xERhX+>EYZd<k+ʴ]yAWڛNr'#i@h}׵׏nT*[y[7ܺ]mL[9-nq=5e-n:~e;ծؠ祖:?]sMx=J-dh-|k'U`M7',?e?)t1wjU_z|s }XGG&3"PmF/n|)ȇ$w` \sz#K${29bŲdeXM5d=UUhg#ؠYrF΂IssL4@pA:xǒ-*:Iȡ!ޢz2k?M+HwHM)؄ TŃMRw- Is<ԚPC MPˈz3x$A I'»Li_e(SD͐ @@QHf `&! xbNzc*ÕQ(Iț&z#: Z-pvEѰ_NT[JRGz&C.%UOJz D"s֨j RN)&*E*O7q/EԛAkJFӠHRQBLY{t uZDHPy,ch 1uhU6M"Peqg^S`\ hO ia zPD@Ɛ%#MZAQw@rR8ך&a062'^7/n`[BA;܁J#7XV/&~\[|TחAH/'yY"}[jZG OL3ubˀJ _E#B7zJ=> RxF n #b:%* *qCh)t/qvP6ÛY]M`_~f-;Of|Va;Y bքTv%qf8.=VxKG}X%8/%OZ2s=@P&H9I<Ij1QdN„4I%R]ؖD:*(,/ 2O}Q 5Pr"C.pgkLx=i)Db#p$/?qb^jBHΑZ)CZcƠSbTrKP]{NquO]ܳ \nYdY^3ٷ,.$LP =AD02WX۾wzaIHeӂl4{ 9,2%Za>BxQY[ab:"M F$AKH.8!(Q{t6kA&%0b!C1vP"/>ضw]3[9^j8TW|R:~4\40yn|-"%Oj[5qj?2zfyE2Gz4&88;z.:Iak%jԆr .nur(q<@r@_.$/ fֿټ7Sx˓|TMjoC'(ԘfAQW>+4• ?/?3g\j.Ip=8@U0>kꎴ+RЮ1Ev.{IB^kuuQrx9SNW%w?P~EhU6ծnt_»7 3^jP&.{o>1h]17wip])zeG׫7^詇-.^ *x(k%GZu*ǻ %_ιL"5qdDT3bM3U ۧ+zry{Z%&=yr͈ӛx"MXv&(As8sIP{0nc3/ExO+O"pEI[4ջ|cicg.J._ZΧ\h\fYKPgW~Yen:ܜ$nĥ%잁\8%Τ Qr]slίJ& ǫ9f<ܣ"2߮%we~ zܴ<^St<\6Cdv|\Ywm3bk7Fz0;ٔԃLjis ?`m 1=otx 8)i5x66[-̬[,B-?CqF'OoLP ?ys{ˈdcd=9ޓ.`4!fyH=D -4o[wdY ۄ;cߘDϲ+eط_t0[n>NG'SlJqxA킴uI'[)~C|}ϔP`hu,O1 Zk׎3[|]c_2CHG{?nuۓD@L$B)!%ov YlRɂ lGߊ7ԉ[y$s[]#nճ!;rTvd üTηm㫧RwQ &两BTs'' R B,q(1e SiQDZ(<>G0|DX=gbA`G}JR'L.E}`r,*a' ! &+Q;5R`ɣ,'YNs5,)e) yCb4%$ NECR$@HPQ ,QCw4)@83YؤNh{oʦ7/'.{q^|nW,Qg䣣(:BixD"wW¯F܄@5"sOD1Ӧbc?.]Mm8aҋb}tyZcdFA,ˈ\P3ENDRFq,'wږ #y.4JQ#e 'V;$QȄQh]P`H,p̻ J<>#MkCr5򼸼8ڿ.2_i}֋Û-r7Fm(K I 0{Q |$Dv;9S(+RBh crnR|Es56.D$.m^ɃL5/5 I(UhEF{z{(Byk ܓ)a KPX"ڂ%!%I~  5x21Ye;Ti 12|q}11(Fs{ {$b4cA; !GȃN IyL4Yalc+0#)?De]TO٫eu2>~}QGVYCox!9N =_~kFKq1WYtpQ!s7hWSj LkadS ]aWknpF5nc G('lD.tf'x_&STWgoſZh*yDDg#P7WS% _pz-jz_yyj<U+i.|{=8*1? `oA.?斛۹]qUPۏo4bsOWmݰn8Uݬ OF? g_=w՛ny8ܝ^k۳Zb$Ȉ Ͽ>za8gts{7TL2Yo0\#P}w>gZ8Be~T Kl%w~ʌ)R!)W=3(ؔ(yĒ8͞UWuw~;7/?|;ο ̛bH`` !|+WFj%GNxJVr[%| !/@>v89zIN7MɃ 0hʢgvQST@%U)"f#* eH 5>M--}OH8%9609@K&F(3dњB(,>mXȼ?O|s)"-!H8.@r*D&Ag qVq|0ZRȥI+Ux\u!W;'(oZZtv`ڠWKd2Nj C{>)2}B hj>H`P޹**$TF ̊hX;/-]fή:|s|as-W]Gn)M*ҿj}5}˙߮F.h}.PXk/I~w}W߶60|oQ_2:5咹/§ѱ.jEkjO6z+έWGBiLNf!>ե suIͱuUzw8ٞNmݳYw-+lYڽ]E7wjz^iѻ݇Y7chau[b|OǭͶV<8F#nmɚ/vRf-taͲgt(V,Lq$6T,mpԤA`fM>/Wqvwc_ ~$l湘V/Cd(.< 4G|8rHy~="b\[˯CA(i}ϐ˫QG75Od/i:Lj+g/nPwZ!+L@p(h[* !Zr\{<*HCs" ?S P j: #Y e=rK>H䤎G>30`(UH+U^ʩ/T71YR!Y$/VZu@'xTcFFAJK8Iy. INQ4\yՎp%UT{1 i:!FaB@Po偁7 T4! R-$+$RҎAQ$Yg|K Q)C%cTZJ9Y1r:r`T||*Y--!4\)q("^ %/c)I D B\#)fI$]G$:0-FAoGri3cAxeouUbz;֒djjܤ|R: Dv7U,#$JG?KLQ z' %,GqRq3I]XՀlU耸Hr,:VrTU 2G}- GI'{r"q1{N|X[xuG1zbqa3t4KY>ipܼҞo8/e_ U#?ݷunOmePRw'mvr煜*,f6TH_2T2%HN'K^zu'T{v4DPBvQ`E BE u(u"`ʦ# xbNzc*B$&z$3-l N'W'-ړ_SA+.ԑ'n (GKI'1}R҃9ɚH`N#TܢH9ܫ>pBc r3h-QhI3*U)bޠՂHJ;O5j`c*fNq1ʹJ&b ~e ?$UU<-^[+PQc%#MZAQT@q.5M a,|S7־;*#oж%7ZsTV>)N:!7JCY|}2Cz ,zJ=>Z%&2` 4 #b:%* * 4Z~eJ}qϠ_ކ7ku[9^1s|3T ӶZI߫z &SNF[_Aqze=n&Y#Mrtsbq99;. _o9M.pC[OpBo>f+\5uF^J*幯 o8*,&g=ij)3GW` OkոCUH{%+&i8V29_1jИvT'oԖY.U W4R};E쭝[)C&RkQ@lyީf Bބ3ӠMh/:c`Iɜ gQZR-ѕu@AP:*(Pɾh-D<F1.@)$;d1r:iO3[tj)j{jCP)W.ĥr.9U :$ezDV".d<QP8pROt,*V)>R#U2=k 8l :%*I% TiXA;IR3+ eaY(zYxTYx*l6rv=l7 *;M>3%)Ac6l`;8K%8 A1TBKD02WMv ^<'’-:-(9j4<2 d]zZ0':s!@۞]O<="z9IPK"4 T&(ř@(q&eGb%X :@;[ iN58C=̿Por-vw"|Z/1-|Z=0{`>0,[pՃA3aWooL[XUb[>Vn?0,ۖqmetgBڼr$֎{I m͝։m[/]ڟ<[Bu3:{03C.|dc{Evzl<6.Ï-6m#z>.\LPj?{ZvmB7=fY&K^xb;n_Xwqg)*l֙I}WjgnۍgEぷIѯ'ᑩ@S!b(g-(]L%JgZɍ_mb%xuسޘ;HHc6oHJԣHJEJ]$@/%GeJ̪ߙswsL(YΧivuq0HjơD"QOS6:#w,rua?'悱ӹPrP>e2Xm&H͙0_r {dYqd~SzVfgӼAJf-$W)ɅœwJ3Ƭ6`T`S}Ns-PioUbPBxjR݊rbV7K5MRSrq\!9H38.K.P2 hQȒ^ .gIU`ood,uXV >nL&uŮz;R97l8j02j1Ls~\re:<#U?8wn]Mƹv76Qfg_ʟg[QϧGAm"JrF?)gOVψ/*r&(y.HJ0//ɥ hNIԼtڨ0͇6M;Cb$7IcлTi!\`Yfy|vOlg;3qOǤ;qzmM?54k.H)Cy%#YCTۂtW}ԉ掠Gc.1y!_ Ρf%F4GS1 lɂ!vo n] ~..%D{ˆ?&~L-s~n4% ǘ\}:>\n4 [ *$AKz'I;DFоRsotLQ SD4X/w3\> c6i";KrHK$ 4jBa=ꔣʱ{>x.P 2Sh1/dbQeIFd,) lKR(OȲRd\3d "u1 ц\pQlIC74-j s6林/=|zIb=/O>z'_N1+x$"wmC)A 8ZU'jYdsTLaS1qpTLg}+]rc$<}^xtG;`{KHyRK0> c ( "u3TV(cO P6qMP*4gt 2K֝-w(xE^\z2kK!.Vnbͪr]bܱ}rEux,&ȂH^\dJ3`J) I`F22MfB~Qj|E&DkȻ,5F2J CC,Cg <!(07۩ѹr }|8ِco`U6&3ǵYK"FF9%Ib ^"OM/55m$A&c(3{LIq.܅efbTAl@Q|THO 16;uB ,4ɉi *0'|pT;^ 1QC>,Q?U" -\7ݣNc/[%ue̻apHOS,ֳJbTG*Sj8*dqpǐ,k_Sp-eac\2qtNttFߏ3RGiŵ_Y@R#GRL$k4nxҮOOl}1"NS>iKOwZ-xFt;Qܪ5ޭ8)3> *&|ܺExYbR?W~m/|5;[|~Bug8gflWY{=OО}% -zfDg3w5*,ƓV??*d<Ml͑u٪`[]dW]JjYVe##-1RV$>\R4 &yP1?|U9Ćye /H"utK_ˇ.??4D#0Kĸꂰ?ko۾̦e$[4]/|vyއǦnmar[Iqշ4ߗ۝t3kCDs_n[KEmzW%u9*Ċe#fzGAՆkZn=&Zmkz}tM<.T ?HoH4$%v,0F-T 2oUb9,eOAeo$jR5-|8ic23(虋KP$!24,82U`Dgm7ulT.j2\6Be#8=gkབ;ʝ^Kwn80=z/9-CaS,1./F5;b\.#urw?ߎGƀ@\Wy(Ul>-e?ײz5gQB~EA1*ɟbbe4D6EDž g*zR{73'\(Ie+<&#x.Wv֝ME)nCp&'OtY%[5A'npܻCf7P)R0-fL2,iDB1Xe !" Uvd+ R$/R*>|p̵+TmX;5z]XM2 ue]z]S]x+ 쪬_?ܥ-7OO"Ã?/hu4޸`V"ٻD|5{ͯzIXs3 !H6\ h!餹HfSB Xj`cGZiDګF@Q1ثyeioJqVr*ͫ;+MWr[39@v"k`|ia.&ۻ>J<ϥ&@٠qԌ&E!غ >6o[|u9E\p0&2_K49Aޟw$j FT%ę3=L"ӑY3sռ鐲 EڀGK2F*2zsgg1!f)M58%^6:h;Bzݻ^ =A dLV|n8X߀MY&+VKiwt.ఝ΅>87)}>Q뙉Tǁ%IDM)ҋd>#i3dF( B8Z4)R)ȠnD3Q[cx}I I10"HvZ"VCJ˥eZu4w\;_>\Q$-hY ʢ)Y%T"JzoFzVCW!ؼ/3b1o(lh0aRseHlr&\x`m+G.vnEVԽ 0 lOXbeɑKVdڒH[x,͵+΂8m Đ 0Nk.y%ҕW.űn{0κugK{Cm9$\#yg 1ʭ ֺVe! + V=!GOwV!GWk*58z_ H*HfPnT+]8iIuT眩Z1w/Sm>ŠwJ%iUHцXLV|g+p]Amu7yQVj*rLj5MUɨRb"ibxFZ^쯨0oEDL-[Dnᔹe";o)8gr"-KTMh[<_71bMA(V٠&OPHG;՝-~KAR)I!Ռx 2d!;U@98T{"dS|BiB1Ti}./ˤl**㘀kHD-Sk\S*$c z]rarg\ c}>K11Aiaq"xA`eLv Ny1b 'ff`0vQR;yok?ㅾulmX ˁ{Z4U./>ߛ&};}؁JhħO<*wP]$v|у6|e)Yyk5jbE4ZSdR ǜ`s:u7+-d%>'ձ6ݲDѵsZr}e>5E7_oeʍfXB ΃"F~Ht5@)Axw~JrS^MPO \b"v7o@ӏ,7*_l3©K^Vw0y;if3aPq!IUmR*oD5F5Q}&lH"0"I"J@`Alp&92I`bߪ01R7U%:+X-"%D̜ɏvGS{nݹBĐv 6^Z'AKOG= 'c5'&∩y4_,ȸ@N>0O(s )Z$b3 k]ۍX<VµZ]+ƣe p*骴`x 1$>  3@R3{%_šG)zlrTfO\.:DYPl~{n1։nN!-kļeTΎݢ&$Sm%wrhO{fG+^HG7-p8-Lo+;&hci{oxU鋿C&iiZdUyw DNe6 l/>ZETFF &Rfs$fX0CFO=zE6 Iu!LːyLJ:*0F}";Iscw͵R}5&k\2i,j7 K!ayqB~}a(&[ !, NڢJ`]4iWXzh Sa/L!_Et.[loV9~jKmTCeH Q/cH QxH  !2$\ːp.CeH ' !2$\ːph.CeH !2$\H&hM~H !2$\ːpCeH !/1$\ːp.kCeH ˁDW4[n, _=871EĚtT JH S&HOUKTPGWxTXyHX*.߯ΰvQWur9K_yׅ@,7m(}va|';{ܭw|]<˻RB)fő#9= U]i{UJ7 *rJP7`Q؊7]=a#O!kebF7iKϮJLiTV*՝=v'qM̷hp%dl97ݛBi^.Vק7Ӟw6}.~k?~9nn7B YV@{9Id%\’537m yϚ\7zbjВ?a}i.Ú90-wf9ȼz t ^XO E~bLmĴƤzL;m*S5'J6Nm0htp<)><]mo:-k;L[<^)"OeDgOpa _?r-4;7<ޕ. ɇ3 KԻeϚZ߼I0i[VL*1M/ZNNj/wN_2YN`'}ʾ#Cɺ_7EƎぷIџ^]nʈ %h.X*"7%gW,Ϟ]bBq78%HWtHz=Rjo7n;>\|gz_*)'V',L0YPɣ)T~_Id)~/z䃯lltj2jĊ hʵ*X#9rP`u&0 օOh Bp#|휿.լ6~\j|+Xy0W(9d!F~h0%bo.ۏ] TW3O^6y噟y *p͞t%Lq' lڽIadzO=T% b4;G+[F(HO;E̝µ}:DY"o*co)v`uTC텢 ܪP fpBԆr1cb0XOZlMLugs|HDF/߾[r빹]RfrL}:;MgZl I`*|uO|Zݣ/լ_~b"\u>MDǗ5.PA,TMpEy8zر KhX 꿺b):keC& VoБr{HtH džD Qd`]g +E)(rG@I(!yІ $FFUA e(-R*sbEPM+uT,6_a@,}/sFvHt ='Ht8:t~{ O6;=ycUP*m@\)D.5n%jUgnk}Zߝ{е>!{2tɣiu?gD:‚;Yh*Y1O RqQ$&c IHL^B}\$0d.RluB9Å8^czcו?#+adVp~:Įt>hWm;&rH ZI_Kl"R|N <$3clP1X)i9!f$kxz1A]٫SD$_H f,'O:o 4"#j!=|}>UR~{''SVʺ3Ľ%;(zR\}2,_>9$ j\Bt&%'(]c]I d ucǜs9]\Fb&F4tTI;oT ԯ"e?zy-^H9YbIZ) eD:`Npw$bS#jVlUtiM$i"yWߚvT yGhah?N(M~ȃE>5ZOIu`r \ O< $$kk >nǸd7ZJhӁNY8<`kז?fDh3LZnt bJP*w+cSN}ltܷF}m/g]L.>j$ Ɲ$ĝ650EWS_/'N+U&ȼ\bįm7WׁX欈9󟆣r:|2tW]K FՈ޿{G/,uDjHg]Èa$7ì [(mʣqkǣ%Wcφ[&'E>r]vfYըӚK+_ǓX}1̓qTq`oXwC^.Ҵ+?uk΋s] OW@6jO\<7WTųԚ+Յʎtv]9JܗֹM]G_y_q 'lN&1oUfsp)K .\<)+}نkZ9מy<ô9%m>$JsU9Kfv:dXpTd@Tyˢγj|כLܱ6wtlYy{+ʝUL7&]k6}/Iه+NC~tm\M)DbDMO怸x )"$K3h*;Absߡe.}qN9U{=y`tyloO.E0 q ,bGߟ^JUhf|vފ-tYw'l^fї}Ǎϓ~kኪ֚Fxp;B }@y3o6މc՜W=漜{c&p%HDes敋G| F! Z@ H&i# CP.ې`Kd2NЅ 'td\ca$؜OƟ.Sw{Lv;-6qz1;߳ٺV¹޺pK+N:h Kdd 5LU fjpt\`w:lhɿ vgвȿ@WJöܮ.Z.dt+@֯)d1d@d_M6=5Q׺L3Ta?9q䋣y4Eȉڵ4|PΛ om2^[Cps&r ׮ZpW:8v5f0O:8o/;h:KG`s/v1o1({sJ A:<@Ɋzn]]yw5-Gxqˇ;(Yed%MCLY\ɱ֡gs3g2h/$'CXcF'Ol=D^3&gPU`92Cp(sQXJ.K ީM Bo?Bo4\;G& G~Rtx<]W]+q%"uk̚X-^x=>:Y36's_C]N@RԸv[gKӿaqg;ٓnEÓ]jy˫漋mk qϿ,ԍY߲}%/ou|)֋ry,Bmga0 ;mu-rgY)\OD2Wep{|PN1xq<'LuE Hor38 k/T[ή6`rLgz\athFNj FJџXO |Sɼ ;w3h\{$3zfl)80wnh5,)66(: d^Jli2yœ\"yԪh빎K%kv (7=r|ݠvỷpŏ?-%E<2xQ%MWWR$c)I\9'Zktm Ɖq%K.rd>162>$bĩZ/ڐ\ I*63n cJdra ޹p%c ,onWSĜ=!Zr3aIImȍԂV:'K9.FT]TR1!)q %QT\`G(:st%!>di I+(3^c)@1ځ72DQJ @t)RdٺLdt$lha&G˹>r2OrW !qk U4 $$I$H^ ^, 2-$mm/ZOXS+ƓT#RZƝqRClA3嘣;3?g>SfXJZsc1i_u&y!i;ˆL/y:! Ob=`A@7,x">f؜:@h|7AiQn,;+.%KՌY!$ja"lX9!W;R6”А{IgJf%wV4K7w mp2)-[.)ՒjF̵a]ͺ"y!y[TtћTZ/1TY W\j`05hzS$BDNK WD2+U f:lU< zz KX+7dlh3>,XNe~d}2ylEvhwe%,4 6bTe9*4>}wx4w]<3ڶ!X;XNzV y|tɭN?bPnDlt \×u-xc%Ȣ x  B}R'XVvf nރ* gbDŽf<Ђz5xj6ݾ`UmӃbFt=E`dAyA+:[P:I 8sl#t$fWEV'J2!~ƀUuuI=yvPֶc@s C DtIl()`{utКEui<`J , yѠ6GT.%1ae )An$LL:0,Kd+y*!0VtIԆ7?X 9xy:"87bW-s z5x] D-颅c4ɘEѩb0ؾ=A#uF(ZĎs-0j;ypXC̣oC; |P.ျS^"fg ɻ"pp9}\D[9hD>ktwz0:  T",1 ' H@'X 7l֫dXDQ!|o/yE R޹TO.^"п[?$ok#Zyke­PQiQpRdEUAYuC2UhÂI=:[19YsSH@jpAP˜c˫Nr D2Tc ƀ?a[_FraڒhS { _tJ.wU\U!5,n3ưdо Ńa n4SbF: x/KFq|P9/`H f}!JUnu5Dt060+&1"01D KPq)z,Fy+hFzRk,FLk{:˭^u#'eًXN4,4 cht(B[gHu"Fc`hMmtiYtyyr}7j"i,[Yw2vF^Kt \'^Ksr]:Zd`5KG̺;.k&b6MiTp ! + *K$r !,@HFD0Om-,YA.nuµRLa5S;Bt,:0B77FįJhgMM_^h>턄_@hf8s. [oKbs؜$6'9IlNbs؜$6'9IlNbs؜$6'9IlNbs؜$6'9IlNbs؜$6'9IlNbs؜$6'+6`"y عW#6GyVڗ.6QjEboQl.rp*? %cwoYOפ׏pó#cB9/3:qoӐת`RzCŧ5hq'6ۂrL+\,R`y2•Dmes测G} XXa͓x,m籐έ(seF~ƭso񋗛%"X+ew{m QyL.YeX;D ]1dXu1BR4_w,d23Vk!W؃ g}r:Egp|}swLNdŜ(2I^yZ w~@`hy `t'bG.mx/+ŧ*4tzkKS|eVe2S2 ^bޕZw74~]s+_t-C;;B'Jw_chpuW7)t}~ q_]Ͳ^S _J޺r{孾FyCT/XZgΠ8F[X #2ggu/jtr )GZue7F)y?5;RAWYM-:y3/?N$M~ȣ{j=$u=s-YZIލn:U_]FՔDΣ<%G*ћl٧Q>s՟a+Y-};}d,5勳p'8&'k ʓy sn]ȣEwHwE~ey(xFu;xU;đLͣ]_'N> &}7L˫G/.?x *u0xR'nS[Dt׫yw$[4uUw;NJymDm$n,?ӦC>r}vw8Snׁ|7&YOogyP]7Ysa5rr=hn~XFqYDoGeۼx!Bl=n9dև&$9'IUaV{g/:H ! >H>3 ,T@Ylo꺮*QiלdLKV&sUkfL@jD%Lu`ކzՠpFYȶ;J)x%2t)\"-k,pqVo WMûTNgJ4>.}Tb=a߿mw@'LPi͔+?,`˞Yp3yR%Vv+;+?s@C+>O:E̚r 1/T4&F_^ HYKs%G8x xer=[Fo5R3_bUM4|6ǙH7tc[䗳aNTyEcOٕv.̂3<YA3 EEx'[ z }!#26i8K9-lBn j oUX7iE0_j~fgwnxr\$+刯 o w TQd>5[qbB~IrAfdKljLIt]M7Tx4u.V'*&.ϫA\ak|d/_W0bj1,ާɜHdYlzs=ŒI64 †}!.B(M\Gn/5%3^0og6 >ߦ2.Wv0Yx6JˆOU^6}-ΥSk|c:n:nFa)Ʌ'$wt%c@*H$aPSY| -pS' Dл5]ӒxX|_)% -je*%wtji?Oli.W#dv`v`w<\X/ GaLt4^-:>%7Mi4<- 5l4۶qӛF/AoE-Cdy5ށmqL\밉XhEb._#fIZ"l-G$L';:5/6V.O=("޸]N bVҎfxN,~ 7?htc,if>Ҧ۪QjjP{ބoJ1nNW}. W&ko?mZ7M4y3ҼCÇwY++ʬ]֯'vkɬ]un왵IJ'zfYfQDQwRSbkm}'E h=C# +Jf1I Gw4J:As,ZunF˴[~Uid4<}rΕw>JfU_r&۷s ˋ׭|ꤸSURNg5Q9AY$t(`"OU̎f}@l -pT<⭐$Fgd*DPd!BWDJTL)SD#-Mqy`$H Eb-S`,(A(m9uY3q W}0=5<"*  2B 4^XsKY "lŁL-!{NXu/ZuG gU’w9 AqΣSNQx5~R;Pbb16&uw4Uj#4YL4#1hiCyYx G1f:q`!=?̀ʷMT3!@ #CTk HZ"Ҏ/snEdȬ@tL+9||Й=rѥR |κ\@KR łgZ5PqMFUdREЏ:F??LMتa@a0ۀx@I_`M>^Mg@1,>4Й_u I_)\P~])gA(աx\(5*]2JXY23 YVzD5Ȯ6|8GY,  *8tU~`}BpB6Gm.x6zը"Hā^r+Ed[+@"p=УlJ"b"Jzu=LV抦zvWJe9 )y44j[o:eRd(񄭗}[g#e]^-&z&ϗ_td`@ːl*C 1nOWV^H:KuѭllF+[F2*AR]\9(u:hA%e Z.UۥWtIX6,B*:v"r6]HބȭS礡4^i&>%VvyD}1ֳ{-i*p_Q7XpjUqAoUZ cVg{@8TmC|}Z?ϝB .,xg,L . J~59t|ȍ>gKHi--A(R);9V" )*9x| 16$MN6qQ7(Yްj6yq)2\/o.凿Tۚ)%E`8C!֌.4(T)yY\onD6Y dmr%T޴1cpA^}9D1XͿF^M/,s zIFyN3i-0k)ƹB,QEɣ8E`[JkbL]+4&t`>.=MmO^m'yxD/@} LҀ*#漵 ia![ςFESL6 dgA}mkCkm)Zfi2P[Q. +P@%u,`cHq@%sIs ET&e YP3q;7}D+Z6|a޲w@u_MFSr"JirvR?N99Tqw2}c `T JH&DFGh9?$4A/\?&(]=U+Q`&ˣ4 IKW@;_u6"Ŭ DR/Vxz}u7Szע??f1ֶbGjʥo}Gz.=KJp%'ʷ^M*Ucg$%}uك(-5{^g5{^x5;z^x5{^x5k<Ͻsko*JFr  *8V$yUe0tdIG6<:`zn4m!Hrp%E-mIIwE6b=b"1MP )K$EDN-bí.ǿG#&SZ%* 9'U&֕.O>>d#:W{Uopyb[Cg}u3|웋iQ >UtFnjA;I>x1q]6!f6L7P4NZ?vl=6O7v?MYl9m$M]f>< 4m~ih2lcWhL3o={v<~ţNYYg"J}튾37CS)rQk@PԨ=qx:-d8)`5<2r-ѐK0ml.&uփU&Cc+ʖW}2s`rsL c]lmKR= MG/~ESȾߺ^r9lhj?5\ɼ8uf{}!);uf?z~EՋIAՋ[;zwp*xx=GONod$pø?O{Q+<-39L^:;ޔșG!;ˉN=BÌ2xQ hkIqWxs{2iIݧ- Siw?&_P3y똦Y5ۙ<3a4/Λ~LTۡ#${lZ{hn:?j nsRoO(/oOϥ!j,!kV_?B1"UH]ArJR/H}J,]KG{{{dWl HOq*42RI kh,,fme0W2Z鵠iG̅e|wUU<_vqI2tN$ G񯢌hIe";U6E7M1``Yl|Z]s߶c#ojb;Qۊ](eDj1" J+%*$-MUe(>Bڡh)iń((%Ϲ$' J$NZ IZQ6 BTΉv4RSL1.5.Ey<҂F0tRlVc;f_0xEoovczRr%ܗ~OKY]0bZz2 NPQM>L֧r|RrLJJA x0<_P WVDTQ"*,*NYyH=Dr&\+m)T*r2s"%7a0 6DR \kJ=!4q0wItOZ#U!2+]Ӱ9[+;1H/G4İdsZE՘iiZ79՗a|=_vͱEB1:WUJ8KQe{qXA/J$_V‚7bZ ,qU$O'=q`Z4υsalhؓzPa L0NTAU+*wa3]ٚ7\8$!\ޅM*$ <#:j>1e i䩴WF1rFëRqxƣoM@x&آ^Y ȠT]! &V+Q95B`ɋ⇭3Gr!BSR$i8 : K$(%HHA$aI(IĆ"( D ]-IcQ4 Q ύ>_bqY;I,bbjƴ'*z:֙\.OU"S+űJd*5?aTqx 2|@sj> ZnBe LJkUeDuȎNsdly*U쇺$ |4s`f[ ֣ݔzFA,ˈ\PF~JJNW1'qږ %y.4JQ#e 'ƃjzv!2aj".(CS0$@a9[#ps:ϔt}[k}õś. )[3O6|D_M{Ԇ Q |$Dv;9SW Wţ1]}=!$:^|vlV>t!Aij7'MB7ZQ#EF(/2(Վcr,G ݜy`*ͬ,B%"qkX3 (gIDAI^QO92LL ?vkziFb2{bcbP(@I)CҌio,#6d?1h $!D-XDt@ `k X-EL} Y9?dF^=;{o{8yIbƮ2en^vǻΕbgj~ϭ .δFvm&V N}gE̦Q6bըd;:PN 9Ts3;k׻;Q^ŞfjI\U#wZOHt6VȨq=Nr;:_L>|1!\*w[ E^~[t>yz⸺ #\o ;0~} 8Qyl Ҙ;DMuLo_M/Yf9gN̵^j2킑B|r{ iib^=]7uT5vX!ӆ g0blG}v CWF6:dSMcE,ļQ>w3 \> }~(v9=`}RIaK%N^Onw| _P]槷?^_?^ex58GY LM{!^{vͺ%EJz:[+Y!7-BlϷܚ8~;9, ~}PwIvOqO#H6\;+jJUN7_Zh>Čr] Axd9]=&6Q?ZF)CO L$S3E"A Cu jITKm26B $&IVIznkLU5|=FڑQDZC:pj]`(T#'*8;Ԋ::EuN5ca;nlDrer j-nA f]{)<NjLa'rbN%%SͱGd*n#X~e ~՗`Xp$JU` )W06T}jڅ4tjV)|P)ܼZqEG{鈋mH1R2Yi][\2w?tz&+f>~XneuU D PG#-j3k4VIL9`^PJ[Xa< 띳bqE50ǃ@$SdB=U2 ,Z#ՁHOi謯Prf4%a׃}8(nPդBMLcEjD$P `PYJ$ZxP,{CEBe][o;+Fv0',VwOH8a救C Ϸ|wMces[i][&^.ww7./ZKmϻ1^Ա}fJc󆱹U:mԡӁfQO4UO$Vq*{:|lO~Υ 5GW\IWZEcjWSV(T'cR->uMΘm'cƓ? s/'gs^|F ' ?yW%Z[F LWkY%[)WvW`\x,pUrpVWHc `}D쪚kݱ[pU$=•B&˳itY:K~$EL8PRpRXLT3-1 Q%u ]̱,! E%@0|sM//~מ_~)}ZR`.hMi4킦]д vո.h5i4˿ vA.hMi4킦]д vA.hMi4킦]д vA.hMi4k["+@Qn5Rw%uǞ[\I0'uW]]Kmu宿{Yd4< vAβi4킦i5fZNjei4] vA.hMi4킦]д 򹑥Vvi{#.hMi4킦]д vA.hM{lڱ9Gd邑^Z~^N/Uܦ(@!hQO:Jy{|>\A&-9Lυɦhsa9d'J"$DV4CU|DsE`qBE&BgE`4,q|̜9#pȳJ60P_^^ߜvz T)5G@.:_ 92jDO(7 dv>]}I@$"S{a,9E2 5rLz͈Li6:Bl̜k15@MrVu YmnOjxtFKZ{QD}b ()`\q$hH hVU{fRGy53(Hb!]ʪ0AU"W %vTL)S$Җ~Eh$JB`0! I'm6 1ʢ, [lj2 ﰱ53gG?_`*=^HBt /2B |iJY]6Q - -]歬P^KI; +6I[ot9nj :O*o|}eSZͨV>qRkbb16x#4C5=/Gi1hmC3$C&1iп2]idEQ`"~io$9^x|-Ҏ_~Xݽ%wDض*/njt,l<-3Gvps.b@t傒C:'+KpJL*A Jthd춲j l<;TcP&d.һp !So@fF$j2!(7U&;!a157JI*"yL> Ĭ^‚Q.<+a))8)) A2>VY f%L7_dG !Fɵ%K@DaȠLC< DadP lm9Ihc9yR?,X1}覆  Af '\eSPJrm,FGV: PJ@( S)FS2hKaV7 )B27S>d#_C}l1 l.[q-9 +)'sE%0#Bѯij0鏋8.n D(5,Nߴg7s|0U"$` Ek`$˛<+{۾awR~m\?~3<@d)VlbzyJg|/ @T")7/G!5   n(H2l{Ö7|:zѳ6t zg=ULLB'LDA8$ӥɒ9 t`I]﬍+ HmqV~Ou?>/$}cX[UyvND0sS|Ÿ:Ηt(W_s &b(owKrvi hJ;5^_>G|NǑy לEю/2zxq)#jb.,ӵ+̐4\V]ۙH<=dIUX`b~c$*@|°ɍu73gm0͓cbnNL#NB!{ G;>*l*z{W"K>ĤTjGbp@(lnښR AڲU'G2 ɖBAP TkdlfsYaZbnc^z_xM!b?q<ɟg7r4C+}"T!]  P)F!kb)bEkDYadKp'LIr.1m/9;\^r,luuG`7΄WN rmeDDkR"g #I˜ ݶQFQR #df@EG셒_+mLjYxHEjeպ69p׬YGc["i#bGč+^;a:(e)$R/ ?Fm >2~Ƅrnw{oOZӳ#On?7e+n81S^lbX}Ľv2z됋 hGrJ#Rmހ얘 Gh:εZc8솴iX5B-9W2 2s||fǿC\a0X@PANz)Dg?^$#{^F I8¢%B!HJ)-mHQKvLJ5oHI0"[n̷|wf}<|7kszPCJY}H(49"cGJ˂׀^dty8@eq45׻=[U]moH V4ofɰxz6p`SM<~o! QJ01j`H-X쩊؎،6 eK(ʨjЩx@jP%c0d9 ZT YUWm1Eu l6@@Ef *yyƛ՞vJ4ZK DoՒoo~*j \/t)89x#]gkH}3g$U2 +}9A@uibxe*KB?{FrdP~Gvv}rSCV߯zfH$>FH"1쮮_uuun2>$Ťb3&S{fBa%r$"A5Gp͙qu_1F068R$#/g)Ocb1Vٮ_늜 nHM #`(=r2I Nx5x7;z"kI1GS4Kٱ1M>tU&*~uLA:5憓yfӜ ~E5S sZD)3ln%9Y{c#8hΡYn8gqÚ1=b+$",0*Ị@3Cm/Q1&x9sԑaYqGþ\$JM$HcYݴB)mx}A/ ?hs@MGjO4Zɥ>.Ln/.E~0T wE)]+ܴ@\p464Yrm˃w4 we iVJlXiZA=p Ua{^B(kB.0 <5(XI픤u吰-'.'cW-c&wus>$v;mF/a7KH}/~ݽDӕ]*;؝֥)*J:[9Oh$wXܗwo&;[?`6]t%MJE|st=.]qA!0t3\֯<|o X3r4މ\.%;Q+š{kA}7[.= !OnD0}nFz)wa4nT2tX \\wD5GcJ2z*Qرzs+8^<:(}쬑B_W3Ζ2O9 MWAx^5,;84SJQMշ=oBH3{9ʉRܳ)])@DHX[?30_q0.73sL` Sk[znIN+ib0̪:]. li8[ P_Q7j:Tջ_^hgR*ww4En"~z+*GR- G\ K*uH.]BGr!RJ̎\ 1WG-J򃏓'*eoGsB%>8y"wm9Ps7Wߡct [*Ok$᯿i-]\`Or+0fTǭfy (3M\:r_ӈ9 ǒ𼭄vb|MdžK =#. T0[74#EB*s(C%Zxy-ܓ`wI#tUԆɭU,ׂHvA6,3w׽LMLGav( YKu0ֲ3rgY~f6{V-| ݋ k6C@ٿ!п!п!п!п!п!p\)+S;ɾ(wE}Q(w_/> ݾ@yψV}ba-Dzaυ'չ'5 "gڳ\)rpJP"p>[[Aj'dEQ}ȵrݖȾ eQjI-+ܔ21~2®rq)y⣾،3wѭ ^ ,a~8)9b?"b;-z*5`nUFw "rE؄ A2 c*R},A@`DX/5^#:D\}3z(V|!Ժ8v?~W}hw "O6,˵%l.םnӹ4&b).3eR"A#qIքu\!܍.SJ,D^G*?%* JpJb,A#E`/H&v X PUҭt<V6⏰OhƂIxIFĸ4)̳VؙQ w|Vp lM\e1iUw!ߚ#*gMK'a(YQUƧ|]S:. 򃗗gM61*Ysa8~-; ЦJZkGb|H!U 4 (ŸO㪤sjjperN6JQ4ꦹY|9,9ϟ^,8#թlKd]ٰ>pr0G^u^~Ogv?U=M$7 ۋC7ÞCxAD[ Yb\Aqo`v[n@~37_/&Pþծx/d_,k?ߤJ0ZwQլW}:-f`*epYvoފ{k0Ry!f=HGy K70%W)3JyP:ۚKhTzM AcG6//߇_]Xj4U '[S m~vC{<WP2Չ Nck>wirPS"`E+9"Zr5[}s{a)=pΙvwi"9]OsL:LoCKC Y:?M}9E6vy܇ ɥ!Zwŧb؀b{9i{` 3&s,Z v{ |K@7RȱZwK[> FsWT޿LR̽mQkiuWe^N/kN׺~jnD[Wօq%y[UsoŸȉsOw#91-inYz.w~4}?{qnދ+`l&,X^%'=R5p5 IdiaR/1",D`K= b FрG!eLD|i4!'.'xi#t>7M=vIѺ„ NaEA(N$LYK$$wNq )L`1ƌ̆`(3*h2:ZzQ9b0XJBԀREwV95x ts2Tn%܂+zdDȒsBu}Uw`dU89CZa KɢȱG=r*ZgrmvG.ܾd^ \Dn%m6vByu!m:I@v1xBZOlG6;*s vnBZ浴v{oj:Eύ!->Y͹hws\}uG5vYEYonM^k])U icb:PzlD\hBϯXδ (ݢb>W>rt*-Yά湵N(zEG@9`"#Qw2 >z2hCygV :-`k9Mu[^ػ^XW|"S)j>ilv3erM2̣JP*+\浯ؒlsؖrK|l6}+/KJ( 0aQHwY팎|t6[ m$漡s-x".oU5PRg(;OD[RC nnK5)$5rG칚}y LTbCѷB̹ bqx>[_δF=b*ߋ8'OM̱#SōǂBRcfjTkDfɒR:<{Ǭo=AzMNoV g2ΈQpoEj25{ρzhRUPQslӂ bv{Q(A)e5X1֎di月LwkV[CfM17dT֖[ o@3II6̄@}B~f] MAyIֈu} vC0 -T2FQ|˓<۵N4ٖ8bDPݨw /J^lو^()cU+x}\g,hTWrj zM뽶%ѯXP6Ckb_`1;"fJZ7E > G'Cz-KiCQȜ"XJ+ 5 )#5h5(`p<;ުtL󐇢#LFGx_v;eW"' a+6yb]/H1 YfY/E1bX˴vňaeΛbTH.?a_r8#Ok!BCdq{. :УKsͭe}~Kz[j I $3y2PsBQ@\-7E1OEA(֣1O?0]yT>rteaєf`~(Q T]ȂRs>52FN*٬Boz! R*%@E-H͎9nWo93 XX_3>2_vv %|ſ<`)n?]5O5hȫh]#!7DA+zҋZ.m͉Bu!腱{Rn]:Vy~\㎛׸c{tyhΕ#ဝ%">_wwͧywpx_p}/e=wZ,pfcp5gI !b(`(>UKC2san&@olٻcv̇;aĀ>;:eG}ٻR>.Ww.׃=ސy 27`n|>`Ȭ7ջg_ ;ֽaL}DžKzXvRzQ?Φg~qI70}8v{E5E[[ O:!G<<ˢ 9B/vEw; 3/M8,qaa9vg>9<7rOg_}<&(*|\Q=%EOfi=qT.݄/vx^AWiߙK-Èdc_e_+wٱ6P^mO6N&썾\| eh+9RXrREJE5;*2'
  • $^"08m1ZkWVƸ)bAEl2 ֐w|uBP@}]E^\*BYw_=:|sA-?G7?!F@DsJ Ľ(E{.>llb\YѤoz <{pgg?Swu!Њѫ'a1TA)-?HE/y؋7N!| 5w6#,uo:P?2Pm;ycb az~ߟ>_Xb3M+ 5Ckiݓe1ABTqb!Ur ^ۖ<^Ѓ2?chST8☓]ȨLsc –=z4A[.I- qoKQL4{[0'OI߀M ѷ0ͽg0:|q^~(~?I bGBBRcflj82@ԜˋQې; wLz6)pתVf)CYzP#(1`&U E%bl 409V=q+7Ui月=~0jQG /)(yrP h7x!U%6dT]-hFv!)ZO#kiHFsXc P.Z*`.vFQ|˓<;xib ݳ-qĈhѡ: BKŖhe_T8UM|<$8hTWrjmHǣFbI@\b_weCۼE_FyZ7E > !(d],saКC@Д40 -k!Etoqn,ͳC!ߧ\͸ң'{"1*N bg*6a=O/61q9ݑ'А Pd!q*4L%)+Q{PyӖJRp+IM^WJd<=8c%F7g #$BTv5Y y\5\)ywjb^L) ͜MA6)Y [ci*$эkc)s^ӄd*')F6ïy\c Վ(?^\y,[w%$ͳ}l)L^rcथwm$I@G%Xۃm4ffl#RҊ&5$ez0}#(YEБ"L?Pz\<.viGWGU"w~nȽv~]`{< #~|GU5G?ex{^#%@],Mі>; zI/*wA ϤEZ0!"z"2dRd2$7Ydb ت^d %!Dם3z+td\e8{yNd];)uK^ bz=颷Y9m >_o38C$AVK9etrD5):R{s(Pt{z`~A4GN#SA{w&(E`^<tK܃xKa +姄.#"B+/7+Ⱦ@PM0 )t t!(Zf/SbbnOr ;hk,mrS/gY#Lqq-jfRN_=B `8TJ\J3.R=Go y)l1skφˁ&z A&e'gHFDIy O>ޟ?*?_#ǻzCD 򌊱%1( wd'moլ%~;c>lv_Q82Kgڞe- ,BVK5@bKxoƍV N) гG,,%3-7&4ɲC){HWKCu>i9k8^ջyZZޖl 9[̔",&@--&̵ѻNF|8xwT嗅= ,"X}G-FVј.Xbs5Bn˅cq֝dp!j_/On9Mv<01Hɔ`L f *\ L ̽wi5Yiqw<I&T L<ن"瑜go\fS1HzЫzEc>o8LZѕ:lya'v=rJ=}>sh'?mL~{߅YD˴mlќ CSzItЋ 'ĜBɬ9){9z4c iC ,<")m #L_K^+kˉy+U_iM\9H3dq\>$48cGJɋ,K 덷.۷`,sϬyś ~ٶ]o7ot[׬dV\\c2&H6`zq3+i|;ae\Z x!1a#%}O~8}`5`O~~9M%+Z}#Χ,Ƴȁ/xdcQ MeϱwŽTdvȆ\ ۝bR?"*FT"<ղeL9-t\,9AY8dB+WuKvUt[t: Jޅ&Df vZr(gln~n/޾ۚJtt[pksF8y?ۓ]N0۷9^'1{[^hvB^rsgӁ/51FDrLҥL @M3qb 3P t>+=+Ltr\ZT/(|LŞQ0θ+ڸHT4pG#0'VRhR>~!~g'?M&\e[}t]RLEʾ2i׶w"fˤ{ 'V2$IV24?Xv2| ΊVGK>iڇ\>hkwe .bCy|7!et7=M?3} ~~c90lq䗶<:ʤn&䠶 gt?a0_ùLހƶsZ*d@8LVK oy <#}Bw:>Q6@b,W} r%jZ 1Eo2=ڜK#ҁ\qâLmH'Ƚ1.:du7 Ƀ^Zydu.=2)y %l)BV'7 6H[K=Glf<̈ؐOQ³Wuc˒c@Nz+[߷89v/3w/R]^ϊL?炓#uQQӍV6Yha Tt=%z %үEkS"j(Ld]#啂cbI(!yІ\Ti72r R0/CݲR).=10eBb Y(X;km"C!eQMݔ1 }6_u>,('QBK+IISE8@݀KXgg,ͤ!p|G-(kƫG7}-9ӏdc[܂ j*稘DsZ*1Xfy<(BƑ' ׍Of:DPpȅ`I2X!xm n:PpC2Xg8{Xyh+{elhj'?L6_ ;ާs_ɋZ:^%*" ")>E'=$9"clP1X)! 2PFsBh ZRi$^WY^଄˜ ?.(ћb=/IubZd>74_ ~kǖ?f聴*GNL$Cŧŵ蕸鎞eT4G$]=?}4 xwk߂H;}m[>;^|Hn'w:{K.խ8) >pF{%=-F#]O],>q}-0wmqHW6;%%1^g~Y1PJٔ%OolXM%V5UL|L]kDG5Oۼܓ< '[{B}wOۺQ[ivs}NtO[lzr~>'tW%v~mn{VZozUڱyMkx+ȣrQ#qGį&89/?P_~|?ʽѯQ-yK_ wNf]8oG?}4}2嚤kS7fI\xN>J돰˂(XXVUU4-BSՍ&l <-߶&C݉&:Ftyy$E~%XALVZ 6 tKgUMښ%l]A'Izakk#x)UDCńRIʃ\"&+aYb"ri3*fc4 ҀNlIՎ*[]kU~yꗌ!|:b4S~œ8ߜm[`Ӛ31D8\8Fd a(*A!:ΤpIЋZ ʖ}W4iW 9z"+zF}kGn .Hm!]'iC>ѧyzthO/_~7N>Mj]Z5'?l{w|a}pUkr1'm5()68 s* 3`3@*|1|NZԇC0Xz2T*GBG u@|paU({QaA-899?0u}Hy5GZ>~`/Y_}t6pKbc젼b% R!*n=%9%>=%%ޑL(QTrPU u')G} #M7@QYsBX2E*2ʣA. j̜=7σqvgꋇj]r1?^|X޳t{`xMrPbPdb!UDRhB ]d,1:Iq&%U[$ud̓6 >zIG<^qQV) xvo=ڃs߉^fe[G]tKNH,w=|Rɫ˓AOJy0ZӓW+䯐\J}vcE}2&tC`9bBW_ ]#]J!² Fɿ>⋷ d(*݆:ZE9pdzt>5g sL Bt('K]=?m~R%8M KmʰJE`3H1 5#6_NGF~!(P@.{xyq^R >:|&I 2DH}g6U?x;zwyϺZ?#5~_g[3reKd>V<ƕ9]j=j.ҳFK_Dw_.Ba"pp"]Dw."ޓ:Қuw."EĻx?m"]Dw."EĻRkF]DEĻx"]Dw."EĻx"]Dw."EĻx"]Dg_rHE{1w"]޲w."!# It񎣺x(w."ޑ -I wܹx"]Dw."EĻx,YZDO;-PZrJ䯑Y."BW_"]Dw."EĻx"]D/%>RFZ2/t]Dw."E;l1C!&;{ ȀɗA_u! #wg: mdHI삏$0DA-0(E.+ƉS5Q昌K0.d lUaeUFҲidl]̜`cb3Mx~ ^{+wLb|{wCuR*A2E#(! J*${p4sa%<&V Z)͔. P5d)F*2%vTl)SD+kE`(CB@a- BT\U+.g 'hgGiGc屐z!d*Uf%~%3._2~*E)ۨU!p@KU#N?ѽ tq(Q1DvI9-zdӆb/{h8FxlL:g9Y9U rQ;oV!kG!ymYa۪VBYGcqfC·o{sS/S:!ON@ZJ:Aˡ=6`lsTa l<;tcЯq|z$2,nAx* QN yMPhmlbEAו2j7*I:zHS> Ăt/ +QL%)#~PyӖJR܍B&YPJu+[g-0?uvQF|rԖcKJ‚UD -%6O$Ea2AJ9)ٚc9Ե d/1vAG)HR".k"Hs K5AS D[4> ̙|rLi [dEN޲Oz,.D銏U4( -:WW"$rx_HNz(~K+W1/q\>]Ye@)Fqšf<d%;AXB${\(R5}eګ>_\hk{bP_|@bJk8jq”Z3VBgJ}LX'g䧝NZW- NQp~7-JF2`4h(JQ]'lyQB_hlH A:&y&AB$3J2t)Rc䐄dB,3~9^2cNBI7>tS7l1^n\"]Wc3e(5'&i -ɂu9C9o@;R v QxfP1wh:[y-,RY7)LC `U SC-Ε=dhc#6XU1(}d*c&E<ݓ(kUtJJ/H9K(,)7HP(f]s-e;U9&hMnq]1ۂc8El<6>E1vz_ȣ/):J.K n% j}$6P`L<妑8[R*!m[6 *\)mNFRjT̜=*vav-[B8 4ސQX"=_R>/od~#v9ZyE9(Er"Y$9@e&7h&=?@հ F`- S )#sѺ̹?b4]K1DmӣvG;qd$ Y&IʠĔlashm)%rP"g2{2rEcTTsZs#43xHk$Gc[FD8"{D9 !OElv)Sprx ^Nz.`oK{KN a,;ɚEUj )Dd#-d#g -xX Gf?}Ud:Ej]=.䂭0 <Y!FAA'lM(Z(cY$z\|8xlt% wg~4`Aٟk"o gY|G)~:nq%1Р]d5NrHugϗIiL.X$lIW7y<]}[Fbs=˥tQÛ= / ^XG!PWGn.A[[ĵ弚#w%#L\w+rA4z}\}Yܢ+M_@ s"!gqo'Z8#T2ufLRoF'-4_߼fC{ LP@A-OLG7Af O켄O8NLl+=2IEQUk ln@+ԅoQ ^(E^⢵B9{iq3e /lyFDׁ{M=<<+!S^sJE1Ln| R䝊.'bQ*-\/wSe>EǼ$绑]OJ7zQew\LE4T[]#'JX!)de4e'G}mkhԩlgVQO^۵bW/zH́ j6T x1,gSpll+Ehv!Q^h- lQл2/tTiQuGU,,{yIfRF#J3醊H˳X^F*FfdD:O`5ɪL=2:p7>pJFRT<7}[Sp-6 g;tD+Y]}}2%6>GaڝI6f^WELWW%UZ^qW9ekzu6^@k(,h'V,Pe K@ןtwJ3dJ̦̽'ϷKy'Ó e3GK&۽S^Zi .coC+ˢ]JNP; BdCO ~% ,WR|.XOԪTg!I>CQ]Wg3y-ogWdU]U_V" 7䀘5, jv0?x]O78|n0:L - l2Va/6n'%Tb:ӟNxv+f!+ ]!\BWVʶT#+9ֺ3tp6]+W]ض佺:F2\W𭚣+ ]!\BWVT=]#]Yn5+l ]!\g>3` WnhՁƮvC[FW|=]=FgE;DWQBvfwCB=]]1F+A֖WH`}!S]jA|Jj(8"q< 2_eRA:q.Ɠr6 rgLcQBTHBQ̦W|O '.dijxT_KYAH[\ďb}w?a+W,1Κ('R]$QfP~^]oE7-+#Zoߞب% %VNbaE09,NSȜ澐(g4e)˜\@l᥷P/f^''Gs/s,Uep_e_jYMWr)8ΪºqCu9&ޣ Ù'-'p]-oГj-:DWTwh=]!JՏ#] )I SBti=]JFzuut%+^jt?0O+UWѾQZR.~eQMC/iBtut4KN0. ]!+D~B=]%]eZvD3SKOWRҞui<qq'̖D{ ӻ\=2ӕxbыS0`s-=;cW}흛j7t%zznSmc+!+D6uիc`v@97\*mȍ;n<4V34p MZdiQ򞦏7ĪRvnw 2"NW\\t]+DT Q~J ]r%3tpigt(1+-3"Ju]+@K i;]!J{:B2N͈`ú "W s/B/}t(-*!y ]!\BWV~Ǯ"O,zyJC;f3=9;et%w+sjN]]\uwBkn;]JKdOWGHWQ.Y{/b@sN& P+nNWFi~N [=X7얮#x|Ub]bRV+Rn ;/C:''p-'hm%'w7u坡+AEW v6И2te'Gz^thMɅQ$\V;x)p<,/S\i^VQu˷'g7nzo-<͸K+V`u)b|NB +7ie ɘwXfhg{~=% *0Tí[VG,r.SRprg?J T ) MB;a]fOE4PO0 Ĥ!s8Fy8' ,=Ue.$nᯓ2dT?X/eOGϟooN~s7vy[7sxܣWN1*dE56FuE&7i |JmsJ^3Z]s}qkpc/ dz(e/چBʋY2"`p>|9ϫ'X?v2 ) @5RBBw4 -YܾOP%>=[\@Ba$V<eXn h[T7ak0CNTHfuB yVF<7r?N ̒FRP/ss `Zԕƽbr~vm p%bĭT2a6eG'Js!&8.rnq N] dn&᪸S^:&y9n %u4s`୍c!TĜ7~Jj e?[حT/ itQ3jr^AVoȢfZ[Z}z=B{P-W9SƽZ>vP%rR Xukixt/Mk5{@HY(^8Q*٨J*\U%FհPҋiSn&C~pcܯ:Gen ؍^`4Ae75# q.LW \pmfK<[hE+ P[lvUuQO OϭS\RR!(Uk7 gI%1RV ަJHaR[߸ 3_ECLďT3z.PG4}yu4a)  $l@E-V^G|gӫ@$GגD%z5IHI)2@+MHk4fS$u!s5T؜B${exenz4z[#G=ՙks.CC硰4F,"R2F]v:ZOX&2hp::%@dfoM)zAyoLb:S#u;˝U{vXZy/snn*9|?Mo&MǽӪ3R{'\v+oYT 8>&:u7~լ 1KwOZ9  WGS 2G tcbzyqe# ߞAY@ȢO7-{Y Yy啐t2bovE;<+Z?b8:p\ͿpqǪ"ۍ?6 ck0Eccko_a)U2uJ2ygxe']G8;G\5ު}F8X*znPa3rj|8}7A&ʱj?71Kl*XТ` * s?LkGYT #hbO%9,ɒ!0Ѱ~C ^:LT[kYpE//_A^B/ۻ/>zy ) ocX1l -{/#9ncaERQHԞ4[8%n/)Řhcb!gS2'gq u|/kWu"LjHiݔ)frf{JAshĉ뿢o, CsޓS(|Bdɱ$ծ,ZIֱlı"ܙ3጖X 2HNeG˃r*$i\%aXuQ*+1#r!X`h@"o \Q'C2؊c9{b- . %8K%U xn<ryrkӣw出m2qd{q )#2fR, .e$Cf~7A|a\Kʝ|rNIR)rI0)KJ~k^[>$d 㘱#",&# Ĩ&2e+D^ x~Ef*dQcJ<4EJ x+""XdF LQ}$&bUSj^_|3+^ 7MfO#Z>Z~k5̚KJs߯  0oOO8Yt$Չ-r>I30X3$k^b[2cc\2qrNNx&7'+R'8qYjr$1 ɪ DTLF/M7n-pNDm:wOx3ZTOo߼]_ܹ=N0j2?{!M ĉ4L.5g}[H2c> ̩n=X4/.FrxKYU`v b"Od|zX[nvI4Φ^%>=}G\ Ndu$PG:oF#ifYޓ&*D+XVYՖI>`Gm&6j\I-+u^22"yvYe`Em '^uۖN5?ǃ88#v?/?߿oz ݟ}|[q@+p66="@B^M5C1.g]/j[ƽ/>nqor+oo/y7fO~ʁIi$R(u4۷SBPbK=ҎEOJLo}5>&6HͫFyd^iIR@OFْKIļUȒV[WSF\b=YS$ ,g/{#{{7?_d>P0%;Ir$\L* DH,HYXCNuxu}q n G Ok =hw~vG-'ʭQKk.LpU܁ukPxA!wrC CZ-x:Sj1U :tDI 1"DzyQ(C 50! t!@i#*fmHKd2NЙ J61>3qFgt|V$!Cctq9Rqdw[kM+t7dLC}uhV{f™:+;L46;tYIXBdS`dm 6*F&.4yQTֻl='($TғP^]` g9.ds[ً>ԶDmEfvw=#λBO%PQ}#P;R F6s{EzQk%XR 'rֽ/8?wO-{cڋ[֯JU^#gL`gI;&.i{z9-~~3TLRfx=xꞦ2- 3's$6ت~BQm;5wyʹak Au'&)N06K[nqwtFm~[XvMؘeƦx0toXx ] .zk-OG7}H%AJcZf!f 5MԒ+BiBt1s@i=ݸ"G:c__LGGK(`I$o&9h ރG0ې}P$<$gtf+N^OGUJ3xd)ջX]Rw ^WKkR;| iey [ƭ_&LEFp:hge R@Zr2 $4%m`h^r;#x*+.] dX [e@6t U&$rnK@O+e3§ }oϲFH-kDEc:FfE4V뀒y';SΎ\9DF-qGR3Sʩ,]$^2BjQ"gythI4!"Х|y-_+=3 hRVQrY&a%ʠX7^:OmzԊWNzzėMƦTJ]t& 0 =0=mKuy8:QGuO*`MKd8GF̢d!*$$D'v}KB/z~,(.(4§!$VjLC(#/?hGA'F3`E'RQQdѿ \ `8HQۺl,䁳Ȍ4"čNrU1p~3%cX$8f4iObhd2sC8I\.Z˳edc~u]]w׹D*G[A3^ ؼzU Ъ#(7]YsIr+yڵbee5`~pwvvbNDpP8ߝ%6E!"(ݨ˪<^UOHԧq͍|lE۬~3 o?'}^>Sw {Ĭǻ{_e*Wݗ~z/Djݾ* >_}|ZQ ?Y]+tj$ELQg .0Cp@A3hgq n'LwB-߾rJu*UkKN#$ fa#S\V])I<=zdInP\̪XyPJX;6#Q:(&Enl9 b51-ͳɢh'Cybzz%&ۯow 2,7m;iYWtQ*_w|@e͢sɦbܲT]YuNR j^W)&V|Q5%+ڲTL(G2 ɖBAP Tkdl&!p͸J|Vq,1Ba7X\.dW't~2~\9bx\6$c*߽+`!*9yMV\[-\=$ M4%#]2};dJthlcn&~ӫbiPP[=2؝h1V>EQt1߲$F\|\ 6ӎ;^^Α6 [Q, _(Q+rl.^>~Gh O7Bge]@; B+.Si |J~VңГ|g#0 TX B!HJ)-mHQKvLYUį/c⬮1.4kUN~)\M};V%8CBy"照E*%/  Fmbk=.4YW9l'jF4{,v;!~;7/XF7oɰxq9_ر&N^?cb/F%v1WgGWwRDD>QTb< o50-bX *Yv킲VCV8tLQ][C !$PEAJޅD&Da'"'A+q5.(sxD[EûS4˫"ܧIN"ِWf^IW_:Rj+aNm#=,-*9,Cfx }o71m_%$p(iѸGI #Jpug&;!b+8:q*p҂C*%WkJzver dzߥ㻺Cvb7/LҙP?*4gOr6.V&~gTCq.8(>fgߜJɠtWSA*CGi҉TFFD ONXZ0tRʑTF"4 UUיS+:\U)%W W=)g lOgud쪴`WUJ#\Jht2].cשg~6jP#aPa.󭠺釫^_ؤ)hvtB9-b6ܹx×{;> wזU}FϿэ "?pmee@%k#7EX('t,!b E`t𐽱eC9Q$Te0%'nF]^3'I5V#'S4?>5ju_g{ިyzB3y=W)=YI˂Z  I#B&Td4D"Xe9*I(y!4K@ &!U(RqHRd3$t"\,fe|xy9) )er\)sݙ.9ij89?x߾]y. ~T(#%J FH֊DV2"³ B0 CZl8 ڃޙCS <RJ:/4VhRy*0$BLO >WaL>ou[=̺ED|[_8Ypf- *E- xu)#?U)%r©MGi? #[u!Pd%DH|rv (Yhme;Z}3[XXS=  Qgw '! ^ x}Cn^,sjH㏇czUN1')LƓEcM !BJ[IV.cOG8Wڱ$0[&0p׳_2>?M:h}.Qf䑣CFvY.uTu~^/gߟMxtvh8Ky9?px0?b<ZxbװODMY|Ia=s}9=gD֤jzV ҏ=?~Ϳ1:<M2~޺;l6')euw{@ϝU61 kCA qil=P]soep@G Ns}L!%I{qך+oPZb`/}ylV`m)Ȕ`L>kn{b!ڥ:{||Ϙy;(['>H_=h}~kv0>'ft:_]5.qKrGN- b`\9vd_*%|sa܆'Q()Bg*A*`FPR%S$lBtw4 m_M[x:#32,MweCL|!M+0XzM@j`,H{)EPfjyiDJvAI{u fpR>E6X $8IhyRO׶y~,,yZz2v|/p{ɕ4bQrEc&iXzzƣ|HQ~'ﰷؔ*T+FhF L4I^^FhkGnIlIx?>(f$B5`E!*52Hrx|-Fi݊v3]%=5ScfSH9RN)THg].(٥ |ʤb(DX)Tb4vqKƺ8<h!ZG^"ͼNO!_ħtSnO e]$5d0P ڗ"[xF!)QE$ҧ·:ᥰ`}ſPH!)8CRf!,`e 'SVY f%L7dG  Fv5Q< AGƱx¼Z AXJ&|Nfqr3a>BEǮ=;d#B!0==I(RlT=oc1:1]Jɠt:DTV*c Rd"o *YP40$ %9OJkMAFKBMɠ},:5V7 )B27lS>d#X M 1T>{Fc:fᕱ!->$O?{WF PRyL/02ݍ0򈔴HZv,dIQ)dY6+WED~qd d|VO^,櫖0Iwcњ  PH=pwArq!h-r cWm߰ mc`|%s 87T]&83եt|K{u2KpCuﰺtkʮZ|:u*"Gr@\s>U{UF9"sMP5Y("`p5GCv%[; ʭ{=!l.鹋VrT1e:G\9sY& qsȜ:ǦqZB96s/qz=NpTh9NF jW0 mht۴ɭU^68L:FʘIqK*չEJpײӃuˣ.FvI/P> :Y7wIQz7$.IC%w'fj@*cz|~&]WthDiLD L#d{y*%x"pOQs9tLDdxn\, Z:.y  ^H^VuDU`CL )@.2Kə hrOIT(Y(te~){qn<};שT-RBܢF'}yu}^tU׷䢵wk{=yEk|{.xYnѹdS6-UQ O'ȤP̄e !&"|#XU2jkrJ9V,Ҭ$'@mʤHs>-M瀈jX}Zq,=c< m*, 2H|ؚ̿b\UO%8+6>U:?6;ObἼm˗\sOrz+sgh$J:~Ȓ~ \g(Uw޾%fzUGv=aPI{ݾWoToZIbbbSM8qgi1* eqPe>}m/pO2*eTeETQୖr2[0頹 *YR킲qMm 56p!q]<w!'y"ڃN I\APk:J. (sZc+px4q`eK/6Fp'8{;l7wn^5bN1zzFdb"W@xq7Asb@Au Tuim%ՠt-Wܰ(2 *^9HPºj $,佴 $ =2)y %l)B$]09I(|Z9H߈quRgaW-B+:sA|]y5Ppn.6dm<tJHjQg1 OyHnB| wБe KA"ҐH$,XAy 3DP{JHah1HAͼ uR\HQeKz^RBI2(R]h$'JDFT4:y`W!e|}[/{Ǩ 1)|x.WNrq(k|B.~\0^7D n_Z7ֻti/l; \_>]UϷw*t@[-wl: 5BeÞt<^N'CWz֪wTېz릇V]9nۼICKޭl5A7% 7 s#Lq2]JqCSr[7,*97~EdQ$(uw}UhSV 4jBa91G5\̉9Da"i R!E3K(rRY`бCA3͉߯C$tsGB FE:DCў7}ymx٠ 6Ld}{.4W҈I4A}o ?f$=9D)K `OՒ̻sB1bwPL'O]?68IkU ֏ag[q(YF#稘D?TAr|,3T۞lv^EK/$jLXvYi*>8jRF<0wgD|XVW>,lMib|)mtiIb]e=<-ü_F~M~ʞUZMARX> 㴴I)5/~R8bL\3$n:9^ ]zm5CDU9wB2Foо] dQϹ}iinZ+G2i[m߾~6Fu;$Ԋ\@H.m8/fnaTYΚVGHmMuZ>}|m`d,l&s?_L|햑h4N~.B|rocBHk9smb6d'N_=bcqt*Q׋k$ڌ:/ \HX}:9yT)?ɿTsѩ&mdx/ΉBÏӏ_~?./?@'_hP8隂O¿4 |5?|[Cx롉DZAU6qqffV#e )|I<ܽ_[ZOq8ݞT^X\At&+4mbe|W]Y֛YԊ+~/Ft .4w]4t[+%<2oXI%%-:0%4y%ϭm10{[KֆkZ qHaK,W>$HsU dH+bRLZ:U:s\\_@\58@u ĉ#vz(wF-[t `zW KdA3l҅?_&R@f1qB$K3 h e!cl9dS!6^2W2봯u\ &GzZpT4'+ -8~6sz׫_&'ӛI|)sȈ$<)Mk1Qvit3\v6Ŗ/jOcN֝G^ R"\ eH)CKm&s%E퀣  !mvaUu (k@Y{Qn@=?dz7iҲDLO9q*m{n pK`ckM# ŝl|H\lT5~r?1:tDI 1"X/w3\> c6iT2i FT̒ېHd3N mb:2}fttςl7NӫI|v_ڡY%c!_vw2Y'; b{|Bdɑ$ozzjt'"lVTuXը .%:̘FZlǸN&$ $$I3{ )(@;tD*0(W"3fh61hښK3J1&*׵m^ǙM6-;6yvP"2]SMg'JAT iw}p DYZ4K)ECt4GdP`DGcBin$\s6|%%zQ+{6W0_si{2W}XJ=\\=S >"ssM\ XJ+>ts1W?bq5]OgnG#aԚ WSXZ32N7I0-"6[EO?h˩# - gQqkN~ynU?vgļSv}g`e ΏQܲ׍ݼ3wmT*L9ث\)Ҳ2?77[ue}v_qٸ`Nr޶q~-~H6418#n{UW׽q" e1XDPMLr:S&(*h7m&ќ>MɡW#oL>OȭR2>zhjQIr޷?S:R:^.9qŸ-gA\PD_~)z9nOz8 =": gq*PYJ*~Ho9sZ7G\O?xCpu+4hy8̗ نm|I ڲO1R2tzSB4/@Gge)K% B*SKH! 5A)΄B'IaMQ@9KYZ6;˞|YL$w|P>nQ?p Dup(AP jEB\RZrD4(m;4v._E!)"Y+j! TkՎFx:Eho'ZZFb< kg=P!E48Ǚܻnjo< dFp߻ohaSi΂ rA xЮ^BaqeEH,Qm$\5( 7 L&VڨS,25 g0 no*gxjM$’=Tg^Hpo WA>㵊`U\@@gӮ ÷봻`űAG{BG)#oy q:]qU$ JXFLk^%pֱ$IOm 5$;D$2VHD@`:\IWT$8bT+$BPƹl $ )6 L$xZĜSSۜݷB .x1y&5W-XP3=ŋ`}}rJ`{A*]yS! &+(Q85B`ɋfƭ`qcJYʂdސ1 A{ )2#*'B'РZP ZbB2*[97bx ~l6 wpNsIdFm(K A 0!GBh'Ӝ330qpP<ØBl|f=:2xdd99T?A,8G儝ܐK%9Gm:Av I옢z6WK-JzB*'GETɉǭ/ypDY)?M 1T|]ƖF,QfOxM 0B ?v'~R]v܆i-W|F(Xէ\ҭQ#Z>xs;>8fF;0{:˫r3ۙ Q~xdp;Y֒-Xff*BN!v{}̣W^s7u{s,mխ;e}E,ĤAwc >K&AlR+SzB[T*)TcTY;z+Ast\~xeo}+_ppL_[ Aӿy3a˦q8E7hZ@|v%K}>Id-Oo醷@wN6t Ǽskr+b_O.EUT/UɚHCiB47b;]Ezeb%QlqJhM,ʭN=\#v<I^CF:kr"`dm;HL)T卓hyBmqME1RNE`1r2 @PHܡWDֱԩeSg7[N1.Pۜ 4 gnGO\ض9CՎC:[.Uq˽TDQA- o}i[mқpeQzQԗͷ8sCFO4cGj6?\řk|Jb+oH`CoWb~Cb6NZo v{U_Gn[w |ɲ/ "l5#o=tuK> WMuͽɄb&%zfY6jOitP// |8wI!egcS @58eਅiUAY ZVOt]K?256HkOH rHH+Ԓꦙ)!߽σ׽n̗_j["6Z2[º K6aW* jjCj :# I1̫Ji˽5+3asTL1 P %p) bTe80(3Y`jDzMԼZ8nM}zox9_<^ܺK&36 nVt+L4IA:IHyEj" w:KD"^ ozRիPFF4Km`u4[mMr%NR^rJ}J56qkB ӲlgG_ ԊCK|ұ\W4>S年4 FVQpUa0BD>ϱs &6F ^' #J tܪfS#J!Ebh18o3 Rx8(Oڄ[R9&kƝ8/kw)%I v]z^w om s7pԶ>ɻ0^Yصe W욪Ux{7òNCZ=Xȕ Yf \g=8A1ǧE ,]?.FWWtRf MۭͲB,Y`ɢWjx<-rgJ[^δϛy46˕)wF{^5n'i2ef3!:{Vv9d5ѫR /lB< !ZrzՅg݁hEUQԮAS P:ƌ,q9F@ݾ!)hA|:;9#8dz%wmH)Hxiu[f8/{.^2%%;>k̐%MAXbbI3LcL74ܐh6IXM5qn iBF@aƍBֈnAДRԾ/?/iٰA C6V*:BZrD& LmH*-Y[K\J@~L" ]BAU9 \)alr6".lݚe/B;I[(+ 8&u"+ fh%NVjlig#1#7FQt_yNW~lb66&uw#4C`cFac>׆&xIbfQ8zUM' f`5-, 92 Ff%C QY%_OH#{k͊V!q|ڟȮov H]”bg,t3)%ER1[)Dީ(<[:iȮ֡*[zgYWcs5i&{:%dO6J&9-#73`L9"SHz_\ˑ?rYU6_lRhJtQE ^ HeDI Yn} )n3d2>nCTyD*7  #W!KB.0iA 20[;"3҈WjSgI:8Uddgy2,)JސMFܺSV1yelfK Uhk3b9d|VPKJ]r\)(@ GMnjF7=1 '9!@N,.dYmG66îeO=mkz+VVE^ :_/wV&{> E[;tU GP9Noq @!5oCn 2\EUD0hlW,ŧW3o@(g1JUZ[LqQHd9sY& qMF/FQŧwP^w҅kfL|% f9\8ߗǶ^b7ZԄ d:ɓAQ=>asM&kFGsOSp 6߼?xu 6t8*Kf+ ~䏕F?.]c7-4 `1~6$Tcaav>oTύe("NK^}A@2n})^^Vhg =Zo_vcu㔾vERBuVӒ]qUE=.^K\&dL~P NH8A 1pm)SdY y.bOlw}j$ 4 [&#g*G 2s|lڗR:w.Xh6sզoG{z鷲 Nr2 z^# -󖣐&[c3 :=" "]_=0<`2E TО; J$).n!%鏺% %vyX"Dž#KH8 M 0E)/9Т20Tf`G@lHʣ=knN||E9qnNno6\3 k{<;.E͌T YRyPpr\0ms*%.MLfoz6ByiS`bRd! M *͞!9JR wQK D8[uCD 򌊱%y $#P@tA&YLCEJ{f$ѷzѵ`5 $g9E#aY@jXgY_ŀ;/DEƥ7Y D:Dx 9zcCC S!79=>fu|G6$Y3- Khe!Hræ.uj CVM o84ثK ؋IZ[6e'w~DBK;{ݪhS`uӖ&/߱9Y8s];Ss̮Fy_~Xf+N^K3r7#ңWLa_׋(~'έXhpڐwiTcg$NFO\<{2tWݏvHC'㣷qZ<~r~|>եWߧGY( BOO~i釒S //۲ox2-4KRjKd8~M_F8WkC֯Zuyo~ZZlëٗ[{69Z P_~j"[t0 %[o]󯢾ר&] T f^eK2}.t.9;١:Trm'G]?}7/gKާhjEFmKm?y|^тNFrWJiJY /]`#?s",VEck$Zwj66w7ҥLgw E;iCyynZYoحڤwcP睚nK|YhʼagN/kkE ~TX7&d:F+ ~[UkezMK_ϞJZ+XCrM(-yl0& h^gEcKmt#v:aADZ$:P" P@߆ t*v"]tI=caDV"y^z4X.W?n#AeVcK'rg ]sz6ϻ`g>:dBm5Z}::6>Lf+Qɽ}Gh:[gw Ih8)5ޖpsY&Ӌ2;nf3] 7=нSɻ,k͙M6P3lvT$B|9q9' ΢H!FH.;CY1&wemI 6&Ⱥfj6&»Z# 8(ZV߬Ab T[R;B&GuvWW_Veej9hR $c aܐh6vjP}0G+iB?5rIؘ"elhLٟ[A[ҧic:>!( ]Eް!KA;^>P7 4J Z{VjRA vV e)~@$,yoF%˭f=L&L dX [e.DPDmS"Jm>E0\غ=^!v"#P*&2',1,0-Z?,vVM??3[dPԲ#!&)T.qILuWz64KUWztP=! 2&`Y ͑Q`02(bz-~io8[jHH#Ck7#B̰ ȱodyy$D.aJ1Hpa\:ƙOV" "_(`cCƚ>ݭ!ZG(rIЛyXwWȤKͮO |\PFHǟ'x܉bW=AfyJ!)QE ^ HL6>԰R[EG76$E($E!)gNP1b%D7 GA 2*S#.y,2#::O8ic,o)I,h%Uyj5qUfu^>v xLy$<= GٔxF%s6f#'COC e \$5[o uB 8c jt9dX ȵ9ANOV 7SLLv'ò ٔhK!IP:Ь!s5*#2 #Д6!01+g'07</)V1I엍ncњ QH=/&DBvZ`c= >o[;'<Ŷߎɠa:էםB=qđzo}Yk 9rEz/* @.5oCn 2\yUD0hg}capv4H9uTӁs?{Vk7?0GWI;ѯ[bպç@Jwc:J09zF9QVimN2xQG.!Md,859S'{j)ylmt3s_~qz3N-ռNF 4&$lMFwGn}8wuo~Lo'EǪttjӊuo=Y6zщGE1fϬ9&ݻt'??+[;,{d'۰ȍJoqƥtp{LTLTfLϗ?K7Z@휀u02674F98P,iCϗWezx8Ns? "csbdEDТ)xtKY`BhꚘAZea ;ːq"ɜ&0NzB֭+UyŝݣF&/J.%diɓuY"(&tF$BϳVs=Wz 0=7pM]D,-d$FH%"`HʆZMͬTB};ߟjzҗǮ]L.(YU݅H'm@a:.ORx{V2 ?sT>`'SٖMٸ6YKb(}eO&b&\.H=bUdrxkKRIN 0hI<&3xIs-= |&9 Wf',<ߦt0㎝˓4Ey\yArY3F1O~ r0 ӻ̭3'J1rtNM^WVG,ΞO66QCАpd)1hleĮ&~Ir1մ/NB;cBpMQDEi bdfJS Vq !DBANe 5G¨;RǪٸ8uy 0ҏcTFD3  ;5^;4%v`uI6H!tI^B"#,gL2$emLD.8RPFuVӒcqUE;​;+2A&H3R2}BF \[fYVBG3@p<}Cw6g:N/#eOV2 < n~|GSϏG0TEmҏ8ЬUzu;uP=g۽/pܳn^NA\4RE64eً=?M`-5\i:<Ş$%g_Ş=ۡ]qcp|R:ޢKF"C4^FY4 ,4N -w -{|M 4 qvl)J<)}YjlO"2dRd2BT 'ˍGnV]n L4HAP0KnCBe2NЙ BSmb:2}frs5q< >6xJ=|추}/CG\9b}$.ׯ/=襟Jҷ~ ~TRDHY=/}" JdyQH Hy@ BU!Zl!;LcޙAJzr@n!% =eHX8N^BqJdwQyQ&!>!=9l*=y3'&_PwwzQWLnlzx{z7p53R)d%ě@Ppr\0ms*%.MS xū4@O' !O6-fn-y@rِy9DπV! Ȥl (I= O)tqy=8f:)qø&3VVaV>+pĶ!RD)=#r간&Yt(;4+ Xh<-4^p~I7iZ6eW7~L sr/e:#[ׇݦ1U_L6Ghs+6zgӫM8ӫ;];Ss1ϯ__Y:_fTt~/'$rc伄g_x~`]/ecL090u}eGZc?/gx[yd.p _Ϗ_y~ӘD87l*ob~ڶz2ɿ|q ^=MxǛp7(oNpAP<A*];Z~mCVrvӼ(Oi~?{ ?aZرtK8?wm?߷k1u2eۏj[Ӻ%a Kc 7hJ޶tϢ;Tn"Tu_3-g{'ti?~4ggLɭN6%WyOϿjr9؂]/-Nb@{-m۲oz\W|iFKLtV?w%I Y#dac?ubV۾V[Wo.yӽZgoVՅ'inuگ{>7.3)KxfU1_4s\o? ҍTpJ:: _weO$`+kh$vG7?*$@"e6^6A h΄Щ4$x$19q̃><YL21P2kAeJr^s)@ڐ3:ϵHJBgŗW+-:R_[p.Lk.^>t+;y,7Rׇ$`c4(P)y%3\ptV1{lܘ۴ :Kkm#G/sg[x30YOf<3X&ރO[7_ZV%'IX$H~U,rΛz18igEY͎DwnK&. n-Hh x%.ɤ|oG2/h?7io> b! O7ؼ4Z9g씿 `bz#bj?@v\=} +ځ/d!M2}+hdM|9į2aN4vqYXTuH*䣼ߖEvԫK'[Ӌtzקߖ辅k]ޥnK OG]=DEԛ~0(!Գb-A1t:"X#-@mK'AQqzm,3JuJE"7 5Ta87a+87tkM'6EPKwQygLDhInGF`oR׭sP(nZ6ɠk (K0JE,h)IcSΟ-ՒD U촞lD qkeYb59W}5$y8ZLw*T*Q'j*U.Phwoo!ezx7#07A[~r|RSpGG:ژ[O OW_T8׋O>pZ5pO.+EùلČWqnU_l?I>(ZA}gUg4P7M4Tz $~F!)Ѡ\b$X6ndEk IHziob+}9"ԁSK@)"lr'3C8>cɡ^8t$p`[ȍ)G;{}i~t9\/&~)[rt5U_Lb{*\l B$sɤ%dھOo kYK( x9S|sck5 s=m]0G>.-k}"-0PÜjj^R$(\qLD"|:-Vg3ӃTZ,d 1AXZ|̽S;,j#TAE 0L^Tfbvׄ!7dxgFEe獩67QZŸ 8qΕEV'4J@fVDàI__x\ž98h>'DU/U `mm$?߷HA$;o[#KZ3>mU[ioDȈj i_a]܈NB&,9kdilHjĞIVIL9`^PJ[Xa< 띳bq?[7|;G29FvهS6."I }()UbY:-8~]J'M@eNRu'H@qڡH$R %EY3*chQFVG#$^Ҟ-U1%83<GG;zGWSGWp A#^' #`F ,L $Bf[!Ebh18o3 R{0*Q mXnф1ٱYstt5kfuPZ˭jTQe//4f<.Z oލ[T[دx^>x3mUO.fu8jC:u#76m:vstIڞKX.]67WYMtY\t#y4 d%˹n_SÕύ-j^+Y>UyD]f^!T.בv|PumtQ5t ;FAv>.m$Cehm#Y<$[8Rr;g]M|i+K5%rIʎֵ E] R$Q6lI`{(B I[,ynum;et%S0Er{az}k@=t ^b:$!AK@,2dXpeH19k_dMg=շ]:ӉK lYQZ%:Hws:͂rkΕ.y;7>mivawӨK}%z;xt_yzkwHϬtPCT fp ˬR ӕ)W9s8Q7* e(k+ʺ]zutdzg6i=Y& $a0q{. EZΐZTgH RAj%"R:|xKCUFXk*Eʇ*Z-t>:ZzCdh['l'XbIWaO4ޘ?`ҟE(]/׮RM* A+-w4TG=#Wsa *~}N*hLyF$a!I&eQvD RBrb.xl7?D\J$YxZ4 椏Ūݨi$̿~cӇ)J} N?^!ӏJ 眂($8FetrD5):R{s=ƒxˍw O> jH}42ygR W˔Gt=Ĕ`f sp%Cѣ"B+/7+]w&vpR8$5R=哾ownb{vJ9<9}_(=~Gv< 91=_=B `8TJ\w<9l6S8O9[G"2gT(GH 2 =ϊ~gxD;?2h]GmB uٍvW%|w#ԅ/4hJn\zu BDcId+ɰէ荵 9)B !nn|/Z-@F"&!9چ$y)=r` M,PʛiδZ`G\jU͜P-s~z_דiZ d|2yEޗ|]һvg%7b2.YS~8.of9ȳxx{B׆Wrj'#pxgiuo y.Mq6<|>̾Hq$xfmeML0:0aeKZ}_fuseWV;=Oz~-y19oƨx7x>:`J֍ʐ|qe'Vd?h(nNpw{]j'n}~ilPwP5:rvDX+,iG9wl4cΏ~cb//lnNIqW˰?rUo_ ~έX*h+p423X|qd1E^q<Iߗ?>IMfw%Al>E7]9pzLKVW+?$w(o(v>_&/8/yq?-(Xy7_WM~חTKd4>Y嗳1_ q! }mȺ:w?--/=NVKw69nZ Ш̏ bEKq(z}?ޢtguy75؅Х㏯^Ƴ#uG/ 싿\/8,Pl\Ob[[w{AwXfGaذ;x26?q7ԴZ,Sy[^|oXfVt3ֺEftީ|K<ܓEOGU8[^Pjz1tFmnX"RM-7Z277t1]bҽ_TwQys4<3;J;E1Hɔ`L # DU8%WB%$M.f( o7kKf^2kGd}簋a!Ba7xm>(2}}_$c,NW40ꠊu{7Eq}t:ƅ﫞1Ǖ@y\'uI<9/DjIfz_Es ,giʎ[--eBaY">-bLS-T.UL0Uܢ >8LϜ LfdAgQ+FH\DmLڢsB8>6'$ueͮB="42I2Q |9z2(ީ(-h&dуA AKND& Lm[ng{T|eJ@U&aAPeJ"$GȰR69Sx.|`2x4B 'EFn_&2',1,0-Z;ٱu&Ύvvx6ߢeg#.&)T.qIp!^ 3<:4QeIȐz*8 hRQrY&a%ʠX/*otP3>N㤵_61RRY;*uC`B$( | 4lhr/q^g:Ҟ.5 oG;m1&X@Hsd,J^K_O^KQڑ2gFd;aUrIC:S}:6բ1) YY g>YSJbR;|bytgk0VP3 :::s=5I5&,g,&@dI6QB8;O6+j:(%%(TYSյHSż`[EOrVJoxؔy@))OIL<Vp"l*!A+P g@)|0P ҃,䁳Ȍ4"[,&|CC@rmuٱXȼ26DnR8 SLXĬ&{Jx̷aI]ՊQOa|6)쐠 njF71 'x 'N,{; a,@}ڶY6׆-AtMJι\O6Ir ^BNgVo.0Z /~+M~9* d2;Bׅ0lWr'-J{?)6. ZM$IV2ɅvVY~-{r9fbԵ@7ҹ,izy2r mׯ{([ õohs4_ˍe/W:;֧tt/~z5X'dr{j{u]3E sעNgg9A2k;_X!WUܫC-UVA8,b`rD檠P"Dm_/RxBU;/30D3;NU߼Ll-wyKwުXCߺ zGzq@ F9Q( 2d .AB霹,ǸV9kdNA]+E\&TB[\ /qr6NpPFotv 8 ._M/}tX/oN'N[F7jݸ:a2p3p*d%~օɀύ%!AYB^%/Apd! iN G* l!\} L4'w* ~g r>Տsǐ?N K YpZd]l$ȣ|^~KE,5pRIk cCLOԪY]ԢzpvQt'ݹRbv>\IS$`^wtrQk[=ռjD_?0{6^]t.ٔdC]nIO3T+##@? ƍE;=/E##/ֲ#IR2%SIK2YdeUE<'3."'`Y帕 RÏ/]zWs̪x[/BBE%5αd_+EWPRpU[CЎ skc A,\,ə替<`糛7 %(:*iC Jk%J; >8RrJPB@P$l%QED݄"nxH&UE 4`{C m5֖FyHY\طmuHP̎tm8Ùbŧ$JZ(ޘZ[gD6݈xȣfmW1uvCq1tE?ℋ[k b1l5`rM>%U XTR_ECV.ǂݬc,x#@؎V׋9r2a =g~S&M yԋԜ\FX"3UP|çr槛y_rb?]= jg=ze{y s-~_W;D|*sk~q`qJ0y|L2ZփB79ѣ!}e1`6Xe՚\kXD9'Cp9"XPȾŘݍڂںZc I:S˹~y:ZhqEDΨ,FC&+?'JT]wEUE.نCvdM`Ilib3۾o'1\bN.T?,tA}H.Oe--4<[9A3Ç@z}vƉjQWd/2>ܙd-뷊NVkU'ܜ.ʦ]vU@n oEꚭ1:;Dmjm|9y@3?WΑW܀@`aq(V32 8wD6B)ޜͯ>.p-A 5`rOqΙAH [5MGJN\Iu?~^'1Yضͭ2- lW@Un"Brg(L | k49o8)o#EV%[w{ gZDցY/o<脷ZyޕeM9yz9cs۷['1{g)M_z׵?_ODTo=|.E%Jyoon .߫^cXVvD&ueܱ,Up=£8jɝٯUUߘ+OI3fU#STD8X1E`u$]^i1VҬâUd#p LU`̛a ]4!z?)3pq`+MqzZ7ICIЫhcJ"` SIW70hM!C EM(1Ig"TZeHBg̉R6#/s@F״A\'^5YhVn"ΪlS֫|>]٭o'1Lbu3徔/(;);hqxzPU~M)L1>%OMQu[M42JQ@ڳ6G,FɁQao% MKVnP*T-IZ%`T5N6ݔ! Xr%ma.yyH5mK]*? $8Qg>9ݩφ^[zwA$"^L QkH"fY;ف;2S 5$c#򱫒ZX߷kDەݸm-Ar3-z*{>;/ab&wU&LwMUAS C%6M̮?m>fsk |~{R=g/ MBfD1@A" t`bZ a_½FK Vn>׿-k@y.'6 䓛Zk ,|wrrr'}8ozgŻņm3}]n-7nJ<'=Aar˟ߵ lӓEޙg\ʎmm[uc3QP}ʰi"]f.VA-r?~?;ؘ?x0{| \ W'쟿qA -tÃ/>_\}p~8z0@rCoI]= &SuM1iN 0 HZn,]*DeIG9FWrX.(" .SQmXBkXS|2PGZbM!Ys r60c0lkyp\ #Z_9AZj}掹&vqvʨ`I ] )6-5UH Va=l6qJ* 8 ,!)_"\U>^F JjŻ]kJ`1uꍌ݆GVYVL΀gt9=>}5Yx|fV-Q&r 3cbޅ> q\JShŋ.)1<{STK:8t49]􇾾<~W=9ގi~_dz*-f/dz_~||Z_!$_f?^7=\gažd.9 v5gY ']X?kҿ8x]rSEg2/\ Xs@3{̲Mz}}װ&͚է .;~X8NDq}TB5s5:swu-:K)!A#)E9( ({ eJ봋hwzdάsfmɣ@yx5 DzoҧDiia&XqHS{DkE%t,Yx#9|[q,U[ ^W=lK@)ezèhyFeL<f؍(Pޒ *b˾C< 'ʼ'UYB.$C ЄpoVM&hn[np_$ϧ .ܻ- T*mw;gYQ<}ahdej"[RzYL ]MR>d?|'We,ˠm.Ozpcv3IOFA/O}K/Xj!Ocyɭ/-NcLTpTs5VRٜ_hw;;D]~<%w9T[kFv㩆S;af@@?Xe3cƈE~LFRh *NZ >M %%c88/ LD+T4h Q $0rfZ̭Eq5СGtPHra{OоaOٜ:rGaj WԷVT!6/'b\QmsįE aP*aPcVYJNOycET"T'If! ԄzœL`r0XLG=2ƨ,11j7aAPI(T+ JD"# %@Q\1 A*!l,׏zFΖrV/@ONk5=,m#!G=9`2 Kk8v,!1"q5UH ")z{ hJ%<: jQP((z,54PQ cI# tVk :᤹GB.JQk*EEx0dۆR|7S(Dh'vD ]O-2",C6 C]*M#r`WFP0iJdi끣 tމv5 yr}>mȮZڇd !}YM$Z)H02^h5x5uўx̍ !JGNrWӴ:3h.M< UK C"y o܁|FlrQz+'RBlsY04rWwVRh'ߙ]^ w*G#-4 B1WSLBHƨq5›6r!Z |8Iq>Ew^,S>WηQ͑d/; aJ.> )$%Z*p+,5gH0Xg{C" @u'oe} ТM&<^`W 4bVlfط9JPX&or4,a!#)d!Wr_0D)YG%`o]"WV"G'>IRLH3,pny*fEe |o+YM!䂏?sz6H~BԽuyJhDȞVbY/ǛV,,Osq8/Թ.xH `E`Ͳ3^aK͞e#8l 'OTa?g+ 8!s+"(1Z_*SZ$iAy*ZuR(E0MJų$)hjSЧfEew(N"1BʾUӏ lrq.ߞf0;}faA%kKVVjIJР%j0Mh#+0S9@Wfi9Qǔk:mOuՑוG٦(MHEL=#Q꠭@\*Lf+0+AJ!"SbhFY; ZDl h}cR0#4'(xAcˇM B/!Ǝ³qƳյp{V 6utp&O'91'S?5`ʓE#;K9Yt^gaX.SP Fazr|=V(w b)R]|<0ÐK1bHԔ"D b <FH7b{&\XU!fZ)MKܨ[u|A[r?\Ԩ4fxgA@F(O4e4C"ҧR!Q+fHTJٕf K3 -Z.c[seQjs~6-Pix`,(3xn23Ńt0]amJaQ@d 9ȑ[oLH8QlJa^iGjdrVk/;^sXe[vVmUۻ*cB{#nx;u d+ͺD%m* ISjz dk\9.ƟZet:Yd}~6e-[v~m͝W[r=?$_}w혷y@xVo.K64껩Vf=sgJ:_AKJ@r{c?Ȳg,{R5W cdJ635}}XxH>LUٔ׷o^ntryOT')E)uhށvc/aizwXJK RM7dߏ?p Ԟ qMijc)3N)h%[[6Od>M'>M'>D\XZ_el(UMkfQrJV(x:{z|)B&[Q@gcN }xĆ?Y_ly|<1bJeKc$ݻr01PƁS(40-xhl~L(qC1Ebf+ˇ3u7!wC:;FD*O`134$-tD$e2`R.$L 0(:6em2XTJM#|Ȕ>U[2!l%BqyM4T4DY}dH)H}[_ V ˥7H:W USHα9%-m 'Z9X`N8]y/]}}:>X] x%m/w.M>4mKFL+ו"5wR:xDpxngBcgP|b桝u"D޼U HdKTKF#,5kʕbAv>OS]I*%) y {PS2$u&r6§.+Ul$ɑu7n-D=$Z<T7b x&@(R≕,+Sl*yGWਖ਼OVg'1/FgY !X"ߚN%/f B>7s0ᯯF`$<?5]f{le(ztx} OH-މ{8գtx7;C<ګan[Q#Zw,DMU%2nKgWsm,a 3_JzR(BC?wS?IHD7l_.^`R}u%d|p/\ZHi' ]t=;zFqףK}Z<=?Skg7Eo/?{:0. zV_&v8yn[.t692O{x[)N ƖAEgmN&g7t줷U%nnuɾVl̪Oץp̯~?6ⰱ/x>߯J6eQ9_7o|F*W/޼92M[%ךvoz-:ܳil&CӆB;kՐ]i!VvdvW)s _{iz2 QV 3gOin;0M\g;2AUrcbΣ߬XPݬsӼogzڶ3ඝ㶝'bgi;gt'I,6/ȔH{?N;S ;{Ї<Ɖ@%nr̡nq-9Ert~t^7hn"l9χ)8 a“hwo싅fOzə1ӕq:2WD냝btF<Jݶ\g~.w{X*|Ǣ+t k|_v+uc1f/|wg g"œB Gnjtt;l|8dыo{oo7 <:.ulדxʙsѥ7nUF5qx}!%LfkA|nCm`SktÇ9ջ;W=xboc>Ov: C3>v :ed}N g/FG_=(=,N}U=^>vZFkycU=d o֒q%u|̱Hvk6vxg"#Ƞl+`a|E r70|ӯnr]q 3/05Wڪ+M=?[N^vN@ѶN_LZ~ËkKg@Jɏȧo'_J%[ Pɾir\%1eƼ-ΨI%WxWwp|X3cB- 'v ڍ|߇=)-'bO #vfOd4ic3{|jv2nN%cyT|̶᚟ŶuE1U"} >UI`_ *GXw +< pCQNnJw,.-uј171E<kȝqK]ü=EWm]|bI!-}鸻UJDbܦh k:qwWw.2.|spZ?_ GP0jayg&ppۂMn2=g*F _V)y;qؒ6*y;)\%1-JbwZ¬\hΙEl̢G?Op->eK4:iج컙tԴͥMs+XADH%i>GBTu;C%QZ4֐yϖZ3&u)Y<t@URB0- s`KD(1K16@7:eZI@t$)SՇS Y"lh c@Z˲:$%LE:"l BҰ&U`BE%Sj*Dߥ[Dp+^et,yrh.ėQgWὐ +D vaˆ "Q<\/wYxc*:$XyB )S".񓏧"L.VkqW^55ksH9QJr-^4kZ)*!w"l:|(p;7I@-r )87 VAb9kOeI&l.-6JRscԈ%gaE+tլkSV(`JPJEv)I6SI*e·@4.Z'7d!R),"VPT@t[ 8'5shQ3\M#oݠfD)R(.)M*јGA_n27p, jBSp qAo b# Jʞ˺ AP救^aqHJڬE6,$x>T@Z>{JC. (JGAEsx/bnHaY-t*X$DX$A  l3hBE+Z'I!EB<^AQ* e HSo9eZ.|̓cd+a vm$|x&YP ל !.d BXz rHa{dY;AMCD` ʮBL\΁@ Oź(59lkL:4i_eR A3h*ԙ@Hqps$eoz B0Gj \t=IP6Z'o"@N):MǮlB*vNlBE]%A/k.6򹯖4b=^!EXB.`EGE6 wK ||PDP-fx  w r,am.F0ZN:G!;T]xdWnĠ3(f4 ,&VȀ0OGdaصFt 3%YDi9D@?x ZA 7Q%oCAaI6c9*agDH! cȡhhͥV 0]oV 頚:+c(ʪDh3ˑlp)BM^z+"t2ߠt j/VКZr/ RF|M:vޕ>q#eJ_6q*pd%oY5!G^3C%Fć8C4Bb0lǏMM@*<+St PLq ,%VQfltK!`a¿I mw,:ZFVY:ku2ނsn)`l cT@ϻWCߴW3Ô@ux0/AFJ09 ָ!vwAs Nt{ݔ弍;ԁC*`Ɓ SqZ*0=b%x}j $ l% zQTH1䘮ӑkW<0ºtBXr<E!JΈ,l ٭ L>iKxH qhH܁6'/dd`tq WVҁ%5e1ĂZ4>‰tsR?{fA2LB@j R2PqUxu*[`}7 B ipA, @_=s!R:)Ap:t h^20 fp=8) ŅS`%qo&?AAL7T yʕjH )@r&JK8!v.pxJ&.x H-SpeL^U"ppDЪҊc+TsU׹nzPip3ƨYK|8JIV?<8'W#*D5*#@ 곣ԀahTφ~6X~/`|>2ťSYϧQytq[`MgQ\ \K'D߄! ,g_)'t1}|'?d?jpKo9/=ώL($% lIrSH{&v4&<{q9Y,Hey\:Lpʊ \/+ʦ~aJ5*76y2x&Dqa$_7MReHaSIvpf3K ]~e^#ͫV|3JN>Pbf[ yr@ЇSȕs-Kn=D=AO55;FW Mbhq(SFyG7E/࣓,zAQM(r hn5yT%<g_/̀`A6V9{+'}w.R74b_[8I:;V^(m'de]>;ǚ/߄V|a:Lgg?4ccWe~c]k׶Z_;h;rˠɞe,Td0c%rEq(ZxZh4՞ƩPmI1_|/$uۣYo`VQ{xm3~_ 9oܝg'j8kڰE.ra:Ha/]`:Nb)+4Bk ( 2&Svw{Gꯏ^$&Xx}u6<<0_dS;HKo+d:g>.ˌt\3V|Nlf5]\MHI%V\X ;En+T4Z.̴B=X<ۍ=pfJޏEj36T4):Ye1 qQ.:O )5WĔчtZp=T}jr|3g]~A;<\5녿a1j9]$VնhոwWݩEW=/۾\9cח!iƌ+FkۺRIlC;˵Md0BEȍa,vZK[""VƂH_`2uZ]lyZ1&'EZg# S‚,9,hZ0/'d?RlaEU]~m=MLnv ϧ&aCтae@O67ז.ƳymgU{42$V0&Lq_ok7=6a:}7;I8O!w7Z}g@e|!sxz;o4EseBU/o_ἠ{"P~OHu)I F8=Ywk6CJMz'PJMS\8#sIlPNp"i8,( Hh$($ e!X4R8E9GÉs{8$t~EBDўI~}ox z_]-zpx&ʖk.'fXe Q+9e8NII/i1M1L1)v'Z֌ݍ:ܐYZ%"TJKj}2 `萒JNB*?zikzG7rނ.`Ǽ R6'.@}U#Dq]SvּBkcySb~~pi$F 9+b.aUo.f=W.ޜ&wgI=]vu:i5mfY^  f0bji9ѳu|Z^r]vIJ^g){)y8(,f2Kbc@i6n"hTQQ>ÙǟNǗQlO^sul\ldotAAyͤXW]_͗q@O-kb#RY^҆@X[Yv0ɵ^]K߬ͫm5Pu/m1ۃk?5Sx|NvU93>^g+0Ś0&yU&TlTz&Rv_ÎHoYB+V+k7ۡuF]-,s[T/`VTk/+Ǜw`nɔku3ܢ[*]POAs K`ɃXR ) 2:EE-PJ4[ϤYp V0^^@`'O}Ka׃_b>v1h٨5+*1e_i;E_DS*AUPNR*S&dNhRI"Q^KoxRgAl(c, jkJ^=e(^q {# TD`[=;KvS4= ,ȳLG){&܄ J}D-gR2{J%)zq/qhK>aK; |+bDZQ gʚ8_ApKUY#cg4Pԑ%qD\L+oV Ė `̯ʃl 0;% ͧ8iIn7ͷi~PEqqx470a}<>w~q!ҫO[NyǛ6޲' *aƬdVRH>8I lbZc=&8SUgj0<0((!RmϚoHh"xfHRC\"a@tIG[RhExNy鳵4'(bFrUʐZAo&j`89 ̀yX}&JWnZ|~%f7G[{uNA>2HkM(2x! RF H(a+QI%C%LݰF3>9}wгzv!O(t= %bټj!@G8MHYRQ0QC&SzHU_] v?7##'o3HBc)Z "33(YEh. DL*XbġUB݋j_jՀgݼ&fUO!"9J̪8@AP߫ )XrWOxɌ- 0IvSZ a` % IBQ"Zl`cQ-Y$ ,}~P5|3T 9fYA1 d"KM_0zrj|h;kzUSy ;F3ذ;bz]F y.cqQ+n'V590jݚ8Ih!54MtrPfo[rN/oPHa4E  B4A0F*6dCYn3`.qݏjg@)2o=l3UI4"sfŖ 9ʵM B"toPWLJhGd1@8k:ڇuCQkh@zV!d.Q{]gvbWL߯R#'*0`ޙ@Kx G*iQ@rZ:zOD-:`V3Nw~O{K|Ŧb]YZ'Pi*FF(HIYPMx0 /W;hZkI`T"H"@R-@Ť]0гe<]ölg'X~ѓ_3e0LDg9EС(<P )BvD5Sq$9U Ԡ<ڙ|(LC6Q'Jg:ptB|Hr ̖KDv&l͞5D1L1{q~j?g0k`#,lALa);h|釽='JG!P: o՚o~X*%*&aQs*\\ y."d+h4Y(xT/)!()lN.(' wNa}K܉.$U#E FQ#ۍE.vr2ɂ2:ѫ\#4iA)v42h@6HՌ`}P$,0nzO,Ea:2Qd1;(+ٺa+qVY:SÃt>̚\N|y @@}l=,c);bZEU61vJ[+TJd@o !:0\d2u˒Un%ΚeoR R|AH4!<茌9 dXPR`٦B$VX Mn\c&ں+!Iikml*r7 )gK+ʡ,m^4aI]M9ncYdD~kV5f'Jwca'dO;PɵCe~}}W͟.c[uJɠ,#ͲO_ t7u$Ծ{ߘ_u!dZv$t+*Zu|m޳2yJnT6rjq?iۍLbSm,%)32 !#/ gئK d^HwJ^[jb-9_^ڈ0(iTIFhtA= 'ճ!^ђ*m'jPP.Y-8|L+={9I-D̺M y q*K8޵ѽv;HTͩ۳Ctх</8UIWLj&claoGE%vYƁ/ik1%m-,6i}J$29 &@>@%CTXD+]P`S.@51K. D9SKD- 0ױf˛kbYs^Nt̰7mkuWX^}!u5|$7 HĔǰyzOZkjc$gNAjm=a (>.j2jFbhp ej3qֈ~ULt'}ggID{7ӓtx| Jf_Y'O:Kyj/f{sv<߁ b5g} :%HuJ kIQ_0eV p%3ƱT P!%rJ`)i H1L5Bqvya4cS,4\t8=xItg? z{r>Iz@&9bGI|gFYtޢ*${Af+ Ug/d% F& F*maN0R )H5FfF8> .PPv w Nc2  ڋ1#Io夒2(8[*4( 4m m)E%P"{ PS%đ1\t9Xs{~#`xl+mcD;D!o<`CVZj'5%Q2IFd;\.iPo[-mmz1rr_[VM( 5vяr 4@hWJ\,ԻoElA{x?YGqEo'؛ ly28;년|ä]iR_> ]Q+`ջ"%z:pExo?}{½ofA #ym:;jc3!_zŰɈqt>>;dqd42]NZ|(ur`:}W(mk ee.8#]Vn|0mlkxwC1\:_ovjOe^^ =h +c I*!ZJѧr =I>qum?7a֞xf\i{vkEnޝCCAd0C EP3R(shRQdF&7zPJYJ4@j'4Hs))xmN!ҘFG1wumsfF7CRIn֭-4А˒Jn)ʲlJ I#{\%S NiqZRov@ }22D]4:%hV.C`p$m'놙s7_QLb%8K~4M0~f GsMl)y_G(@E4q&")&ԖZZYRRЉVй^w?0qFMa,t-(X{H!&]ڱ@ѵvHJQS7w \I$߀-Wʊ7\LOQ1+9{rx}D׮Fko5ldWc3_Os=d}ZcaIWxUZ"O/ƅ>PQ{uF)Nj \rϳ1KѷUft@E;g&o t}s)Bܻ!=jx`y~}*~3piY'o?B#3h=QGɐ}RmOL.2B*sm@އ!;!jS1jgf1OQt-XO[IJUZn[QA0 b,p.9٘J3Ks!1А%t}(R&VJ&Ly|(s7Rqm9>QӐ59R;mt)Q,{pvD(X$좻jhI'JTTl8Q6?.穨TfF/OXx.HŔ4wUne*wgjz9_S65oDQʔIYk G ,ܱ4 tgWfTUg~>#SWnO; jtI\r@dx}l&v(dkS[cE.y[ -MP| T'GAj_2&YPjSlH٣[bIGLrУ2rwRjt,f#>VFXX$||rGHh}^0ƦdB}~coV[0Ssl0u%rEtS_kDgѐ3bTR kJP @tԪU[hE `lZ`v"F6(]})H&(,h͆|_g3q7fkG==Qc>!%0U8@ fw5$ ٖ0o%Fo/z~bޤޣG55l$VD$ٙL)+!b]0U.}J Xkuͫ'*}JRlL&%^;62,18Vrl.={?YS^(j a1 lb`)(W(KG3"jۛs%ѫ0o۶#_4%Q)}y;]˯ϭ^޼z)/>lm.oY^<^ɋ =tL^6ŋ__7]~so~>yurd:l?>/9#-ض{LzCO~Cߒ~|,ʥqƍ!r]ׁc_Q=z~) S99tܿGKPLTL6L#ƹ,|IۙuG,i˖ry}2Tzb*Q%Ռ].-5 fD3] "YĻf!T&@%3' ԚB%aoĆ⭐=+1>v _;FmFRb8ySC1()-ObghDdpc =LHV?kTĵ}c+ߪd[I0FN BPGK}2g󵶡R9y=?a~rϲd\L]88vz J:޸}sO^5v mԓֵws{dRRkQFOj ܻ6F Hg(&Pd1Z3>UJHJojKlk[z3F#0s(Um`B? ݂Wkxә՛iWW S>>>}w|myT*Y4TJפFS PP_=)[IcN_'{ZmMxbO5TC(oE)!he݈ͧ'Z9sAm`o)Bt,dq%Wb=1J)0`[bȏUh+S4CVfYtA(Td,Q1\EX73n<\'֡gcG}1 Fİ ₈zO}f8}$bZ9ۄLT5cu$Sc|@4 S W.ċKLRXz`Df݈ft\s%4る .*7c<աk  XGzTLs`b炇üc.xaWgvN_]ngjEpG5䦌b]GRDՏyF=ޙ Ϧ'\7s ׭ 8pj3f zck=]CJ-Gl4(a"yb4%o 0.{c[aO{(I6ָL-h^j9nj͖Bh>45N_!d@q$%X'l HB*a V4Q V㋉97Or 3gϚseҁAiNmY #1yy¦q۷::ಷٴ{cgO}.ӞM{J4C[R夋a}=cI9X EiM0N}k "<"t__7`2xٹPK^c!;aX(S jEۑW=blTM58zC`p$moٰa9w W!fG Zf'oW==ۡKбCϸmzHNpEq='omXx91uMIklIV@[/3xN/^zvBς<0ʦEif@C ^%>:!!.ɐeD M_|+56o{W?(pz~xnys<}wᚮ^_Ë]p߿[e)_]۟gj_kxsO~+mnfe_#>TUVaץ)qE4-~sx9b{۩su|'}eZלGQ*B竱O5?u<}P?bM?6y$K F5oB5yweЩ_*1:!XN^p<9ST9{z^4¼8e;IS[)h6ܲ}Ns˜*a!9IYJ3֒5#b]`qW0 "[Jm7x_ҿ*qa @|v^^iX-_\w7Ohha/,BMݏHTқkbbCeFFާz5XnK YFnsYlե6iʠn5&?RU]Of]Pc,xkB3b$!ڰtV]R5;g쭊X;& TYJ' YL,@A1'H(TZ]F3.XwH!AqdPyCt$؂[7y`t5 DzXߤOYL6De>v1#uD:xOKu^+^@/"LJkqCoK7%:=kתgQJQ,ۓ Z1\-;K9]iimVxрmP9  }w+1IY.t{=k*0H A8qUS 5Fg4!EzǙ ob ⹐tpk8qcbtXfc.3Kj<+|H W_әtR#f^a1MYE R#~8)` Tw[0wԽ)oǮ3ExM/6&^@ێPAkߔK&,b2bލ39(Gfb9c)R[KjQ ̞\#p%.:$;>Sq$z13J] iTL*-rY,Jx4 XxFY; ZDl h뙁:aFhNPƶZ%6u⎉B&ףgvg>L}bS#/?Cc6FESyUr =UT7 S1cUsKm@lQsi#|ypv@v?@N#' ,Y9C`0J\D$2L1)AdYb Ԥp&q ra7S3t($%S5H^Zf`z'L:,J#% DݱE[6*USkT&)*;-w=w\":wR^Mǚߛp=䊏TaSH1S"qΰ96HHc{bk4<%UUGGU#j,J,JkMU @VD#K R'np搾!2 1]<[9vd52oRgt7$)rӇq <ȉR4gKr_+RXA@"TH<5$OE`YD!f=b+$")sFR'1R 2cB)7L(rnR#"( -J) 2ضwQklDwWBpevwhoY }DHOpI9P,WGāEjkDδgR<ᨕEMP|Or m ntj d82_'%Gb"("-5J [mi7 Q0!S&aJ;f2 µDX/5^#Zpt5="FX$I尥eNf&˛XtY>}&ײM6i[6WH`s6i FFELYHH%-c\Ҡ5!o[14>mqTx $4:&EKaf-0 -ZI%̈Nr /LBR҅0qM%ГuGX5Ġ $"eǔA$@ɺeZs wDHyQvອ驶m&dS[BE^1$VGEHnqDjw"ݫ w ^+\0BS!YP@Hc51Z3 #.@1Sl՝O?.ȸYkL@6vЙ~2 ) '|:}8]iJyQ,'LI&Yw٥xxW!8eyOZ~P>dU~4l>Ea'XHa\HA)̦SoO ,'(ZN `UK5,Wa)[;]ϓ$0?ϥu*f#ӛf,_ۏoggBG9aMU)hr>cB9q68|}0>7lOu*Lkki770̙ sm߻ƖGpͤ~7~ZR|piK/ozU3de3˕+;> #F1Abf7Y9`'+[%h}M'W,s>mH̑0e8v"dØelXw{ T*@'* 3]:~7?6?1Qo~zsװ|#0 6U"hz1k7nȦiP-^Eޟ>h.-lMrQ!WtIRxٮY 70ͯ+G^{xKRTڗK"D15/@xzZ~_:]aă8~iϘ#90IGFJ(жRDgG! S0Pv驍 ͒1׎ \s]}ի߃`|I/!l6B_~|NTX6=m n:DC@*aM`)wN)93ZRBi#G TO=_X~:ա(kur3]m@Cb=T`!c=X\.;;RtDe촦l;mW9G7p1 r&اc 쐧Jɜkmn0]LҝcH̏:y Ac:+sNj|~ %ƝMEjƈo&~"|YaL7gVkZK9xAXuvl,M[኏>@yzh-7i(4U+WEXcBݳx;߭b\5AH B',.U*KiT W/$bS՜ ,^p2Ǝ {-\Rk B)@c*ۜ,W(%#WcDJSfƐ,SkpRRK+C&]AfqUcNEc,%oڿ[=F~j7͡׮vW & ݤTW=EWUXˉj \ei;\e)%ic.7~XS՘C5fU}?mPAG~ڄxT}TL$<ߡao4 8m35obrMýFV֎l#HR gsSQ5jF;Ob+SWMF}f; yjу[$&1L.kXScnR.L09 B3,<WY\WYZE S@ W/#k\eMs+R҂9vR^"\ no)XSr upK+ L>}J41pU)pj.) W Apkќ3Y\ M+У?㙥-\DҐW(0c1pŅƐ,WYJ-\@29ޯId0 9YjҪ?-\}=pŶz^vCdI\Fv@p]pءB#Na'|5F vpRZzpsMr*e +k-\cr]RǛrlQg +.5zexF#Xyӫ!^ځ9}!8ՐR++"@*պ pQIET6KO}Y[Vpg5Fd HMah0FTƛHly#ur)\І`]T_$с>ipTȬӮᵉ9o}@:iXuYfZ.]Nm~>_R@$ߛ&6lǔ;xG ҤE,vW覬di>U,iWU^䪊Tdw5γݨ}x]@ .j`pA Ȃi :V;2}|.'gcp$ BFp a0:-dN*.GyUCM,WP3pktu 2&SF,qu~=tq n?+G8'zHuu[J@,H Ч4^"A)AB " T4$EODNo(B"Bw@xB~֬Ꮚ'66Uc~jXJݥLv)U5ykzWk3- 80l|izK W֓GbTx\Ʋ=#wd}xu5"&,(h)ItvѹbB+(YVƌ=NVАy+n-X#g8EG$zM0cZl;$y.4JQ#e ;ƃj[!2aNi.(CS0$@aml 8q%I7tpE]wWWGix~bXhopҙ~Z:8yb89 CA%p"d 4L&jţ1oT/:~P(/ݛJM2x71_Uϓ*xh 6ȈV*h"Er3?; vRPq NPsJ?.c6mԟ QDA9ClP Tl0Fp)GCDQNM55mcH $'^P(1&ŨqrDy/D<$X*֐5Wח}>2 $$$0Nd ,w@ ;9ųL&OQ~zb *;g]]Y_uǝN~YCoyLS[ӕK h% Ug i_v. H~:N~k0_3.Z 3fю46$ULa#(RD /`yrºW{$hM];K׿7YSK-V%i=!و Prq4(7bIY`\FlQ~4,x۹%o]œgg˫; m-*AӀu"Mnp#^Qm Ν6ML^Kkrɓߌ/^t a~ Rw~1[n}8 dFun?BV4ciM tNj-maO؃,>o]kdejZWayr`Y(/FrO2f;ɰqJ3sJ k &߻@uÛ?&>|?q{`=N U"0%GU/W{VUUnQ:[=-ꕬ&BfdgSni@rA7C;ye>k"#e+ tBrPERT|e(!Dyb4ÑAR\/2p0eZ~tכ؉#h5#y,C%tS*E$R-޶J&L 1Y$Sh $%[܋ מH~{:HcRN ,1>b-A3CRDֱөeOg7W{oLw]gM'3WUs`&GNla:,qc;9YuDsظHry9 $TLj|'nSAW*.:kuݠg,NȦWέ*sT }+p$P"SswLX v?u? #OuvOkgg 99=YihEhrޤ<19wNe-lߪJ-:DJmIUm2|?'Sv:׽ep>̶Y|72wji`mW,WΛI8킝u 3,mjh'믤v7R>9!|2G2Qso| S:Su8߽=ݪ* hCPMjם;drN͎.>%%{y:|1'X2ns%w3UC?w e_~ytN|d=Kl{;<To>S2bs91) ԆiuPkҖNNJiЕP(֘+1WB_ dm^x%DJ~q%VGwI!cS @}KiՒ–n$ծ8_7#.N-1< 1FV{Z_r~ϫ ˧-°9 -jxa]pG[woj :# I1̫Ji˽5+3asTL1ֳj AJS ĩve80(3Y`jDzM|V6q|e[GL)StYXcSA&E#R2dA೮$ԣz,ҏ Qlv8 z6ͪw&ʹ? 7ppHM0{{+ؗs$esEJs0oe3?5+yl%6 fqAdHt-kt":<ԳgVy#ddDuDe2cHGB N3u lCӠW:fmI<"gU5bͥm1l/n1wTqX㌥.ΨOfͺɧ?FiA) 4CA]QT-iKޞ%dȴČ&ƴچ0XL&9K6:(ȼ,kn\}Xt=dh:q-(C1֥am KR"wƘ x46Cb&Q&)V%Q\eu=2m]sEmH2(q|ɉCBOJH0&++j5qn?qXh4 QɖQX9=ɔ%{ZZ-G[^(:MNsK(OP ($Ӟ Cx`&UEFet1{ʌ2$p"T F꘭ѥhrFST\\R8pU[bbŒ$eZ8]g ?Ɠ'؞Q6DV%%HfnFf9T 'kr غF' {.  6A+Zب3vȔ@ЪRfAʈ]Mql jWmAmաv`7J1'-#&n5WD Y9ƈV{`Z)n,1%o$Gr$+:@"-dT0Aq$*i)|2 qڹs3N7:ckxl-"ʈ;DqcWiJD"H.6YoTɍ;a$3FR\ݴpƔ&ERIڬ4Q3>`0dB Kv8UFj܌>$z՞pqԦjd[\qt\%"fA:tq0]̧*p/lH!G<wmj-xh@XŢ{ mrQ0`K9U;axC=7 nE^[¬D\X",QkNy=b]87G&P-B:fS-ɁÇ[T{ӭ -A-6hzho3|\s?aās!)++u{-w E:)h1rP$x2rK"0$ PK!,BJCϒ3%,!n:̙*ejFgo?jBnNewo!+ ^xa8N'tfJ!"@,;6&M )yw0~wU{֮iR_/yC]l+^xZFR}˕ˌԋ {<_w1 18ce،t×G~~/{ݑO{ZzG8~=_/B&sO'_]/6-7?9/kM{M~O'tIZcwwr?_J"og4~?V),ϵ"b?iin:̎=rig4<;]ٴ`e~~`)?;Oz}_<V2~kϢkNWf^o@p} $hʀWoGo.!Uq)=@{ tNOK-/:]%#Z3cq!y,Gثwlܛ_.kBO;|8g~k~O'9dڇyOZ<՛-v+w矛csFP8~̘(07Wݏ7H47WZ|vj6pmvwp::8G w,Zrnif7I2Ń`R0 c.i<d=ǤK!0̝E[$H0n7yzP+5<iyaБMRSs) PSl}꽙mzJ4zy;ME-f8h0+@7?f3g#W zjl˃lt4H<__wWo|gY/!/|B%B1Lƃp -Y$)ʺs`l\^U}g3@dVBV4L21f"DSdW"Juh[0qT^4 J7#|1~rCcq|s[f 3hJU#0Qkh(ct"\pH%VQ޹thnk_V<\?nM|ŮV(v13C}3LãdDMy75pFb]HɊ7lRFICg(iy>>ݺ||=EgD2q(()Am򊃗ѐi1chdvNjMFnJ}0:e`)*ȵZ6&eeݛ],b q&fp06!k ̌>_C%G%4}7HFԕB M?*YåB'&.7g*e :""+mln%K}"DY)L29ptdB ^$=2Vn!.BdWK]:q,O`xgAC\4%q8I7u,u PYM9 )jޛw UrQ}< EMP<`( 7cQ9:(XD %:445$)BbswP+}+UA^?RJ%D> ^WE=DswJSuԹ)tw7d2QjYkp@7Պ"?}j0.frcwߑ,_kHRw޿Rrxr1Oiݓ,vzO޹O+.bZ89C.ʕI\1z 64UwRJ`@r:gqm0YDB]򪴗ȕ6 "`E0rE]$WLkM)+z\'tHrEV4E4huAj/Wݑ+/" "QG+Ƶx2wm+OIr wlz P7/WuQݰ\ '2ԢЛz2gkroKpN$W lÑ+*"ZvbJ\uPx0/bỗG%pۇnﴚ:_{EVTVthS z$aP# 7Ok(0  +](v65. J|@rEF`qO\rŔrAB07e-G+4UJ:\k$\+u(rakQBv<$reЂ䊀6`y1KQb%Uwj)eH 'vŸJ"WLkZ 2cW]+љ2ثpAE\1mSQVJ{\yɺ 3ȸkl$SbjrJz( X|U-`M[WpMyWhhȺGZ!W}^ HB`q"WLkrER^:(WT?-G*ܠvjo S׳˜Ϯ.YJƃ;]ZՀIJ7_4 =4X7*[TtO7z8nq61_|cP;H[ 0ͼzp@L5@#\:NhƔASM+i HۧKA0%ӚSZ(mtH/؂ FWPh_|rŔUʢAeS2p;LzӢm\1\uP"bW0lJƵ$~.WLD/W+NH؇#W[Ͷi\1-dJ{bv(4Ol\j;۴3X7]CCI\(erekȕjߦO*\1Vm+lԺ^N"W9Ur0 S *i_zYD P!Eea0rB C+UrŔ.ʕFi\0J\1.DfZl\1eD\) I؆c]1ƺ"Z +l^N"WVbug劀A`qM0ֿ'k^#WNZgM@rE(bܦ7C^rŔJrAʄfQQD/6i_z!)]/W_\.6?a W6Pcא+վM +Ɲz'^+urEJ^(W HA#Pim:|EM߯R&^(Ӌ<>#]r j9D\,,,yٓޔIjHsqu 茨P7_|:z4[JX/66ox- ELwQy2+?Be9a\yDAn-ߨ_ʳXV67ac<Y.fBǗS|û;<{qZxk2$G}oTi'&JRm5'?yR7vstvMG*Ъ0vB+ :+4! *dF'"4ic1p^^\O3ҞxGY~{>M[Z"mFbLehBd24Ё>VC\DDj̳TASS)m8.c9'q &,m2[hA]luh.?V~f2!#K]< U \N7Wծu~M)>3PrxvNIxوR~'_vy]2??ץq+-2^~??jf6Q3'\q?7x}&lR$BbƯ*bH$S G zhYC菕 U\~e~ȓ0v^&Yaۻ}X/G_fٶTgطt yxtӔPfy64^6ݔn=z*}g8U֗ ~tzN&/Gޢ:;ӛk)|x^SuhQ ?\wXF/QVM?-T²ŠvYny%\##<%}bIFb(;q媮Oh[Z*6}+|[z-[YopvX𹷸ѱBWEӻb48wNY)Ӕb"g_-vnvo6xv5Hz >jƶEtb9~ A%Z:#="7vF'Fh y9YdXd8G%"IR2\"sdR>. (>0&IRLDU9t4ǒVܦX/^M*W{?D" @^|m=O8y^懻'4n( jyV .;,?fqyRlēq2U5bVC!-%պA$^J 0V{ħSnNF)p %imژқ>Im6 IT)Wl:$,F㢸pxZ'(E!&:O /oоݺx^5BXo3AN ,O J%QeMUnt.M uhȡ&G7B2R(RP%dP*^!@NeDžG7 RgIFEqa~yNnzu<9ȱ?a8}Nj(LH2 dKbascYj+Цdll-`A dv5A/`0W^w9aIJkt$Q&xg)#-}|םvd8Se+S}!OweyWյU]wfߦ&.IiYh4xdZ}ͧ-irۣ^m^s[]-㿻w'?0?ԗSM }>s/w՗U'v[qc̓Oy ̃1*+<:)+P9Ujħˎя+Iݹ=坩]S4ff.LC.Ջ-l@Gfb:;x .֧<ec#QbPӹUz_٠ٕ "6q@6WZZB!Ɔ ,^ 58\we\&q2d e"oi}fTbZ8yH'.9rlh8,kABfIWOFtbh<J{2mY8ݬMFO>{!Y$UnM1B'!ƪ=2kJ^}01]{S,[-EcB1Æb |ɓZ}g~:# :%j#cd#Y3 &v!J`|l)-lxeH*%P|Ȑ>1`#Lp<%tA 1:l&Αt@PAMWUixJABe7; o4_/n"ݫ&%O:&I#*'9(9--pkk`NqB)T̝_u`+vkެ^Q_iyMS/!]wR:lZ{8>p_kBU."ӊa946c_p,L PV"|#蛑,:+jJC0*flQь$_)fǩ[MMP)RA8ϯ="!:%CJRg.g,۔XRhCt&Ạ!5,/~8Y6?dΘd\fCrU?MѾUh/"\YHGQׯ8;u;(!|'M_<턁G;/y8^UB3nKkkDՕ ?\^?xs})0̦NWc^[//7~ Į< =}Սٍng7ו6ZOozybqtW%y]zVvl{]|Hذuq3_.i(6^9YV=Z {\56.OO/?X凷_zÏʿ}ozת/'fy;o/?0ǿL|~p6Gh5WT@xN+,6-k^7KRu[]WEi~d7>7^j<}B[.D_#gcCRk!2茯y?ڡސ(`-9T"Sl [/6l\^[ pv_{:Nب z2&"d)4p:8( 0c=&{:K<{uf:qcAʠ;YCvvXgέvU㱝9ђQy6H~,³漺݇ܪ]~yWy,]FQ#7.H=^umzJ*Y {hR#SO.:~dԣy{?߽=zw||ߵpx`~NKgmp_B]bu 4zc`T:BTgT ͳj!Kͦ*CTN~_,.ʮk־F}^)rO\1FOYz=y0 sL)H0W.(caP#cb뒲qg3q^@88Jۑ5vi:5;|ν[Q#n>tU؃h} Mg^]m+rSН 7][}',ͥb|ww_Vwos<\Ӵ4Ύz|n>-_s>ow{uHΫS3hx\=b}[O'tzlZn ]mɑյ<1vHm͟&6Ő\lX !X: >v)wp|^1%Hov# 1f1#Avhdl]S̑A埥`H\lIIGө Vh}#\rz$e&d=iM!|)EKqZ79FW:$ۄ;iE2<2ِ*sϗw7T'ůw kk6"GJd׵!dk9@e1$A rn]exH[v^ 0"'ƭd$SMRzHD6,,ⓊIۖDV6ɊLC]uBPAyr.msVxx"ME 5H;/{ECaBH,F%ہ̸hd~*EXlT +4Z/E׎?_{ L}xXyTHώ>PQNL`.zG j'q-bheuot$%!j켊b;ǬJ|#)O=t$EGRvIUVɫ~V!#0 y(N `T198Tu(#HN%CdxB-P0q̢'k3z lQJ8uS-l1Am2LHxDJ٨Z@lU cT2mcﻆdXJ(+$muBqo MJ(N+i0:ȩWúޮ<印D4Χ,}JBx_#WޕKP.rdޠ,ַv5VhŹZ[&0@@M1ۛlfGL Dlq,z9*Zl[sˎu{1h\|@?RU`M) .(Z EX;,E`VvѦ +n5J7 d0{<Ʌ(;P:"tɤeTʠ#?]_*ֳH摞C1H 69 bS@@b.E(=!!x (R^@5n|iקտ'^n\J]}q ?FOZm|d҈+hg=mJ]93g[VMsYAN̠Fh}18FνdiOdZsFI HU&)ZLF]+%E䤏d|MG".{@k@2ʻ0l|d Ș"C$|fcR$,x1Vknf}P_OOc mGtG=^xKNY-Ai@!y[qMQ2j@6+b[ֱ!Y2lX'5:P){ KV:M>gk֘8=SW/.7LKCcG8Ž\C#AZHU:S Wç,Q6ˤB]ko\7+|Jid_xY̧AՂZv1aԒu[|5#}uɪSȆ&\ \/x8w<;7|Mf?~ُhG8{c e1YJ`34 sP26┠}sU56B9xk#@#qMibSSCQ.RSqd097y*)NN[&m-glPlsZWx6 LWPk}qld.\ ^夳6SՐϠ5^Y|RpMp'}7 ZZxvܢʊK:X HRҤrXD.e1-y* ./?e{2Idwh˵d2-9fd`͜X'pp<|2;~/oE@o2=ڇ-· <ہ9~;GX|uԔCuʓU7uZ oԉ+1kcM6樢-E[O3O^ zgB/ yBBm:(Mtj\+S%!# D܄< O)uqiI|?3r<"+YjʙsI@`R(JY*R(!-&zZ Mġ(R4; ā?ȑ x}՛.U=ao*D˒kpI1-k!UM$]!VXZ%V㷚;V?]\SӬ~ţ\L-J߷ o~mj?|x-358:v=Ot~6gg/t8~q@We=Ug?>D3ﻃNO~5F<{}rǧ*ˋ͇nw?9!8x_knO)[.sG.kغI?2_}rG.[@B ֽ_|8pe:yˍ"΍~Kjh>Ӊӳ'|~s|xmt|aל}h 6r?O=O?f R{՞c',{9JʯfZnnx~hr~燳8M&g~;84-^8]Mx/&ŗ@߰rǞY¾cֲzw2{믎Î\1w89=|FޱIM@޽)^ ǔb`rGݞm)7\^}g!An6ӨmG-ɜphѢM& )HYƭH#sՑ q!/9SdxzI0K= P#aZW<|a֑Kߥ؄%%NImAZ9eN>6k)sz^^l9p?V7 =V=rKB84U*p jMvgG~0\ڭxwgVZ?5R[e̵WXYZѽ)C$<1&Q 62. .rnkDi}ś GU^UiLI99UJrŠs$t~ ׵:bØx QkNҩP)$8:AOhkc!-4dmG;I!x 6{KJ56zmP*DFnU{ 5KA𼆾{ ,@ X0,`V<46֮qs6 χVx"f#nC3!vT}1+cOpPRp 3ˁ *Q kq)9nkLv/ uR A}2d*SR`Hq 92rܠP5pϣl_i,=*)E8ppvM'5HTTDpr":RD ǁ"3 YI uWH&PGwXj9Ӥa"7ȧ}rim+ RYD|]1&@TX}uyL !'acQs yP$PidLȼ&# g,TɂN-kAU>D<je53Dw <}XAׇyalUq'sU Yr01C 4o ߁]j$gA/F9ІHT蠻%_>JhMjA׀R")e`{{GV@ns)A2PNJuX:t kT(d}!c&$VN}< tg e#8c䇍E1pQAd$*>P*.ArQ+y㠈YSj1zP&JBTRwUτUSH!cK9c!lK+Y?XGHGmY4g&8р(:BJM|$BExjƂZFFUVP¢(j[(Y SŐ-؞`MNr#[FᬮWR-_.u+doN@0u 6k(9 4#p(6t,aIeHu7^^'6 1W(RV9 $y=ܲ4cPfB4}GD/l%V4B88Bi* G R3rT -lZr䋖oS5>T2'Qs\m!X7Ϧ$J/KC&"ˡ/TBVؤ53"`aĢhI!m*?u+"Ht)E{ T T>6?+NYd 3rJf AB?e4`ةZ.їڞ,{fLfZzFZYX7+i[/pЋwZԯW{H@0éZNlx0@ J XS2U}E@0N@S%T 4UM@S%T 4UM@S%T 4UM@S%T 4UM@S%T 4UM@S%T 4UM@S%T 4UM@S%T 4UM@S%T 4UM@S%T 4UM@S%T 4UM@S%[ [C?J イ֚p+`%kyM@S%wmh}.`wyXL0I^2 x%G cjԶU&Ǯ Ha@B': N t@B': N t@B': N t@B': N t@B': N t@B':rv piw@-mw  @>JJ N t@B': N t@B': N t@B': N t@B': N t@B': N t@:_'cPt pUw@֨;:G'PB': N t@B': N t@B': N t@B': N t@B': N t@B': N t˥@/.޻_eiu>//;'u0N0%qI:d\;%@eK64.qYуPCt;CW]+@(i;]!]!]aتv ZD?yte%K`yg+thk}0XPl=䂀W'tg7Վp9)]n(9m%ٵHW=UcWv \BWVH]R(3+lA)wNbM0ѧލl*y?f5**\ݻՄb6M0V주s9v5WyʼSNr{a( yߛ|_)˝O2ϫ7Z_)rϯ?wZ CgV*Z9yeA8dUk5__O~_֏̹*(֫k+ UW`&zP $ޯ|7MgF( U0uaZBWwoř P=H7ʶ]7Ơn8Cvx1wU+:p3+VCt WWخUAt骠4 Jik [*p ]I v JnΐVv),mw J֙VUA5ҕ.{Of;C? Z`T ֭tG]\N:j/(9A:CF2ҥ]Zw Zzu(9šo+^c8?nͩ%' FW-+]Q}Zj]Zt \BWv*(C:C*qKd4H Tšj*#9iQ];$c7Xh2퇕G#ܤӴ- eYO *Y}S$nw_Km֠=A4cH$DED *`^'sP) ,Jyƹ,d{(*I 0'-=Fߑ Iw,m38Ϊʺrc`B:VdEڣJ+XmmC\l1] Z1rݥ',( ]w ZnNWHWgHWp*I 3ҝivfZ+(D:CR7ȡt ]t cR!]#]iR0]3 JKNWR*+CնCt%;cWmR3+K-CtU Zv S}`'`0{eѳ ˜no7RvN@U:x+_GمDu +,VlQP}9Ky@an9>slY_U?b]\3=~Bn3m n;zW/\?^xc 4T>> /N=r-?~ߧ幝_S6)b?%MqYfg~Jmyǿ-S ~54,H*42Q*Q&g[yUjr/Q,zi "d9Y2?B+"9IbzA\ j`9ߓǤ11HgCutOVz?B:0/޿f`F5Mo\No..8IbfnR>i:IaSaS RqLQM):zVkaY7҅GnF|؀5TJd|"l߸˕J(d#dm{# ٽ4/d]FDdcC&, ʩtI$t&x􂀌UԴ Q!<CqBf22r@@2,95(ebyYl=.Twѧ-'h ؿr1_B_O;PGؔ}b#eDfTĨg1N\pncPdUjHL3iS-2W!FS&Ʋ0RP3 3%6Tle' FRu 'sik2/㵋^T hABt1t=U>Yuq Ʈ']55"5jH50R=R}оL>}jZ!KdMӛVp{dMF i15G+l{/outh[x6^]L?:Kx[Q~'WfqN:obBL˅rTa"ZB+n2R𠢷D=$-A^Ԙ531;oD!OPS\U/fK>eK:[qI%8)Y \LS3)Q+Қ)=l:*3_5K-Jj|\m6XLa$Ľ/.H|)Y;Qr&.%\hxq{c89IĤ-[y4GKhx18lڢq9@-(H"ٗ(U_^O'ɗSˤ_3eIZKe|Ya"oJ 1ko٠C$.<$بe:hHo V;g$(σ#6At@RdJ3D 2Q 7@ K  ѩ'F%rIl (p: mچOD$9d,iJ@Z&?/&Zo?~O$63I$" ㄥѹd!!kP"SKE,ӐCL4jWF*R=[{:6M-{ >Z 5wr;߶OZ <'_6|6_e0=U_ qbm,ˍGR'+CDrJQ4`=kx{$Y y$V)&yVSiɘspV*Bav&y:XJ kFY3 8.aG,|PKr^-(ݾtu[OO?-h=W4S 5ͳc %%Z[2F2AxSIbm۷v!禎ڷ,^w55٨b<ؠo19Q|Sxx h|ٿ3ń"Cy@Tt"XU1cx%wU$ObTsHT*Qoը7jɂHNJ$D $D:^iD\+yTA9\ύ #.%eKu),Y[UYd% 5Ƅ̼/Ӏ禅Qcp^FYM If&P3^JX6RM}tn8ۉEX͝ JI(NIU+;}'ՂJI wRɵO-R{~ƋPY6LB:c*Ҹɸ+!A')à N;m QpCrXb:q+jX& 褕Bi.$36DJ;Ǎ4ᣱA'-Qy "dꘘ756׶cb7necy b7m>Qb܉9i9GSGﮇ/c$qէhY:HL*ι 4Revr Leď>p n -xx4m^ζؔxaӻ1EX~%D%m*{?{Wƍ  GRU>8sT9T~up4d)hSu I#UDt7…2dmJ \U)i{˱rL!:VG 5/6MrDFT4:y`JNHThe6XS.C> əeNg&&uFΎ";3dzd؆xX7霦qR/ZKl;trZlzQ*nͲYbb޽tI:k;Ԩmr@US_Km}@dOPZ7_ZW֛[:]lɌal6RˊZV nݽM{;_cZPW7o\>.~OkO!ڊ5lkƶ8GeiG`eu˭6>A},/u}( 뱞vQ:e@*h+OsT9yoE҂t X<M92Fʥ:1c h&c ' ]]l ȵJ燗'= z_x| LهD2V䑀ՏU>JLy/]%U{쬖eW>z Ys3T+]A2=G<$itXfy<(BTpn3M~ᕏVڴ/yRl.~iѹS},hq):A6 9"clP1X)!2PFb9\[RL:}Qy]2%][8jVY0ε@b^>WGq6j5ב4u+䢾e?a9QIvX9'zq)Ɨ p$ՁfA8|C2x;85>FuKv28 !C1.8$i:uatqPHUkV3@\#IaHVUP&ZntA6Nj1Džp<+s߮'LO3ٝ:|oHӢym_/?9::N$q;YbvNhe7 J+SphgsZ'gǗ0<vpek>6[Tί\xq>{:0I18's+ܮm ƓO&%jrG8[G:G:mF#s06x'2eGˉ/gQ.e }yIq`27:6t'~w6/p/_W7/o?x {o~xA+_hfKn6d~$pc_kC0.z1;E2Mqma\-Iv/a=,0؜xU,a\E] -o!IO ? .h:U͊xbrQH;>h^2ߖU9S+F*ğ@b'l#ER!0OFْ\RļUȒV[WжSF\b=S$ ,g/{so6,U^Hq{m:d>3p.JQ2aA" L3g}`O[{|oXջ<ێ$lPBqOF!ڧRFbL{B%}/`=B} tE* G"!TY@T9 uy9m<$"1tF 9 "sUq'D AcgU&c=r|C*GtDf#AM*vqYG+ #0J5I@vUVЋ#B !1B*i-pC$w]@+r *5u kt?٣˞|<ԟ^iR`0.ٔ3h|IEMȤPd !&b|#U[S!-Q%y hU&eDl &͵+Tך3rkJg3θ.BՅW 2>u\} *?M>k1Ovɪ$-aRw[g9N *9I7y V\F$?b1|L (jSj5 L&ێ\IԓY4t]]c8;}ں=݈STAr2Ǽr,CH)(4(8$3@ 4Tf )8LȐYQ"I 2ҰFMKC*,avvFv}%xl(ƽ]}5X#^#q#k%Xf-9H BR1˽D"ϝjDg9cڐ iT$.ke2 r3!h5jn$ɒ.u=Ic,nkٮ?CzՁb,`O{q}X/^/zqǓ(3+dA:9#%-deOe%tA1/<^/>^}w>tӇ:$UMNpx?*<7bޏ㵒xP*eTBP9-TYrxj'ľ,2G3ױ9;:y䏋fQڧ .XhZEq:[3}?,i%>ƺj*+~#So38C$AR9å(,ZF-G!M$Xc+ :^#@ᖘ,pa΁dS S"uٮ!}z4GWK񇚐||[}7%зj"j1hyjP?sù+SP3S2HrRyq($c9.9%OGgr6{zӪzz4O6-fn-Y@rِx"&zB[ ٓ1$#$uy:<8*|r-JoP9[G"2gT(GI 2 =ϊfJ)YBRߙeݽFXBt Fԝ@Qז0l'.xE]Ѕ8{j\zu BDcId+0+7:4 [` 18&Ы{v5M9Ge+ kEDj:^UZn5Eeo߮:.c9hiCdTe |y/g$NFgqzysN51>y;zHLNsvV}wo| 1Wo^E9|F`p/~[6 OinIΟGG%kz>o)_tYFx,!s_zlR\t% yM`Ezl0 #a?Ga I7,gϙ?iz_]:_V]rv 9PJJ+:*N/ӌS8CLVmxHQ򤪽0IeE-Zmg-ΨMMx=z&vSY Fb@mŀMwn61 ]RݕZröEWG$AJcZf%#M`*rԒ+BB&D3CzϼH/@yaN0I8ZgQ6d>>7.3)hfDګƾdoFnq'~P*a7(Gky8 `*2yO\t>Wb|JRB]<ڟ"0=8w&S-BM҄"p6s"0MsN.(0!EJ@iXDmXnڢsB8>6TvF.H-ڶJ@0Pf YeXS8Gʠk";ecomQ36;M g [E,8' #Ag.Y&1bXɥFj2 !I%iެTj(7O)ı)&8 nحn nrWp#{:U  J-3Qp=Y'H\&c".aF9)(ap) qۡժvdnZ`!:DWU;CWj tutA=T5 96G)h"{5`kdž_pZEǃo e ~*J:$@y@D+)ʠ?.R4\Zءyyu^-Qgeh>RɸrPɲ([6RqHASd P3=b(y]PGEюޅŇՑ+mlk֔-5mp?\Y=wYN~Z:dEwVpUg"h"v[-ĬV`! ?i;ZtF+tђτEmOAJ"Dݱ"\BW e Pbτ"])R~p3qѪ;Tp t.Uug RBW-m}1d'CWf#U%1Έg09+xvR%T jCX3((dQw!|7hNřYy[Uo*BYR%O5(XIJϒ>Q|4T^l\HOqjk1W6F;Ԃ׌}Ee_mk?YM77* 7e0Lf2WɊ񸝻T m&SR &_aӣ0@~l(mۃ͍*]@tc ¬l,Vmh ?]9]Ε%_0GIHƠ74hP^xQ|mXu0zBr:N3Gn;*$mHI6vJ^C!D)Q!horY;[K6uRk9䑆;U99!pakE+~ 5͚9͘Qj)udz,ҲKU:ijwEm>e;A-mzC)[ķ|}Ӯ]Tt0A\w-Ѩt(Yt tE)r?^>zJ0cDzeTŊU6,aݕɴO'ʑ}';(g˵UU '1PYY=w8S,stz^:hpW92#34g%\f=eR2n`R:H<C<f\aXtpig"Z.EB~!a ۻ;T%# 7n^֔H*tH+:C.]!鈶$QsIҌbg'KW,Ig*e+tJvQOWCW:DWSp ]EUD){J)]`EEg*ehOW+I%5~ pMF33G9Awy180cx|6եӋIzYg%ۗ/NSZR- k:^l6sE6앞WL6,_p˔k=^ѕ>T}@e4]pYJaUZ%#z "0^@΀Dix\tyQM҄ 6<$_DrU$%РUr Yi%s/w9?qxFX6ek˜̤v)_2D|(;Y}CL\_9Cၱ,F1SS^ [<%_oqICʟ,^~T~b)+_x3x[7=x/z瀢|kBǒ g|C~?ga3&ifZl͒`6j%L^ft:<>:8ϨBHIe @5X"G2D@7&z5gBDv: Srnǐѥ΀"\OYzP /7 C!f2Q$#j EHX 2ݭ#ߗu4:[DyF~ҀI%7idSfM$EZe71CDM"`&C"0 ,C!f`{ 62sٻ߶rc[q{l> {m@ - cFwxlؖqCCoYбt 1Rim#@YcbUX@DQ>UEu :#g?$ هAS9fʜy1 A*ځg :@A% ڼ 煠OUNgC:djd:'՚$LtG'+axU%t$@lUE *m&y|6L+x]$dʋ1U~EhHۢ}m_oݻŭF3wl4moC.75P' 7S_x87Q5apkNփyZ.ϟ>jv:0O1p\FOgxkΑx2UUOV;Am#y:#oFmFKuuf6xɼK'r7cχjJu9mn+yvifRF珓6Քj6NLyS+aGyΆA^1;?姏?G?;xfHP$Qpo?CiG-f=`h]zkUS2}A!V0V0i'#Zxԃ]sB"Ƀ b~p]T.U-ys>Jl7-d2muY;hl$ ^xIr1]9Ho=Vْ"fA%1Œ J ٰTyM+=>F̥;۪Du"?ʞ9imx;p6her2HɊH /FSF?LgJGo{Yk~3g{m`oڜ=dg.o$W mot5xΤkXՇL؎֛'X~@GTWa~D A^N~xX]Ojo ggq-@keGOc"Edo=(v *ֲZU<rF L-8 nsݻW?7OJ&~ HKnY:&,[>Wu6l*J.s>PJ &2!@[:Pւ S cYPE**5IM‡Pv9!]Ɛj^SGtbCrۂn'?or޴B3܄ټb{l(#YNQ0łF+DJ"rq>"oւu&%o*TJpBiJ1%V1@'4)%' R&YXJ9%x#zr|9ʶճW+_OMy1[+bS~vE{!GbyYW#~|:W">Kqd}CF zi+,J{˱r1DV!$_B . DͶI|Mor1兏^"Q u)Ĉb<K0(ԁ1uQَqggh9u[dz6(mq\8.^o??ÕxsΛJޫhm /=]/vû\>yն.D$m:I\ vw]{'LR1ܺkfzwNWm$ݫv-n,nw-~o|q v@k-wC5|۩s3Cx-îvޜ̿r-zwf5_6=ZnUnΧc{с}s1@X;h ibL%l$SRoyo} ReLJ$-s`9l BI䔅  oy - Fv+ځ\M z~x9S/^2[áb JH=daUsGu6dLJ#<&VkHa37WgeeNt!%R2F¨zR5oʢ!K; )e}' #Ӈ5߰rtLZ`]Ie:B3r#C)$=/B[ZjGd{5u;*O7wT7o kǂ"JhZbZ:,*q!(4)fY%pmUIw[c9"V$6a&SMRzXd,XR*N*m) JNieQHfdjPB9ڦ |D[Ic9댜RoǥX!d.Zј%ˁ̸"`d~*EX\2 r/]oo=_{ LyX=RA`!0 julIt'f..sΦXfwVaY;DQ]()' BW%3,AqT;7z;XGlӚ&EDdIQtQL~ 3f,2 =#´Z 4FNE:c t,|ˣ#QL9M( =RъM:!E^hM*^+^$FыGwc qCt,{t^M/E/D&Mdu ThD;Y$&/lgjn s6O|QHJ2ICй*hXF%!^($#!)BRt0M^E@Q!#0 y(N `T!:T6i(HNE00UkTM`B IKuRRo %x$ IRɐ5j!̷& z(N+irp,S!S.-nꍰD0$} V1BHD+JPL.eAr2F31ދ{'`b'%7X̐WLg' #,Yb,#KM?Re_vP hkU7v3w;HOV_t2e3~{ˏyƛPl>S Բ! $&bS[U Tt o}zEϺkws=%mRfXe6 P&!L" =`1%κwv@ oFu[ytbcl ysX[,zyvԳXYCC0jSY`ڔ h 5.@PѼ3(*1$}ɭ8#m#HAtܪ^"!\ST"S9t-r>zEAwz&'ޮȝ#tk?AyEis;L")h>1,dTq'Q&YlVmpx(-K@Dx]/LA%" ShAh 8&cA팜?2%RA’;%CaǓߞJ -#dR0\Z3JbI?mVF(bM4̌d:ՌֻsQTL O޸\0(R 8VZ4 +5cgVSWP]h;օׅO o%]UdfၥӴ͹y 0MQRr(Lتd-(;`rJ5)X`kD Y ҪMm]-d0¶`0 ЖDE$Ψ95vy,Z38m{#؝8d ds,baZ6_ASr!ͣ*n\88rS֕L= ,R JWٖИ>sz T<@<׊B`D1,FO@6" LiA-ddЊ<24*=>i/8"F8oY<`Ĺ+>NCUÈ:Dq팗p7<T}M4N˜l2Ig\M-* m6єPd&IEyPReTGK:>;5s;"~zqxVM{͒Mq4&Ẃ<@BpB8'Z'Q'I0cGh .ۂ͎ wa *ދTuNp#y?*8̽jrgCeHb(zF?<cvF?_Ы=M&ǽxG;~{+N`tF\dz7b7^l'-3v'dp,?p#VWzVс+A?W}aYJUaCv:*^OUWW$>f5;[Lq\)_"#ߨ5 l{=^tQX\.FsQGQQSl>ZMw|/?_ZRtD])4Q-PɔjbѺcEL1ͻU}}҃dKF.eN *$UR41 D%*&x1q]6 &eXGhk1 or4&}G?/EM57t[Y7|NfևN@;k_-76ݗmOA~=cPCPiN'K42<E.)uDs}:`x:x.Fl!pRCQNE<$ jAzي=i~jR TNΤ Q娀NO[#TpyVߞyWgݟY^oMCKI TAJJ! &?8_>ٗN-_}Q>ڝ`trxvwqߛ|/i~hvg/h_E?]˟~xƳOy'/$۳m?KBȎN//y%q!.{-:;o%dzFWs4. PO~BRIqG²tǏ2}LMILj9{yUƟ_ߒFՁ8|iɑ/ |5Ͽz};:A,N=7[JeWunwAD3;Ή`H T("r(9L'j  |3yoӨm <iyqG 0F9kFb.Y'P2>^&ҁ^#gO'8N*KƛmI4.6/ /l=VqfkRumqBB WGX,j;:\uf 6Z1-R EJܔ+1~9iu.#;y#v,iK eYԘԊ((*jPLL!`7@@x P!r޹PHQxvk鲖#SH#Žƃjǹ2D&eeh 35&Ch" +goOy^U. 1o}+uTJHDZn%3xAs  oB,}[HᴦhX#qo 1kBt:Q:«n[cWA.Ǘ})K8 gh}FO,^(t,.Db?oZ=L2#U@;Y=T){9 bMAZR0IC2h O㚛 _:41oF=-A7bF͔^v�JH4]f&Lg]w5<11F?<1'\/jQ {W{ܛ&M{ȏv"aW⡟qG[ +~=@e;{_vJՋՄVUNJfw4Y2I8JR+=+mMyYM\@EY@^~Wx& ^N,?{lfgau\)_}ˋoT$^Şg><(p6\w<*o/׆^qQpu/;Uq~8pu/)BJdW=LhGpK{WY\)PJEyWOfX_LNsR {1|o8׋~8ZĕE?~7`..?RUz*r_Y1>;ڬskC |#^w91._أK[*_YWjl-=Q|i7ډv! APSgh06 8\85:2 ꈫ[ˇސU/D7d)x \Ae{WY\n2YJS+0 i O2YaW(p\=EFR&P\FUUr9 QWO6\.$;! ЎyZJ^_x(9Șw_9{JS U N9 %C%55>j}= @qm^Hؗ[F/$?,p7?վq63\I˸ޥLriQkr-yI y~L 򫥭s^Op SgXps۳oS{8ž{a'6ݶ=eK2CU^}|sMnU"p'X$,*MT$ D JfZ^#֝0iY2s7uOjQUb~q|d>`/)ak? FE>Vs R/^ξI%z8!Y\B$YJ%:ɿU#JRIj @++"@*պ]-u]y-X:%"S㢦xCQ@,&fkDZS GPKwH1mu@ >ipTNS4&g_JG2gj +oך$Z3׵g/{Hn lv^vf/hBu؞ؙ pv$uRItHd"E^/R_9˭craK-)mxzt4+̚6sM>e&opN5DlbЈI"!c y.HdgDI05$=2jϠAh$9KHi-A)ʢh*)X N ;R-eU:_gDw1A"0Hvx3cΜ->cx^]”G/6/=y!SA< <Ō}C>ͲYaf~1u؀MaA%r&9A\5Cz,yߜ(mK4ɐWdDIGd|F^Cj'sDm?ۊߪ.Ti:ݧQ;ʭB4Mσ/g4̹]]a? n'sݻIq=ųxe{6;|mO6em" xe'iRCT5}Hr!u Z_D or}(9qQ[Vu_=Q^zutFWi]v!,kl|xElnIslԑBAG^jԶ>R(*Jh|!ֺٌRR}R!)ƀZb_W(\Jju%oUM3t;F+&\m=0HɺdŀEd@"KE0Vtn%8y cba2\%W\W(Hkt2*Pt79%x#zr|1˖͛kV2f5,we2 yE`)^OJ3$zx_T} .^w%I[Ƅ!lBԍl0;zGW:RGW ى@F!$WB 6{I ɚ0'wuLDLpIO#Js 1zH|J F0ɝskY 86(mkipQ}QӫAb<^o_?%}sٜΛJl^g잮]OOSv0;IMgWivkTInu{uh{lOmv;^Fø{6e-%M/z^kj{}7|3ss׷+-ofw::~9]d]jnr%JEA*y̚>wORs>iE#yj Yl0zĘ\#JPI>>?R 6ө dv9ٺmdf$/ %S4F#;Ͽs[T@Κ~x9S/^2[/¡bTMs%cU~,60HύG2 sbOY3J"ͧb 蠘vDvg-fСJV IAx BCT(LNFuQG]CabRځV" BWӲ1"62@{KZ˒1:쌜woAC.RVJ% F eJ\@6>X1^myvFONH26J2҄4!Dm4 $״JbSr؞Qfj!jW^MCy7;Y\YjQ3ꔧwyziw\ meji28 O3 ~j|wJ8A~I|U}>'ұ!P25^*'SqXNlFqtu2?ޭ'ct~>Q1Eu6ap랖յo[?W~j/p9?[~~'!K51p\FOgUfnל#dMzrG437pH7tm:v0 mOV=sx>9N b6{eԁ5&m &ePƩ8Sh$Qpo迿״#VC30.z9k[2}Гvea% > ~8u8K|Z(;ufdxXWU<ȫ0,0 -h] sB|k=/Axnߖʩ%sNho6|Ue^xJrLf]li9HghkkJ> *)QJPΆkZkt0AQ=: 1c§R*3)b"z,*Q$"攽st35߷:[~qvC@.l9``;7]v6>levt-4a;7nf; I:) *7Bn06}anD Jx@#~PPT#V} ϖ]U0"z 0j҂| jr4B !Zmm OR$a5}&X_c˔(IՄj@dʺ"g?[Wh4L{w d52 ~,=,c).i;n.k^OBk7 zAzI9IU ѨȠ1:#V$*`s2X£g]W$A[t%꓊K+gKPt6D;='>H]5tGFH@  k33 #,Yb9YT/cG$c;Re>woںnvr.OR7l2U͸6kQHuԐD, o>SEM !Բ!$|S5تJђsIr՗^CQR:eU*#/IE&!L"5z! YUr_9q9 fT}MK6汾.lp[Ȭכ[׏Xr P[B>q"ڤ 5ek:v0mcJ f/ FTbHH=>ӄiqt)妐G&GzUG*ZB-T@>7ޢȝ#l>ayE KZ L"zSHĐǰxzO5>y :YLrFmҰH-$# s(M ,ܱvF~AOQs.?먤)%]*o/Jk (Z].[%1~* Z("̌d{0ΖKzcg)Gȁ+lQAJZ3vF!patagP]h:օׅO oUx_Vd|R̮fiYy/4a4| g7dQ>bgFYtޢTNI 2Z)D TMlg[ض =YۊHu+rk85Xvgq,ZZG;q*X^Iz+M.'H`P$Ĝۨ4@٩>,0RNKd D"B`2aH5 1+y1V%;#g>lw1Tخvm{ǚlBaȂQ3*RZ8f#BT#z'0ɐf)XE(UC2dURbKЪҘj[3rk/gď:^3.y^E^/xU8(2CgoRh@%8aD`/2^/>^<}w>ӇO:,*T]Ap}D?z!QG?N~7>巭(/8ϒSdj| Mo'Y8IlbXc+Lp^#&(@Ë)j݊}bwkf7w}=Exl^pG /|#L9Pjr(Hީk((c1ʈ9큧%MfB=kG@'1 tF6Q!438]ɹCCӄZ\st}a\8Z -x"#|ۛGMyosmR-%K0R)Q*DbZ>~2hx|˻o8=?>/kqnZB=VB._|riN@2?P5XJHU~V \HXIć_kVs?TחD玿{蒒}Fվ|eG١暯߾ ϲ`'Ǩ/S'sjM&VɭxHukZF-/y,{0tt*u.wߑEΊt?c *_xߦ?8A?k# ȃd=˗9+5ګ`X" kKy1i3Nl4ŭot?tBrCmMVyaԁ$%Cz?d0oV*`yFv{b/;VY 28XFLB 7_)5c|mZBh|+Nb8oloUPee&e =r9uhTr3hPϟ 9uɞfk^r/4zKiWآQ5 ;+l,!1 _UbVֿKQבMrar>Ļ-NO\q5Ε|קBM'By100Ӂ~z`_Ꙓ(j|+o y}ƓFŻA4|`s#ukiu;{ro wަnjӚMt}P> /V]ܛwXLg"p[ n54#7ɺLtǾ(F RN3/"'QSDPq\9o|֠- o;;ӎRG j/x:>ᑩBrւŤ]NSo=Xe11}OV@/sO4*q7RƷ]I/6/wۢ^yW T}jJ"G:ʃyT~>ckIڮ}~WM^e?y¡Ⱦųv+Vyd+,V{##e~ѕUK%rYO^9,^5B/dEQE[!!޻NB\o[/x /|-UaCv:,-וQԝ7תJ'j~S,3pZ!xTOl"!j)`@n,"x2@\:{|qbz&")5nu:0TRж}iWeB$C\:9yt;'#Sp#Z'jkf<&F6 _v(sD)DP cFGQT,vfK>H+$䤎G>Cb E#U+I=pkfn;\C 퐣ͭ˕ ח˄הJ^yQ*nK混yE6IUg3oTEq]\FFAJKr!N-pƭFs8$ AC(Pф(*GJh(X!v2n)Kq& 1Ր*0tT.gIh :ENժR[YkQ*|FA"F$nE98ɨé\DK B/msOSJS.ͣ`DT#Л;A, +,7Vk z;툄|R: Dv7TʌD@wҺhqِ/6=pDh'6D`HT5;TCU3: .}.VrTU ΅ Q Eb{ik#ObCpS--TeǐA &N9.V[n,Y)#-/RDHM&i)h)@}u+Ν:=[:=" Q VKnqÓtr'F @dhUQ4@ZXh.P፧:c\k\x<҂F0eE5rR=.)ash{윯ܾٛr*~\S9$uuSkJL>*WR^I\vKhK AJ>j*c".lğMr "00VdEٍFٍhi_AF6>Xy)Ht(E(Tc%jjJv"Ma)Bo_ˋnQ?,>?޶E>#=$F! I;KTbr4:TɪRQwm0Ghu@_Rr6EEBmSf]vZ1Ӻzv>ۋl< Kv*`zUrh T6^FJR} KuӾP׌ZSq>4F?⊑q}|y f]LN[׭ڹ7idvI9^7oƼF6t3>wՈR:kŶMAZR0IC2h O㚛\ _W3l#Ɏs9፩5fVf/E]&0IZ7aV=9Xw̬h5%)W(csy }Cҙ- iyYQb%e cAt1*%"w \[/TN,qז.K9JQ#e .76V;E!2a.t{!>9whڥC7ϹՄ.|7%3iW>9%  SM],eK}NC J)N&.$:hW%-[ȹT3\V{a鐰眧h:"Mp$&z]V% (3(.*,2HĊXT‚vhtrvgF)"HHP%z/{J4˪xt%0 ]#h U 1צcN1K43O_^sȕO\~'v/*npSnN= !K3pU;WZ#WH WhG(8\!`g*ծUVURp2I 4? sRM{1F.}o0@qEGahצ0 'x_ƠOџqn| Nt4A_*k(pɃ0q8|=C;!P· %Ae 6x50sj ?Sؒ 㿿wI䜻!RƦڊe!O.C9i++ɐ{݂c7@A KH!)pLL096IKKs1&(As8I(Pz0b R 'j4'cq9v*yx|XYWz`q:"RSۛWd4D()+(9}\POo?|O7]2%}]-9RiL ;W\)v2Zt2FpJ0!B lej:\e*5•L]L#lW\3p*mL%=\@R0:ΪMs1ifKYGz,Q=$bGbUn🏏w^YiJ*)[:4)Q !ncK#͕JFqqaX. W2siX5"u?]ޠ]:߻(̟]sp_y6P=MGy=+ny yb9$0f'dQR''yT^G'wc:u(E o#fgHZ}=hjWFJړUoCnYW|7~ξN;jU\ =,Tguo#o uEQ^4z-]֩t8x}iV9pRRTiR0C^V}0%u}mTo\} p wj7p110xu3CD 鈭dUG -g7>_|cQ9 nWM964F 1G*K,=]YzGe^TKh!d#}(#d3ZJ}R!6YqzJeM:TܿjܫJ۷lv ]rG6;NJyHɺ`Fb"D%ZE&KE0Vt~-8yЅXmRy D&z% VJ lT%г&g;>+ŵk_?t *MSr 31K'ҏUmsG|Y@ڲu ie24ށn`r680=e9&/Hy/%BHlAe!+t& lФsF鄋u>յ*|>S[9 mX JxoSs[_[7ۻ9ELf2۟f!leݻxo|WG;|v}䯻:3حewt.EXs]owr_7ݵnMsSr?Rc/^en7&n5SK~H|jNz.䩑b&QrȆJ% |?S~. LөIAv9ٺ/dd]j{P9e!Hc4Bb\i R)(ZG[,Q'"x\*2F5V"{罉s7?g\Яdo-uԐ/xz_x| y 16V푁_!,t#}9]'Ql>&PLnؽd}3y?!C*Y3! QQ$Q{>ʚu=MOS=zrakc8I绪 9%s06%]):UXejS1~E&M 5n+#l٢=uNh'[7s$ԛG:FvvsUY޳fZO>:Ϧ_9>/ܜݣJ}Mvu억u^im#u`i\lNYUq0-WVpKuQ:ǟ嗟~Ƿ?}|?>-8'Ts;t ݃EǃP{C_[*a\#qobܖ̎ŔRF}{;ӌ?>7;>yIr9k2#WU4+,6E9ꢕ72/B:έ qC݀_n|n%Dr;R;hvs$xܝ]sď$ AQFeNqm)),&Sl ԋ k״}\^{&nwӱE\VbO,`Kp1[=9e {:iqv@̖ ĵu*Vb%:59sb{6;unt A#%f+QtWKI]vEW)PGL]_Tex*ܬ:<ä rd fѦ/̍R8p!)pϏ+q=:Î i D'#zĨI f(V&WI 6l2Dm;LDVYM!`} uD)(ĢGjBt,le}s7G]&tQM?fVZYQԴ*Rl= .k'=`Tk2F={ ? d d+Dofxيbe |jzseia \ .GS+ΖCl^Rۮ=QQ#!5(NXClo&y~b*[`x'<d`'砨Nicu}x hbbbа{Lq kƵˏiS#$Ծ?O}5)L!`"U%ђR߼rŗ^BDj()2*c0YeTL mXTv&YUP9q5fT=']呜Xߖ]1})VzlW 0PQg@Kh&mۥ ̛rdLM*5m J iv?fVC-.Q*؋>dg$0D"`r sP\pX\03:U:\-' {jOX^P~‡F B9ܣ(.11lޓ\dC-dznlFd0eVI|L8[r.1mK6e]),eF3&!pa|aoq__hzz ZUd|YkiUy_6hd2:cGd(1d/Qg(<)yK /TN'd/d)TR6`4vDOk"u}K=vM/^7m= ح8Ed0 ڋ1#`2dERF%D#@â #4 3dF̢"Bɤ2%j0b.:WcK&%5t:ƽ}}=##q57*F* 3*&APWa,!V英EY9&dȀURb&-Ѫ֘j[7q__ud8__ݧ޴~MW UJhJ8a+}JhiD?HB_| /7XElJۂ~vG?ξLc?_[ciQN_?q%>F; F6HkKuy޻<>/k<$^+Kzyf#2R2ГъXf:ALUƞ6Xt2l *PZ(p1S( ׽Bi H%\& BgЛ8;yƐٷvRڧ.Xd;dn,q-r38͛cCuoM_:<6_tT[6L1Kzõd,F%T61BT&8S#T]^ A"5;$,ސ"xf`Ar}0g10`ܒrFKEd.s|VIƩAE4RNJPceώ7qv Ә>yGWkr9#t[mbob{7ڻ%j-75q?/:aE[Sy"H6V)CAR ap<=9( y\8*3 %Bl^MA %i&$TY (()3xgj{H_]cuwU0do[$nsWY0-*$enᐢdQPƀlӜyzzy¤; VF;8{fK:D`I2F' Y"(<8%(R3e}#U-2хP@um]Rّn} PϧsZ* Z9Q i{T"q7x Krф$%MH\O#4?8 k|Rʳj~q>4izn  ,:Ϳҫp[i M `͛.d-#(鲩 Ys ӢP-˚VnQKu=EZW>S~Pc}9Վ ӣ޷gw!Os\ 6'1X9V߃n u/O0y?tz?=!-yM?w}2.cC㱾X̟ /fNާChM'FN.vt5f_[6nRQVOG"76q7mXrea1'26{tѲa]m{7ݴonrm}n7w&dI`LA\x1L{IBF` ;K>[l{3d^PGgy>^БMRSqCO$tffSBl$NHyZ\_&oWs/9x; {Uc(XgF) ce~\MT-C)Lg͹re\mi~O\,S6^gz?җeJboe?z߿Z{ZգP*xߟ$ 2ŀ59vE7r,ٷo 2<m0-}ЄEQk"5LHжm']K^VtL-P 7lwӯwU{>|}hT˛[\6I)VX jYAY\.K}G~jSq=sGU&ګںݺM/ϪV7wKL5L;)zlrѸM ˠ,̖q8eNr TTv:]8lNmIqͬ"07x\Qxߜuf$^k%'ibeF)DB2 YO,cV-ZNFDNi(Qdul3wg6ġ}N:NOE&[6*QGu=BioTv*A+XpW@ {EEXtxZ57FLr{ @N$x34PIo(0X:٘d&UBHH9/\ 0p,;B`k(0Ik@ 4L6)iO2GYgpvFxb:rz`x']D1c 6rA&= <~B6 E`?5ft)__oz~_=4ʕ<6+,(䆈&xZ2hXGmFU3U+,&4IkGB!kcYtT3!" 4w\hPD/1#qtBDlN [O~Gk;U!U3yd>1O3B5 `) S@rZ5f4su#Bخg.o4)?9Tc[ 1lL1eaMt3J "ʐ z/F`xt'keW;!JGNrBMyYXlBM:hrqVF(^r+7|RwlriuRR0ejLjOW+΄ߒq+r!e 9AGݳI@dV)!!+\& sJ3AIީr+ZȕF7>t@8+q>}CWv˴rZ_MH_: a.>!.TӖA1-Dp>Sh-?eNdX; o'jo/Avɴh{[67<]`&[m{Ƙy@go,7Fwn4Ix~1_Ȱ)c\-翩.| #Y12Cm5Qj(sMH4=/{=`z=`:83"HZH8GgdbB6U^q224qn j#`0`7ɪ̽(N8 wgr /IQ< kjmwcmLBپ~(Skp]Nl8 0ހ,M%fWW͎y8WZ?U]6f}5qHR[Rz!c & \)([Sdzc'Rdҽ"DfcC.8LL:@R2/|45B^c۶mZiܯjUiϿoOwlQB|T,8ƨ1.e!HncPdԁ+2YآbMKkr='ͱn ̯)9lMTΩ2ڮٮtkWٹzBc?ezY,UZ 3kIarZ#IMz"Ku}MU;Sl/ȵVշŧ,ZqxWT12>l^o;tXzcx9YPBIbXW1ɜgdD 2Q Hgpfx(,=SϔIbS-:̐lYt4AcN,|S ١]?-4N'$63KI1qhyt. fn9/29ģ׉Wvj2릠c3&l-3ܱ/gƍgژ(xn{>s xn%Rm?٣kBK~v?؄>vA1}׳-n׋Iw`ͰWswb%Me|K/j o2{~ݮ{@XN`SW~j WP?_\]Zkqrۥ /tMyS \TS0ƢR}nm-nٻ }x +zGCόTfŲi")ߑ55' ƗB p \4CQ ZҪQmxjh&]`ñr \ņBWQ}+Bt iWb@tE%ӃˡUA w*(Qt Jj ]`l0tU" S/^UAGc5ҕҖ)9 "|@. ,hOv<]r xt8H3lDW.UA+tUP*;++c5Cr(2p \CQAc8++[φ]`wUVޟ bj9tewzYg~p3 Vgڻ%lAWvC[uN]á+BOW%HW(ńN]Zr:_\H)5<-ߢvP5|:I4? Ps@iYѱp#hsi+6|ΪںR>jʊ`璿6 "I>jF8c0zʜH1B PxQGM8׃lU 5;n8 E_'7'3S1@0_gB79r ؒCIhz#LMrdgTnlw!+ W!qanyw >.B9>Gr_8O;>%\NNDEH-쁛1YqvRR0٨S1Xo ׃jQ甙B3 ]HRUmjSUjt7xc%^UC0dh ܍+o]Us)V,&3Yn͉!z-@JvsG[%k`Dbnь}'+S4)|n-L4o w>BRMR֖q[EMw!0ИU֤{{6;[_ED^WUmxJ* b Zڋ#Q1'Bk+P/Ss.ޟ7'!ИUj5Ժs!DAu@r=RNZQ >]o50'HwSt-!vf8mI!a##k$3$ H/6G;'Xr!xQkU {KFC66WѓvzUIO̚)Vէ@}Zu)jԍMɀ1!KutAڳkJkCv:v4&G-/VL ]d0`g,a{PQAQxi0mLVjY @rܪCZ]]L͆Jgǚژ[]JP\Zlh 0ճk.keB`T /$Xe=Ҽ1 k9PѺ J(ڑkj o2 8K)O-` t[2 `,dc!]A l\;`AV:v%&CijdLB5Xvl:!L% Wћ+dW40w\͐jPoB+"X7(c=0d;ǁ9 qN0t AC\oS/%J uf9pu>VM*9oL&vr/ōbUڛk((Sѝ*E4GRFy֞eAQ 2#}@PS(H6vS]T eATݳ.RS+եa`rT^7Dr!*"pPB='Y,k x^! C9TU54-dPgm# ʇ~AC{ՋqcŬsDr2|bL±yUt"B!N&mvL.'Yʗ=!~TA7z|oM/ L} h"X 3 o b9PT8xigM:V%S0+:XF SQd4XQ|`QBgąs՜ HDN̫ fʇq_fd8 =h^= KH`Ge Y%cԭ)x $nGfp6*YȩՏUC}^nygUcۆj¨ZaBTD4X!wOg''ygͼ ȓd}ȅBOeL3fXbM=A#%DF$rm !/fsP6DRD :8Blk ) Fc*2["f5BeV]!۔%VBkB-`iڳ64MRU#MTf% MݢrɎ[oU5^"`!-##Z6HªL$` fP}AG `jU;R@4Xq0{ 7:r~n§dM{~|vڮ8׍$Ěj`֣;Vlfѓ2Euh0NvPIlf1hmTךB$KBq)'ޮ $hYeFIJaväDy آsE6G=Tʍڪ--hofr%5n 2TA3J`(HȺČ,mAz|BA3rwU߼Yov+P!>n;"h$G\k'F4$\sc~?"oQ^1 bp GYTPcD4wv:yT,~ xCڨKecT-x4ΘO+qҢƚ5AF %_64O5d, Q6#TK1.dP?MN{Bw *uz,qVkA\c땀;*އY[Tڠ@4XAapF9 ZfSQZش =W&= M F:`Q%#p"Z{j(=MGPrm⊑d 6vA< \T*N6fTS.\;oKC1ˡ옕\7"kN Zf !Kk|A9oۨBb!(92L0H5w[FO^|vv 1dmkQSP}Fqwh ںx k6Ѭj)FU۝_\V{ObnuV G&NNƎնI//5~Xvzq|`@Ɨ_o_sG_~x~iR/EW;8}_kp?|HUKpChktr6nzzaCp7G qKrr ƦŸ'J;4'stY2w.N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'Џt5[^Ή86@ 'sty}/N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'Џ$'l9N ~1N \  @yM@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $Ng& qINUq3qm X@ )9H@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 @_oGttzztl5lo776hwioݯӋG@Pܫ%`c\p}XqiܽPKzҏb\~CI4d.=5] 8<1]՞h,Dy#c]iuھW!COFhDW.\h#:]=0[CWP^ 'c܍ <|i)?~[{ao[c7v<~߆\{4/]cJ[h^_|xhןۮM7#dc{u#k.*qx[xTZEO7bF]K3U-M٨򺄠mqAN AbB^x6]1y)uLǹsע#DG؊GL/8vǰJn˩ܰJh:Jn$c%g-?f?BWՇNW?Kb޵c뿂VWZ]i ~4773iGMZivπBNzVqtS@7 $(ty|@bU"YVsWWJ[u +Uy)8\}1 }ǫD%UՕP0`/]%r$M%j]%*-%UWϢ`M/H]tuȽP;PxuUWߠJoKZ@UR[TWZ UW@sliJ]%rŨD:{uTzժKWW'] #v+cWTaT3; <@]V]=va05F]J:vju%箮D[TWk>)rNJ׮mw}R`HZ]z+J)$%%7UxmԚ]V#T¾v|qB@khһ 2'.4.auL]O8@&|2AၱiT dV0b]׼8 `1Ue^ede_|)ZoGo\Jm1o^PHYaU ;]o\?pλ[;tUܱx슱5mhm,l'4ҳm#^͛>ջW?T5 ]zVA& DPclPMk@'J;\VۤV28dpqvú(IU1F68R$#?g)Oc.*5X@Uvޞ4i^f*꫓<_M9[1.ߠ%g&YXZ~u`V_z1d?TߌKj{ŮӃfگcmt?*R>#붛MMuuBύ"8Eoa"d\SDa7mrِ ƹo*}#3Ae:63 B0Зx8B+W9xG΁Zg䀖IWtf8L(s# mwU')4hrzۭJuneYU5p TԟN.Ld_|;Ah6"U^M :|H_?9H =[@0܀=״Da|x )Xoqyÿ k>:} a\$cÜD^rÜQf YcI"|ZeD@"\H<7$E`YD!f=b+$",0*Ị@1CmlQ13ȹMXG9#1rbrX+_h$ CjsDqK;;nUx`5ўm7-9Zh8Y2\bNf5AEjiDƴgR_ugч^ʹlhf`t*-*%sAiL"l,7*drgDF* p 90à gEBfw-Y^G*$203,* NVc SE~,{< !a(NFy +? @F4D pz)PnYk֜BSOiz.t p,FDj6b,u`m-!d S!^h#eƵԬ X=LY9*j]:'ډCQG.&,(NDEK2sRS =5/dh :ָɛOZ }[v~,Ñ3@ oPE%l*%R3WL~ E'WAvEyW# ; "wi!F)uG d>)"1~R)pmb :[[TÀ4`Mr@Pvrq]oab}⠐O4,Uc/} )HY^]㍥݌wk躃JAu.U& DG_7F;)!_o~._M/ׁ}bɩ6_y>cKjlׂ#|2KJ<دaȿgյXmKoo麮R ufyBC>*`dGv9Ū`skIJVk9Z-R4;,e$5́}~MʔY:Ij3uRَJ? uǗ?c|^߼D/߿XLЭ#I(1ko/95pd޲i`^/vi5Pn\erK]ٗaP<}9b#-&xă}/d_WnQEIOwLf]R(.-@K`^ֿ2eFzaNhM!`KRR&Cd\(DHGz`ReqŚKrv^cMX.&ZxwuZ/|DPbc=isOB$A!ƼDVH-8E2V*k:a:ɾm8Cc<g@G0/q~a;feg.cAk +=2;UBYkwt:lUmivON9ư6fZARn~|ZHpe[ߙA[Uw|w~Yr{$@stx\FY{TnUz|k[h6B#~;S(gP>hl~sݻVnsc>D Cp N]g^‹eGf ?Kh*ZbړfF764-:TPyHyQ_Hշ.V5߯e(k/v-ܧprakضH4H"4ii])gZvyaqֆOJo9h4!i0.r1n?W>rt:XY3kPƋ[?'RL ইr{Le=X4!;KD+"%!W$Z9%(1pxSZhN<*0VmU8hC+abtVL^D"YYڂ"QwLHhMNl"d2ku1gW󼡔eL3Qyϐ!HaeD))g MIgB۔(h‘aeGJn' e#]mo9+B>,Hvqs;ű|9[$ّ)ewĎf>X%}ڲTF6ǖ(.yUVgzh1(̛ȦA5 X"s:V+s穫|ӓiLJ+k >QjGSJJeNXѐte z,פ 2zp<Tі ǒ0)2脌H+g#HV$-G\]P7%-LkK!k y*!QDI$[/NN`=+Wʝj10׳8,^Ԙ,D1#5,NXߌg7sc*Y`z'$` E"(j,/l_{T.v ۺ~1h\|@P{v{>=ͫJgVuX"jE)%5tt+UUDl1‹|:gmnyzV;Q阘VC$cI٧KE%B' tdUIc6ξq/Y׳sҭ/_j fݭC)wjGPRj>\%)~ГW+¤ߧ̬w\iDhL*v7v%yT3C0liBX.0ʘEʔIg5"U@VHj ̕w6X(c6e|t)e3]QYPdf=Zkj:%b"уΡm\ZFa0Ŭ o){!f $V*LL,QUFu3VyvvQ4';ƻ(S]Dž6^]t.Q*-rKյs%'UT2Ija5#28(((7EFCTB!J*̑"i 26'Fy0HFf< bWi߲ Mc,# oQxC"V~Ǿ_6#vgϬ*tA]r#ddڊNE=?*|B dKL$ΔZ'oeg j7CAm3`ɨN 38 &(4(" :9 4mm)Ҁ*J/+)&Q5-EHEj1j9BרIc[]6FD;"∈[GqȓSElQJKTRSJ 0 ۦuBˎdf/tXPɢ,gB 0lL1hgo51"63~D>U!u6%k4∋[sּ*aax@'g>b`H*(ZAcYB#.FP톇{u5/-n ~`mU Y駳Eď gY<?~b` 5Ӕm.i C*^kb=cz+Ӽ=zJwvH}>3YB{, VmhH"LVU;BW`2d *(-ꂹ(/LםBi HFS2{R( y_k Y;M|]LJCct?NOy7xbq& 'Ak ZyJѥ)i[󖸬="QOI.YfTh/sJ_cec`hfpR!gCP>͙4=ېb~-]-7ui}V<TlUZgQx a+Ap:%PLg=gDq|'ԨgFzr!d-D -361(J2(E`EJhf^?qަ|qd>Vv'~a0_jՁF닗ΧCxUNY%6Ș(I%nX T%YzLQ*ɟ&5.i͞ Mol~\+ M d~r<-<`ӗ!q~t(tG'ge:geA0GNjk dN<aX`&5rwW//5_>?}7L]OoXv#6Z!>K㟹kZ@+}z靇HcuϼgtN)l3pqoˋwry~lo. Xw)yzV]Y(ŴsC7\Olqn Bn1) 8.\yENj7%gי r# d0b-*o [2BGd; sݴI VAݷF/ uuS\ǍQN3?_ >xn%^]\̟ Ŗ3{gzy3HwEs=fw糓O=W}=o[QA}="NmÜ?caB˝U\[.vԯFXnv]{n5sM}n7_L(A)F,5oKJ-uE*LʆbX㸔f)}>!zy1C^5x gx&7j-_6x> fׇ#zp9%c,b&WG`,<_\&o1zr y7MJuI^g+1|tϢxgGt˪G5mϞ)*:/v)H~MJRټkz{%{ONqO~+P h mLmOEeVP2%%!$DĬuJej[S*L^ςI厫.b)(IYjgy :(ɉ:YP3s M*OU+)[ #QRN2b ymL'R1(-" eUMGHE4AE/T%"4 P(J{JQѦPuf|,px#?:ˀE"`0./I q9*: FV$${u)2[Y,PjaQKf+RhOCA>U&?"kKF` @,1eHV q?Kn|z'3.S6 ;.SJv?9hrܖqa >vfohi(!Z6! YP6ltF[, CXL͚6fj,їմU¸=YvIӽZ"S2T(xakNz =mQAdZJELKuz9g8kLu_Dv u˨]GjQ7׮Wc?qz;- :6BRHyT{RT}JJz]irGdd܃$uS%xD*MG>jƽ[ SmUlqkZƃ-g&%IcT* Or+PW&=x=pڿzfIfoE#y& S1+ Yv9+u?r {qwogjwojϗ&?G'R߫?M>MǙ Yk<VTQV(2#[(:rd ^beΕ{ 2풞SO\e8x&Xyh6ov9a\vUiEУ}u4~1'˓]U8<>9[]`VG'| odny9&aomnž*}ÄG*¤;a\M_7`ܲƁU-adZxRWV[:bsICjte94+OhǮOוQJtB]7cݾ/ޕWCׇ)*}yo߿.qfs9. zxtẼxuqvZDY77wO)LB{}3~(.I<,ybY_<`zR?b:(- 9XP^>G~Okc̍w?_y7׹/Cwn c1aO)s"OdU9*yeFNRnaġ(h 8nV2Ъg4Fxh֘py|%$O&7~«W<~WB M* M o*xkG.ۻKTkdQ; W.SαG҃'?֛+ڊvY(|fsrNڌ 7Vt P2ʃkzt%Cֆt`RiFWB+-?3ʠV+ftD]r2ʴ=[]XkFWte_(mZ|C2`ftehs]WI;7j=곴*#Xy%?ҕf+zW&p|C/J7xdp.fmZ(E]eyxcG7+~Υ+ҳ&Oҕђ]WFyꦫ*:N9'VuJn褰a2[-;wm%I.;|^|<ܘ[@چi-[a& ʀѕj36Q2ʃ87]GWHҕvw-r]WFy uA ʀ_777 j.TQrtB]ErN]C].73hcɠQvgU"&nٵ3%p+]j}\rND-EWYAc[hهuea*X;'Ldfһ̣]j}8*!]plGWtt56{]WFj UsT:qɾz;Ӥf4Hm[^.pI$%zцL@7VW{`lqz\C2`poEWFR-Y{ߔ 8]Њ6ue6]PWiK }fteL2h*(kk`ճ**z[ӌ 74 -WQꦫ5*OB ʀkFWe׌k2ʐ7]PW904+3 [ϢAU>%ϮǥM OEBQʛ[>ܐ xfږҊ6p2JV|^&},4s@K>q`uk1*9m1sj\ږ4sҌчPƀ7V2(k;!wh'aRyRa璱Jz\L?Or}>M8,~<ܥxΣ]FYۂMϢi97+ fteZѕXr*W+aډ s+2Z> 6ʴ}F])'[zfN\2L2hT}teEWkU*8epvAÍ]E9F¦* ]X|jFWK]N]+TjʢDԐ 8 qyveYj(MWߊȝV.drj&,>y&. Τ]h+yGt]]1U5Ќ WѕF]WFtu,LUJD1T+Pj ӪhU$yX]vR/z}sc%ME/}Sn/}u~_Q[ _e-ksBP:ǮA%ܖfXx{i_½-7wߢ<\*_p v,o^Mov (X8os.'>Yw[O)}qW ⇚ڿ7?^U^q^퟽~yZAEDߘOGGܖ#Ά!<7·)P@Q!_|{ߗ٪߂w^ov~~*Ac  i&!t^J}G{ISB_]/G ڇܷо_\~v~*~v(*a. SsO:dEpst].Ba;'^rpI4Tף? ›tHn]D c[ tDpXF|7"xUn&0S1m7QNnRS' |鮣K9Ӈ}2cm24fK*1 :|*wmFj)!f9'ӮlI! !BrSG$c}n> YŞ!{Cnr:Xr~*!ޡB0fhENp!t 8OCH葂Oq1 0ngfd\/;MZe:53LyNXωGeOC@GDt~DlĨiBA:Gb;xđ$pv "Q.܍ C[t[7K B~bCAFߚec6 FayR9Lg7t}QGxYA$4mN"Yh8 S9 }h@z>I0a8 .S } R)v}d G#k@K cQ<;A!"$4`&6PM}>',dAq7$ O! eK Ks QXO)_i"!]ģ 7]i0d<"2Fp8:Jse(e".BybO,}p0ΧOCu!Qu!RxDpw\#sGnV[/`!k1 :JI"}ZRyED+|RZ0ۃUu[(= AmE":t`2GAq f|t[PuvLt*,CG'GimL1*!w2}{ƒv?ɘYՑ&jQ!V. ۟K zSm t5eP P0 l`BކE@kc;&d:*䢎6`ۂU9wu@Qcuk#$-M44>p  AT8+cn@$@fRPEVJk}eC1%v1PDhr`<) s F:̑gb C1ʉTQٺIQ; ηNKř B5*)8,By!m*%p]^%{,Wv*w3(absYB}7 wNKXy1SNtmZSo֢1wy^ǭf:.UIվ]gӼ\SIe$` ڢP7h!@3 [Nng(v`fQxjQ!58Ey貚`P)@qP'%^Nio2F z0{ӀrHzs^ 7T2jOR S:`8eR"NI%d+x|p4iӜ )QFåΗ.c,`VM}/""!2l.H,Nb@kAH襁56B\T~@UޮHdJX B3[^e"i3ŭge7p5d ՕVBӏR+K)g!NO_oZ-w^VZZE73tW3LTm Z|{9wL_a:_3<{W'sەZG{ݰ|w"n%M㏟ֶQ"iܺyYĭAG1L&e\foG]^@R~)ש/̧/b+,g#.!o~M5d|grdѿܻ<7>h3d5qm}Ǽ 1}rNii9'4(n~*^q{,|yEp߼k W%`c*CK QܓC[A7<= {[|U i%n3oQ;ݥۏA 9攭R dZV@9ԵR\T^k$ 9H@r$ 9H@r$ 9H@r$ 9H@r$ 9H@r$ 9H@rQ@%` k0@D+;|@=u5e pU*Wrx-4wO0Fƭ_ׯ^|;ϖݡ` =R^+gD`O8VFE͔cܐьfd4#hFF32ьfd4#hFF32ьfd4#hFF32ьfd4#hFF32ьfd4#hFF32ьfd4#hFF32}V枙!@фѬ8oF3-,sd4{F+\cH+2\ Wd"pE+2\ Wd"pE+2\ Wd"pE+2\ Wd"pE+2\ Wd"pE+2\ Wd"p9 W}ޠv| WWFGgdr)K+2\ Wd"pE+2\ Wd"pE+2\ Wd"pE+2\ Wd"pE+2\ Wd"pE+2\ Wd"pS5EI^Oˌԭ!= +/1 X^!fj80W] g~aY?9îT{; 9bRJJJ;\(&zpOw+|pU W١U֨}5W2GN#Z|8iIr6?l2{69?8:q\TECz=v@v[F9ĎU9e+),7 N*Q2WW.l`jJgr߻ܢ\\͐|tm>pa)Rܡ_'iћ\}7ؕÎGs\=gCyW9=|o%i( Ng3pQEQ-;JRrjoFG0kxY@ɤ7Ԗ9f-i5So~^z1:!З}VphyX|}8Î^K;)/_]vڭ-ݩ_2A_y"Z m Y+zd Q*FCaϑ2qHCa؊ p \hk!m1%ِ^4 \pGU;\($zpUvH /઄+nnVWwz~xbpuSսµE/ZDb~Q=[8Cs\q\p \hJ\=CJ~p w\U{J Q#\ ӺϗI;>Zig,7b"iɝclMtXf.T*K&E}Bf{7MQ΋$_<6⣪Z&*7y\=y[ IbtYШl<):))> R7YmUS|UW(w%jcjYq㨖?/nkJjE$KZU9te@Vnzk{[ܶt.MȄ`AY]ݘ(LїL ť/Km J)3!&iIMm{քlO O|0{7z?^^o_B|CӔ95[͚f$XeiQn~aw/fǭ>o%|Ym2TVWe[vFXS}|ܤ]nSHM\V{E"zpA\9 \!\:o䬷6+#SCSpe~Lf@pUj0pUU)ZfJ&*JpUTU9I "``઄kPD; yJܱ!+[+kW;n@YNRAwZZ .--ӡGI-Vu+2\v=X@a8ͻbaw͕p$Xܰ?RiffNx9d{Ni/n= qN}$o|п̕@++{Ҫnv\\3Χ-x?].i:/ "PbpsWr)ֹ}S{OiKju9#SurxU0d*6$knSvy`z(5Vϛ&,OQ|?;K۞ی̿O'kIuUvP'bXI'y:ZҍG I* .Bnq|:Eɪt~گb=L/Y^]"t>[t-}qe>gy/.\r;˿ayѻ0?SXmr88\/&B-W O.J׶y]_ BUBrB-zu{ڍ7ڕ>lXY7bXq=bCn_-pb8GskwEb'iG6M&i~gKyWL4Ul~t*n麑9[ @US]e$cDn[?gu^sus[+ÉخSz>U,cg6$t>f~yp^\OF&עV"ڤuA9gA(C:ĺJ+V]f]BZ&ctQ tr#D&]b&2BV \!l!: VS\ZZ{j7,21ߛp GgU} eVBwSΗ'1>kNͮᷟlѺgM2a1B%Q*+kjefms@o2`T' .{ !kr ,,Oy֟ge97 }u 4JyܸC.![uG]8 Q)cWFr}ۜC**hVXޢ+HY>tFؙuTż4Ⱥ*ki+D2 4d\I9rSWɱ5ۿ·doHh|j6&{v: @7d^SBR^+6DIF6md[t4RJ K,aЉ4&;"x w:"š@ NcJAṪ1"uA B[w&΁ q_x`4yu.SZq<ݽ޽~@7w^,颩x~@ G^‡w=\}uNBm1iYk.c%BrogU[7ܺںY=OtH7m{2q?mܲZ[v>㝯y~}_w7ungOqcx=߱w ORssjTv@ I)FhdSRUS%1^9#)&&GA3 A zϵ1WYH(1Bi  `2( \TJNz{qȻJ0ESC??\~8g֫‡c_bj \sccYBz.DabdrOYZ"KR1MQ1㏇=IngSnءN(Ɉ?JRCВ(^&LBBRn1MY_+H$ղ!C>Ą,2Y0aoMgPíUqnxIiIB=K|!KwnO;˝/:7]G>HU/sPsH Z[2 'Z9b-4՗v9^y-]m1{C6]28dvwR:lxQ-}?; >[b=)nI>"c5_XQ3Fi*G0R+ɢ")E hTR&E9D3QnX>t!)v Yl:Lvx;UnnkԮ{eX6fEԁ f6ՂF2q*Oh:tjY'wxoc܇uffm H뷣߮>4ſy_]\ּ$vdx\ W)p&/_^wJRYO!_` !^/ 78/m2(mT(x{GӇvHeG<%Q1 Bڡސ(ْS@A%3Ŗ Jz>fzsk!$'qFMjF2ZTDZ%xNEC3f Nw: Z^]>S};e}>Co:qgAˠ=HbI9` `{7]vvЪ!l˄vevnYuȓM56^ElE ǬJBveCJY#jb*JK*PzLP)LȴZ/,jAN]Bb.E(=!-;;]p(O) TanV׽&%fXWlpb͝Ri=mەX1V诨D0 jJk^_K`V w%YJCD6 R8isܮgєKO4(&׸s\**)f>qttґG0$,#J؄ *@-H傓trZWѸ'W#z6#lϡ|E ss@$GC$|fͲy@FgCRe[irRqi…*IPB&SYU 8M c{6yP鼺Q͜h/q zӓ'Pf%}P8'U*/Qwx܇V\@G.2JVk|FDbSACFV( D}Ag=KeM1BgoFB_ DGhQJ0Hs@no\7,M3g,4()HtHY\/(_5hd2c2^,qNGJF|c("H(YR-0[ :kLًNIiUaS[-$ T@WطLPL;~_aN64Ɠk ؤ V9%C}ѥg=Rξ9?**a"DO%'R!3xA&Wt߫נ@%ւS>!t=hPHhAQGa1zz y>^1U~bt>&ޯ,X/v^ע(ytr+~l# >>M?2 bZ2N,V$]fZ+m^.x0(]b"*f %X&(%c4:j, B$= ̿@܃yKFGE QO :aʚtI ZXcUЛ84a:&Gy_9z9D,dYH11aQHUQ)R7?Gcޥ6{(8T2PDqaO'W^7z_h~UFU i'omV^#cNT u@IL՚hOQ00дޞWvFFe>-狓MϺ Cϻz~C`Z/VeMGq95M=E^W||DfP.vob՞ |ZFh:J)G`Y]|(t0+.χ04wf%9BO 90'=AveA_eŲ}T/dɦ4=3TU_w]X?̓8¨Qx+= ϋBswsqv={[f@́EX --ߝ/=pI p5={3ŅiQ0@V"V7kwq:j晇\[hI_uBկBXnv]Yr,gJܾh@{8PI@h%mbFd@j- 0ɧ"gRzbܷ3hwx^R6ULncD*.5L!aO9CJ55qzUS9ft}s5tc(5xN#NCjIcJPVlh}͋}t]QBu9/Y 5ಆpZ$VLS2w6TIG.J* ÿK>jEF#k|JE۠ vq"JJD|d^)dԧ1b0j!# 8h >tHM<|ZZg=k|yꤸc̼!phLUHAA;Hh֤J`!(h,w6^ZȸBfhg҅iĔ- 9!Cn6=*P Y;IQʑ^IIP∬3&pvLGYoW3[nQB\ϒ%3.ȯ6hj,JF&O6iU<h-M G?_{ j+J/ ұ!A" 6|At}>'w[|Ŧb]Yy4GQ(2))aؽ^A^3qkOAۭ#G|`IE$Ysj/h(&_&X>f(YHڭl7Sj, sl"66##QrNt(;OS:!ENu*As@ ZHH&5p~H)k?v(Еs95:w~Okw`d^B *Au: ?+ x[]ĆK.Sv]E=قoG~LFY=Ӕ.zĦ:u:ʹ/!Df1:1B3Z4Rs9f==eJZUZgGRbpjj#:ouvLvNH,.!趝{,KN%ɘN>IG|$ JP6PZwm%c~yH@q wWKL5%#= ~@V4b;ȆQITjdфX# 5mT:r:r@VW><~nۢEI٪rq1[1d+h4LYmIu'mѪu9v T Tm~U۝m-#w}&n]Vk2oJӾB<(IɼVQI0Mᡨ]N?)hNݜU^tꭾԀZCہCidxRgN#>޼5v[l梇B sx5Mir}s@?_^t+ޡxnk^hJ̷{*'WDǹ{vt-n3 4AB1gdJ  Pkdr)-YI\Agaan; [mڑs=ι~d}Kv3_\zovy5,/ϭeכy|~WEo_d~Wˮ7g/XUfz1Wdny3 &Ȅ! kͪߤ杰N忝sWoi|cpcΩ/7hTHڪ@A߁CQ0XH|d|٢U& 3`T(EBhsaHa .dR[ٽ,ًKk+o:Ri bxT_ qF[ R)}|):gEVZ9 H&BU>)lx0&"e\.;t9gH3q@7͛Gcc!(ϲJ<%c F$+r` .Er4C˔{~ٌFF3JJX1Ӳ`F9U?"k "UL2[5¿&𯮁wL8g-LAV g }PԖqY3VVLv-֧N`}l ^JsEr5:MƺZ[W+֪LE0<$2O ̓A$ uCuB`gX[_!!X,$+u(m7Ku~wVR~{HQG8X`' *ǂ8s#*w8⧝]2Χ9KB֜6҅ ;X:LwQGUL%7n*k/k16`8m2hRr 02b ئK d O{,)V{_*i{Z-%׶kJZ0..6Z^A:p\wo8*ټ-U{Cg/KtjSէ(Z].[4A0FMVZXۭSf2 GM=q\bu8 $*T\|T9Pv`4 ;֞8{vX/lB$v}SC2%`fgi+y/;JJ60)JRe]_rJ`ъoK:[~JBVbTch0[-)Лd]DVb.:Wc6aؓe#c[Ǯ6vGܪcm EʀDR}=茵 BPĠ<N&`DhjX6$Cc%dT?sLɐȥLZ֘{yDI7į:_u +B>$ p{7o_qld',b%?bɊLmpQSjvUWzٌKՋ^t^6gQ2 RhqV) $O%*Jf1Cűf1}ӇPaZ|Y\y7YDs;MяOem_ϏPeOTŒYKp5䃓EM|g&pPyy`I1L  "svHH ҋmԚUdÜkOOl۩㖔3B"3pޣS^l-( XQj,9Pc[mFmb/A4gCP^mP_#~Qx UwTiL_lyOSGXp#\1‚Eԓ:)#^,(a!zu&H/IlT=yi'Q+$c !EIˣ16SqY; l ((4OSzH'*2?('^g $h!RJ@$J!A濅Eʖt7QvzZl]G P}D}(jm uFx< v -҇rUR :crYwCd&P]**xFrXۥiC6B&5j|͚;l>>9N_f?h~q8~%`s~VU>?:(JGǧt6X拣2U>QHg/}}lN8#OXj0l^Uإ|;/yHKwg(Ne#/9%1^2T>2H _Z4,LJ9C%ua3)k]pizv~oCp+מhsm5?QPMElY=o'vӼ6k|ΙkP'х]I] iKgCf +wְ2i-S[3OmȘ lS%R,Ue5Z,Q)5X>%ʢm|;8keI E @Bx*%>P 5١)Y׮y9&|/z:S/yskk_9&E7.gv)S}~yDuS\Ơ ) (x) ԚUV l9XW$0V&<:E[!+d mhbH4z9`eDlZ\D (Uv^k0+k"P8rXgL, f(g\}}bna!d&Ydƅ-Sm02bVE$̓O/E~bڝJM\5'Qұ%BQ>N6> j5Zd&;i->#Ⱥ&y4UGQ($ee@R4!ۆaҟ{9)&cTG{xbϣ=mG;|`Y"XDdΩ D)1i 02*13GWDGiƒO?V,ac[ |.9:&"d(zAeȁ5HrAI<ډs3W:G(#:A/ucMdsCQ e8.v HM((/d`[k0ц$Ω;sD"]H^ Ir7 a$NE*ףl]`Tp)FDR j˦285 9IX*Cz&O,1nrNh(ӡd%f܏SUQX,.. DPi%Kv9HVlȚcEhnkә@;ȀZ-\P00f%e/rnnٱ dH'dHeAԼ 4Fgd-y'dAIe Xu0_P[MG|űN ں%* mMqXpY *9qQ\%WF%cXrV&QcAjQby [Nxds`ؼ:Veѷ mUg|"w^^ ~P}V٫f3k- u^y!o^բn _/lakqp9k-yyF^LeӋ.2xw9귒:ni9xHt2#tZ)ɓ3ZR: yGa;7 aN%-*3#)18h@& QggT9!,:_[O|%ȘNM ɨ@c&&L`Vdؓی]WzyLULxjryn^ݠ͵ ՂVk~zϧ^ԘVM7@d^?yt9^UaV$΢ԝ:uT:w@_SKr0Y d^+v -l*H'$"\.6A1%RT舂aafMj45ÏjnY=&>ϏW&𛄤$ְ2]]Xns2$ 6XEٚS# p+?R-C]t2ٮOc /[6dzN}_X9vP1`yG2-z KuӾ?T}zSՍTuWbG8ݢ4edݜuD;e?~uS;wK={^|a5r/v1{݈]<;+>b[ql -P1ɳ髓d xS]k?>[Zm={A:"R j]ybq5o1łT |Oۍ%∌*?{Ƒ-#! 9d8, S"L4=w>%I#Z鮞Uu=L(̱DDD7&z5gBRƉNw]Z5`k U cJ# ?g)Oc.*5X@Uc9ϦoNZa}̸F* Ffp>;$_ AcFG/8?:=s9W [~vKaR~tBIQI)%3ZF14&zG탌S?:ze#͟zLtw(vݼ!.ZTj[q~Ts]߮BȎr2ކ"N/WaPL]:1CwROK <{׭m雵}q02hqK~=,~ASq|-+6m-+a !- }E]L/;Ϣ:[F*%1uk[m0 P[OTAKy3vp$Uxc4a *2gR*C#õƑ:0*d)̉w/.#L2%{Dܟv=6oZ9_Uf>:> Nay3I-S;4)L`1- &PfTetE刊`) NPL]99x-x<Nb[p]m"^ zE,_HL=uNkࣩȥXj&j>JsWX3UKEUyJ}Uwq U8%CĔZa^  ^`+#%{\?xa#},ޔ(a"՚,0|`SD#) C wRx Fd@(:o;cƌL"^ˈiDҩ iI:#gGS4S%b~hbζa%q;K 3EloMG G=8j sSO ÌB_į`Lj|>] `aG`΢JƘ.qIHc~^s/*ϣ@*2-4b[L TUB }.>!!Ȧ}_Wf@JM Z8o<#7&oבKSط@)s ~w}|/)BNxʰ6dێw6zY}0ߎ6c̿]ar?5V3)ͼΗm5 ׉Muk]+ر%^j}v&r#iU]g+G&Y6wz`zb2;0_n<ƢDƲ(_^R έ~3-HJ$ʡV1漏J$+ռ֩EoE0[ہ) &^Q GW8#%G>zSQ y Y^zt.W$7ZɄr2g#C.!&bi%~ gG?U/5gw4 sV q;!gߛ#'T"SO7pN\ F1̉6_{]^5Ż]2כI-lNv$ZGb|HmÐa(,o> |JS;{buoarJQglm 9Z%~HpR7U(HbbX^ 7ׅ덮`; Ż?.o?_&xp $&{@!?uÞCxa04g]zwW.2}ǭٞ5Qz^zW5g%MtѲ85A#%~ l~M]R9U"DtQH;>/^"p߭6Q?ZEґd>b+I5a/Jٷ(7X \(DpЬ#=a&ʚ{s/.04`]Oy c'!Pc^"Y"dw^+N'w:U܌fzg{j?"AcovI'5*lpۡO 6Ac9VNƣ~$4 .-# dRKJ16#zwo^/'t L wE\N>F#kj\J˶(w\ox LqˀY,CLae$XLjAjxɫ*?`ȻYd΋ZxHFd5Zآ\)-^qg%Sml;0a ahZ|)Ia#L%3qr" e?oEHBG%rhrZJ]KT25WQ+ W\~4vD!|󣙘ELEG(SlADh C<<)RYp.>Ar8>Ωs8̩s:8>?>Ωs8>Ωs8>Ωs8>Ωs8>ΩNN}S9O3ZT:"S bv4.\r4.Z4Qz^˩B*3L9S{Ԟ3L9S{Ԟ5=4@5/V퍣vIqԷnS WFYemEY]]w@ t-iҢD,em i]w\R6Ͻ AI:[:C=xS#pԃC'9!VM5>nv@%/USx4N Njrl|HAFYTJƘ.A,O"Eԁ}!zyνX?6]|0cU8/hB` O'EP'.a]bjw@sH>M̤q ! XWkHz 5wB0ތM_G-eQ9<(tSO꼾W1k|\N(w3Yv\5gp}]DrªGڀmQ9蕰8jFz%hXd^WߡY_u[Mu+K[dj.vz[_HZ5֭ʑIjS۠.&mmZкj]67wsPmtfK6e -nݽM^ϋ;FdR!r}D.{n <%baL7Ϸlڔml{BK\#jے&[^R<_W6?ə$e觘j FX۔y\YKk*Q PytږmO@4/(DF+dM}a3oj79+Bd$@ U˒2=ř-IQvKjFӜ_CjsVKWL߭q*c3jŹD'N>fl%@1"ZrOm%H**_]tڨ̡џ::o83s^.xP#jl>.)~S=CȦoA-ఔ6бՉ˙S$3a܇[fsfdvlgH`ЛZ;2DI,)A ֒,Z2WQjgٴ с XUd='%Gc2juj$)"`rv6L CQ\JOgc,3AbvbGAs`kXk zxRݛtm#N-@1B[ROR @GSF_c/!/TkժWgI_Tkk̾USe.R]M%"YJ1[ p Yկ/~~ p?dR1E5 FTO9үäX5*$?2]xǬl󉜿ѣ09p]o> saru_"h`۴ݟC͐)/'̆H뀄%$O8>y;T_3D.l65x2k< Ir; QlC[LVo(SQGh A\*c[ʴ&Bt%Ԕ1x5MyP䳰KP~z^~sZ}Ma<`5jΧbz<1UsT[ d1'1޻Cd RRC 6FƊ<&yH%Mզr>eb&T1G*R˵$.N˜5}aCb˹%'gz(O)bLu^6E9E[mAJԜ8fG$l`cWZDbc3ԋ+{ٶ ArZLGx6dK;.m};h2\Av:.o8뾅oMۻ4/O '#4ANbMXM/ijλRѕ(lyQ} VgpWz >T L``Ĝ@I,jӭY 0;>vYf31ir,>q[j6Y*ڥ0%5XN*क\Qj:ky3os,kQ1&fSbXHr%q޲8lJ{D(-WZ{ {F;ÞApU{L2x6,+fb* j)w-^Nm 5[/el\u j-U%j`Q#rm(^ BT0q?qW̧RY}aY, Eqp%:Oxٳ6bΛXLP*~Ր E@1RL_RUBRRq5Q8Z)HFa< bwWo9J30 OQ3= ea]woS>;t}X1j*5aP,&Y(JJN'W^fg&X]Է3J&ZfjGs?bsbi\P;,0;y/5Uw,dF) -2cLtH5i>(S*-ТzDЩSYjPRnxktma܏GNY:0ΆǎҏC1Fĸ ₈wj|H}fF)xIRkd\(%of@V&-{RVZ8kg J(A Tz#d#0qGވ>ꢸn:aZr(.H ..xg-^ )ks >UӼ ԦY6 .>.i\0Oa?m_ %E?"qã/ޟh"~nev2"kQ i$qT ᔂB*-,]1|D˵|i~ڤ*w OApglQ+єWcɈ}x٘F^CWC@.o( ,ܔ'@ud&TC97{ y`:].~sk8;-{݀sq.ߞ+96 3:#w?:4LQMzp٩^No)*) “ Bu[HpE5;Wl)%dLNx ܊`2ȫCyK1S.1jNXG 0qeNf=Ayz2;V&_04ņ;џGmuh}z-m,:g׆Hq=땈OꝬ6H!`[W=bp~2ߔ h%LgGڟ~uqyÎ>:R[_*7DIk:!MA#z71[3Rdျ||0zӝyldQQ|Szs^y['z7mV}>wzr;LJ5Mx9pnw"ۄu~>.;o;-S:ho4\Hm*ܹzf/>؟Uy_`vɭ[M+K˿]pOl(|J;C>.Kc334hl+E@\dŷȽSYBFmsL$aZ{e3 Lkr65xG\j۝䅳Ǫg4ԓ^OZrP'%sw,j:w`6|lttYR |4IĉS)v8{œbH@Y'zf0ړOEUх"Ɯ82 ZSQs-RoC) [rhu0`#RŒ)SXOQ1bAy`G4]./\JTT"WŒ2 PN5V6 jo:05^WYur&A45+8R Ș 5kLDȥGVA#B gR׮x"cE$rc)EMlKT޹q$ _!Yս%ƛ=hud@]1c75CΔZ@')OSmZ)vu]>I,/b>)yt~ -Sܕ6]ŭƺg-E~Г"#I_}{};_rr)?~|JWOcAN_wŁ|-Enpل ,^ { $g}LL X8h$ ne%z^'"zȮP\W@aO_J*!lERt^Y*TSUxV}o6Aq; %~{ юrOA":g!bFnA2hl䔽!wLu]m >MRys.|IO>N)s)"h)_N>2QGۜ+dMrߊ(`.pVKJ^T\YLת<øcB8ɫ1̔V?o˫}g}4zn/?a\mt8g՛g?:U~~7nϪ٭OdHɆ%g壣{m>wM6tyou5m(䔱1̮5CD!;XHԚAuSLacHwh!;oP7.yzdא2ZI"wf`Ϊ8G-2B;`^C/v袽|>$B%Dn"=[#QPې}&8> }85 QN҇(nZKVJĩDv"].ѕ[C(]WJiäSԕW|Ov3]/عMWWݙE9%yv}o5,w|0|- րeM|gy׳/۪/;O2]6/l5nZ>FßwpeӲ,sfٹ>f-6S(h|)Wٿg~qh'%\ 7d46e$@?uT* T7( Bk+=nPʵ|N'nsJjt])muqJsNQW 5UWW+Pcu%a-tu:rHdjҕWԉxD'] -y)%IW'+!XW3Licɠh`zD `ER`[O2kѕҺAfƞiM`P=]DWJMR]Ug-]W9v28GGZwd/G']m[@HW@lT+=q+ÓNPWJ|ڏ〻}?6}۷?6wRo9"j^T.cK7}Kݩ0UI@/ǂ5Y,v8 -qlt?7# F0ij~RtMLݠ%4t-j֯tM5CnU-SM44ԚdQ.gf߇> PE#A52Ԓ(3 SIf4 n-X.lŸT2Ό;wx 2SSǓ;_7P-Z(]Ja j XOjthjѕҒ)]WJ0u#t$Xq}5633=;I]9gJFWMEWJt])%Ѥԕwv};uJqC5ɠҾu%फUpJ]FWy"* *%L)*W+FjtkѕRRJ?%_܆UΌto6 c]#dG{@GQ)LWnܤm?$}0w<k`8eeDɟ^y- "M E_Z4\|מ4}MmE`PC-R{e']YXYFWKP}RFM:s]J=U+ DWBKx])%NϼNQWPW] 0Q=krIWJ{5yQ)s.C-RZW|2>L:A]踀}8co7a񃦔rmhפѕ1 4dPic񃦄͔ o"] pzZt%wv]٬ɜ#ѓqXbY]ÍGHZ:NHJ_-d6IWT=pػl5FWJ{GRtuBt%h]).ZtJוR;uB)I=WoW7M+{?1 )X_l,H1̗ 1=5dD-5AAq%nZ:2L#)XϼFP6߉,95JCFW{ɻ#i_z0RJ7ʜ>`ɬgxյp R=f8K-h1$)ڼ&ݷoξ~(Eܷ;i}jPw~{BWGKc !󊞿tyy-~ʗB7[)mloFE[b 3XͿ+ݤ'7`1/T^XW_{-/V;{+WJܳEif'i55j^}|)OUWXkc՛5˂ΑBD{uXf!lj[Oȧ˼S~2Aď鸞뎼igWL~5~+oۛ':3` ~6!&kKna#9kr:h.$p-$KV^yww[/P#+탗#ŝnmWo5R yĥpHap]> 7ycbQv!t;\?tG)ޚ.fr !֚!#smgqr`4xӹnRw:h '{ڮ#a_!vKc$q4 >SߵRDrBIZ r}^Z(rԒ]+bt`ҐS)rv>{VNm1Zېw$z- sel$ ,^( 15)MJsO]A[BX.fсf1n]scse!mN8y Ҙ)?Gg}DҠmh ݉6xZiMI0r0FMBm/-W CC;)H #G'PDw@Y7&Җ~no.SXd|H-`VJ::fiY,sbrغ^[w.n'@@Ԧ~ 3JԶl fpN$7D ;r3B a9)>HIzK)kmȲ0ۤz?l ɗ<0gD&Ϲ$E҉([6mG7Anݧn{nw1įZ)|!t mDi?s.RHU}i! !kmMфjr5jNxNŢKSmh5ST:8S׾"FnJ}Q$\' H4Y[V RBv$JkCvYviF*Hh"X # o b9PT8x4olJ?o:@CVz*&{H\!iZTUFB1L: !'`Ge %tF\8gP4I"i Qk2!`(`GyOC |Y%H Vjx$ۑ o8tXTuJrA#+PZDbYьmC5YK1Z1ة`U0~~pϗ ӸK =`Ye>z ]{ #BK|u)}ў=Y /ѡB-10 zv5t3E=`Jɳf !nK)5Y$iNJ@K^Bk=֐h ўU3ڄ`ctGJ߳yp (tg@ @$Be#kvqAص? F!tJ35)ٕ P?AjDPqkU{P2,,HcF8&gو.BPmʉ6BkM{JzBz֖Es64IxTF+ j@eV o-JPti[ #I mU($`F}+-u,a* z1dkR@i{'{be5o`y ״b2I n=CnNN44z63 {@(8w%[Y:64k*޵%8O)'֮ Ęgr9KhްBg' % xKd]ж+9]PnDV}qnQD{}A,WR(TP=`ePˌԠ*1#K[Ƞ rwUyÖ Vnw`E^QH"NdviP'7 B`|ևDyE aLo(ƈMnƢb$ӽw:Hs[qOҕȪT~Qc@sjo6i]0r0Κ b %_Tޙ4/Q4T Z،P-EC9yye}z*DEPxZ*;k*Y[Tڠ@idzZS0h¦f@ 2)z/$6O34%0U2'ʳ֞ +JOQ!,iJ M\ѓFn57E!iކ`Yc6ՔWn6KC$`PrAvƒf5$mW%T$iBϝp$@Br6Bz!(L ΃rt `j?zDo}qv*n/jA;PChkЀFq~*xjAH#ʢvi^QOA-ohjgo_o^޴ oVm!ש~뷽u)]0+ৼvlo˘'̷m;q(4d7ٌ&d9oaZzo>t|z:9.'.Vkj^]^\H -/#j :_)7/Xљֳ^崿>_N~"^̐X#nm=yfS^f]3M7we'40hݘ`#Gd\=;;upWz+z7pņ+6\ Wlbpņ+6\ Wlbpņ+6\ Wlbpņ+6\ Wlbpņ+6\ Wlbpņ+6\ Wlb>ʂi˜ WSp*pzUFZ6\=AÕ^ +6\ Wlbpņ+6\ Wlbpņ+6\ Wlbpņ+6\ Wlbpņ+6\ Wlbpņ+6\ WlbpuV՟ :e9襻޸ڱa[Z]m B^Z&iC.ƏFpha@}0BÞ?]bLt(Fq,t鑉t ] 8?&=oxh pc+Bk=]J ]E#+ldz;3uj,tEh?&NW(c1];v|.󫫇[]= n*i>]= ~dt%@W}^fO\tEp ]Z;]J ])8z:~bU׼m z33׳K~_8m=V*v3S t=8Z__pc_ƌ9=@w|r7;_LV/M ӱAH?ruLnVIt>p~ܞ=EN[Yrjj0Z!zkJEZÆop}/ -04i`C9i2Qٷǘ(Ո*e76j5:+B%W4ON8OverGw<1{9h*fn;L .jqi=_r~5go v"^*\z-ȩ~PsnLɵ B5v,4Mh]|4M(C`~4mT -7'GCW׌FUZ';]Jɟ"]YIfx'++kXкPtY]pa4tEp ]ڠ;]e"]y'uӳ+G]0uV>IS%ӡ@/Ǥت ~#*:]8(FS Z'pJtCW^=^j > ܋U= n5ۭ~R=2R+t]/ 1$GCW׫ЕV|0]=ERJa8^73ެ򲭗Ӳju^®pK%4Mr9䐂Cs*FFweСGk%lxv?w;gWž*i]nOhj7Ӻ_Y6sa=y#Th[Յbo|8h`-?{=^0touC/6k)7UXs6cr-8M+'o+O_ [j4!kZRQӤԽs:Ek&p(._ځ/_N8T맚K IǨTBOunj yaw5dOC?ߦu En|wppݥ{FBL~BZyY\eM)ehܯH:W|~Z^6AQ>_^Y/(9]\-Vz˄:v4%A%cLr*vVUSy>@_.^P}8U[Ū25Vw!bj|TThKŭ8:æ'?mT/~tUZoUm=H9n_oJ#$ &I-- 5珔|8r|e3M_luqm"i:Yܬ߸JmF u|G>~=-s|-avX6CY C}w&}ÁFch#mxIMy.Ζ-;OqZh;Nm-}7?mwt!^t+$>Gd>{gB[-0tׄnNCn))v?a|{s_][o#7+zmyN9,acȒ=3OuZ-OzόYMV*霔:]iΆ?UزEíwozj|vDk-l:i{ 9ZÇxaϦ7wmo(\?h[nZv̵fƒ:ߐ>Ě_4j-)6,@e8/w>ZJdoF`N@LycyΨg4_ˡVfwo=Vat\5z{ŻL".ټs @IEXmma.64tV[V: ײ>^DCׂY9t-A0ƍfɷފЮBYAkQ=d"e&GYpxȘ&%$viCP89*@Gnuf8j1WZ<#gHEJzM0ik:EDrRHH`\qYmCe tA!knPqgFś 1<eJF-G9St X^<>γ{=^p1rز,=[t]7 9 E%gD#!k'ӌBLLrz ní_u!.^`Y>^IL|\,y4(Ԕ[AKhEB!~n8^YSoNrΆ8;:\~P.r0nWvp%?A㿪bVN3":pqh>uE*̾2ihd G(#pvC.d ̖ev/gYyWK-GJzB*'GAT% _4Aq2Py6\~c1~JT-ev[_޽ίo  \81E*իO\g+ :lSzP7.ErQZ?z:0OPg"G4]^sjnל#tmV~ɷ4W;GbHm@0a֙ mir$j:,&\9\s-~@xɶQղQ|p2>?Ng52iƐ>}{Jt*'Uw6ԏ~tsTG??>\POy)$$S7?rkCxEZ.ߺ#ƕ+{_| !q+cv[n%@~$|?*?~z}3IO>`Zh2za$zyJ.*~ܦJrZ*<@K=ҎRhA^hML9N0H$ hPQLb$X6mdEc %dYo#aJz{@mu^3Ee1_Zx)eTM88shǃF؟trSٷпg'6 h/]ژ#<@ygG|7AT2(탫OvRhrٜEG =>p1/rRipTgvfCs7s0ZFnouJf'q|*uFPXGgq+=6Bږ^M>N&-V}se3lmAdS2nڔ>j% f9 bF)l: #ڬe}9Hry`[IL"cYRUzQVl׃-go to=z%^[RdZ>_d6|%UnQ +,.%zg3ӜbfJ6>r$8^[C)xf=띳IA+ipp,HəĨvu 02Q0,h5T"=&b:茜#ff6+K;f6kp]3D҇ί6kY$_2I]QEt2$ɩ IBsKbkR"H%@KB  1kdD!FƭFxI =-U @x, Uv,sXw⬯@sp|fO,/onNdnW+Y@ut3$Z?]^v|;4Q)[8yJIMZvԎ'gG'x~_to7PayR@8E\* T\AMRjo7?\mJg!k4(ov-蛏"_9̠z{y1to'ߗcy@QhL2ok3iLϢ2/7l+^1bH,r<pRCQNT fYH.rMpT_u!@C*&cCMP@d ;G$˩]'Êa;% OHw%ޕq+ݽU[NQ. &w-3+r(|ťV<*JxB4: <2)HsC|S#3ͯBzZUOy6$- ͓IxC['"Y:Y4JL͓[ѣsOq _Ysy1'{ {:D HAq,c&'SxX{LO2E}%ТEf]  PԵe7r \.>ȎRJ< p~ SBγ괔4 Cy}kq!T ুjj1i:[ OGl?_I ͦ׃rtY8in|xr9oͰg0~#0Y?ξ)VLYUl0iw۔>̧77uowHn*.k"I !0B~h钘&Qg3Z'V37 !HJTɄs:PlMu9=b Erwr6?UQP㶎E,}>5~b~w.kT%;2T1Y;gt8r~nYnji?yf@7kjN6do[Y(4z+ӜU=̓An`B0L7 \}?TlZl#"|Q~b=?؞]'\ކrbH]u`]Oty?-.DŽ0~4ߑN9^an;6l|o`2AMg&OFMZ[&iu9tmq7f]V]dzlha7V]i N[Ɓ:n\ۂ *"%c{/?gঢ়d,uf8jIzA ꨝ698"ߔqDP B0OͼJqNI!} rCHq+˷pbW=jһRS#}1l*PQ߃G(~)>:i.%Ex0;9ɨ2p$XaJ%3=/*g\3q+ Q\EHPjM!`RJځ + ۼ$0PdH5G闊@`@G a{HBsQi)]dֱuFΑrvzeJa|_D [BhHN "#. %/@JDRoa:SC/]>BO2gbM"< &IHYrc*ˌ6= jz;iF$$!$Rf!Q;8xz' :W(D'H'Am!ԌCaS,H m%C%nⵀ% _.KS<"H,9 <E+|c[Đ!!M!xDŽM`t4$g`O΁5kpţ;XOkn#ױu_vBIL>þy⧓$_d9vCnYq*,MOevf!Ύ, ~c?HVl*Ar|Yr%70qɌlny~)bebV*`5B̶LN MI s`-})+U%*7'`U3HNHg2ɸ LѲmrYC/EA.` CVɄ3X!Xp֩ӥ6uv&Yn,cB(7dYP BfK y-?fx\Υ,{V$+LIC +UzxVۧ #F(A2"0V 4}9Dž7K(4Ds*3EŦ~- ' nS l\&)Ch1N9Xw녦OT 6c[Kl\G@BZư59F|!C"J*Ԭ8]N3+Ń;#RV@g\jCoZګn+Po걿{AK8@@Ao+˳QyjX'e8.UR FIfիdK=Gӧ1Wf.=g -r{Jn;JҰ:z%Ԍ*q QD]ªs5`jXlU}O-}V6DnjO'e<ѕoPjwE^q_͟N>nYVu=+>]_S~I/S/o-0Ý)r"oLf\}_ށ^[?D_~5IJdUXtmC;X%?y"Y?lm)fϒ6M|IԖ4ὑTk5G{ls5]IJHb,&!ZD9?k:n6H%`*9$@T6S*4!*FԷuFȸ⽔=pI[S;~%:{mI)$iXM*yNOʶyXN\Ӹߛ h+}c+GF{[d %Vu48HH>چ 'U:iɣTjwvqzuӖ5uw ɫU mV+z1-U|}?Y;)4/b-=eU W,z^i䀔~lB]𭔖o *94IZBk}-Fؔf<"n42,cW.tcp tԕWa/W'/!}>yy3v}T?JraDZ-HGiēP*9CA}StjEu:1\5EiώԷ3*,I%q-f#3hpftvZnSaѬc*^:Sv,Q*TA<d#1xRH|B-9[k6ƭ|ةRJCVe^tƪЪSkmNrTJXco6 |r7g:1NFǎe2ˆ #kNHo͸C d`\&%֥hMTt 8ǭ#)q^'PœmgRI9Puw!g-y:sOpG0#~~Wu妫bYɮ(#bXxq{+XJjE<Y5 w)eHVrMU ,^ fSCٍ@a^^#Uq3nNX%A>" n~n\o ;.RBr|KEsXd95c0T fWvs&9O܂7'ha". YSiZU_H5Bٛe?i(ɖ^.MGGϿv :[GJ"E)3F*8vb:rJe}WjxUSot).7bdBQ \OQ]({nlfϴHN!@cD&w$$Fƹ@CVBKja-ԣyYNsRNqRAAuIfTH(FF΍T)Q F5[=2Y1Y&^;`Yͤf*?FB$c j*~Ln޷JPNAnT(_#3A][wt6LE/Rsg܁s֋{{gPᯋFƄ ՈCBz}T˒iRx;Ewg}f=b9؜E`ri-Gag mO%3+Ulai੡զ@7-RGhG^ǎiFQNnDŽ\ ڵ_?fK:]ݛoy e|K+)ƧA*TMcUC Ƒy/'&.%چ_7`,w{~>٬-_n@P`!x}G<]7FoVxB>69\ԋQ!xσRjCˋ+aڹ bL[sP0f(P {agXA\rKP.qy bݫDW]!ATdU\dKT(*ߤl58lc #E׼<%cMt|F,~(۶`K+/'b>ySި0kz[,z@zUo2^?Ά[kgz}8?u𤞺sDy8ild B#mܓp{'29!2-2 ! Q:0UnwF &3VǕ|!'*'2*c*59S{Iw`8c;XC{z~{{w7g.;grmόo6BW$ )Qy[gp=U-fk`YeFu]ߴÞdZ^$x{*Yʘ}.*?Yd[@gC\2}"^]C T||Fp\GU!jX])pvq\d&%a5ѫrV5(R,L#UGN7BCȆR b+ BT} 9n#LWURjc i|cy|U !")n[W=pu~al%z.An~|^.nl9.@( 3*:;M\v^tTQʭKStC`;sϯr'JbS~7W:;[Sx|zvy>kҜaQ=Puf h@O tokgϣt%^˽6_ԊI?xl)4l*"ynjRgY,kX^,Xݘ)&G^1  K %-"2SΗg'usګ%맆 Q2Y&+6,& an̗[ FIt٦L6au+v49kjw﷒VR+tKY迣??.ήև5oo({Nȳۻw>ok}'xr.\<="Z<0!Έ:0S++E}>]uvFgNq)O 08yGk':J^!]9keFt|A̅:ZSxWHW^I])`z`Xt2υ:Z?yuQʒ*XٻJW q>tl誣~TGBW70#Ra6tz ]u2yRCWȡ7'2< 0ѡCOt$lDWOC&'-t])`յƢd.t;Kt(Ѹ^!]n&Nnr7JT:>3p4q?fn򵽧`wmgoq| _޴.E >YR%9w=,z>HCr!3RVE!p .0q$*֑WH+JR^!-`zStV.'\w&zPZJ'MJ:bϠ/vDVvx~!W/ԚJg[5Z}e//fΔ:+Μܭ%hki F{4jҗbGև j+TӠcǴTVhU!WcQWZ{P٢ʯR])cҜ")uUȵGCWWJ![u+T '+F]j:tuUЪP]̀*ٻ*v,Pk+R1ުP]Y.RV,kw%Џƅ-aMRV$}Xj>۱ӯuȺb/.S x*eTBP9-EtfASqBݽz`z ǟa~9I⺪wzps9+R"gv6oSzT,3=O<=Qz.M;N焖#FKKӳb3Zg]MBrL!io^u꟯Ҡ4լWOЧ.^l@k Rb-W+筭Zb& B#hkt?F7F7F7FwpH˼hsi+7,ɆtSRn.:dȃ! BQ,FfLb:9I(|ݍFT[>F]6Xۓ=W՞]T%`c5uнIoU~[c[{SnxdG#%>:Z]H < UAT,ӿAG9H }C"7HY&JAf1$_Va"iL1Jl"4db)'Y1T4*Nk#N+iꎆ>$_ {㪘}ɇ?_p'N'\]MZpwfmɽZRG+8}:H;_zl5kzӣ_#-n be#VS'Ly/]%U{,`sB1 A1e=~ASk&SkܐCbitXfy<(BQj]B="r1p EzNXjI 8c Fզ 4qZ(ƊLJ JE5qJfظb4_mozfY,|7kt̡u\dA٧\dJ爌A`eI CF2B/2G:_uq9gswYɽArԬ XGWݾ5L6;]VꄡH9უwsXЂ$yiJ a<qo /:ś/:ߜu~I?} ^:?]|kѽyg tsKO3)oo7ALk泧?4˷]<~0ʞϸ  K&Nh)INxWws2:f?f聸*GNL$AZ>MVnh޺||^8iNn1? |4 [{ }D7e{=m9tr5\In'wtĝޢ.?iۡkd>pt~ ׸̻Ν1L_/[kkEɏo',X Ys?u׽έts9?_%޿xC췘_cz!_=]FFNC Vv,oIh~,X߻OUwapRUͽuc%<ج{ |HXm0_3ȝ6N ;AÆF5?uǝ^;oo~<\7ߒ#u$l~}9<7?vxf߼kbVgCZ495>Y2ڙƅɭHw?}Ouq@߼\ {8;(-Vv'~mfpn74[KpΨӑ;K R,«oT=,{V;<:|~gBkmt:wM\9}nkc |)Cؾip  E;R軓;κ7l -8[0.ԖKfNJT/B"I)9DX ˬR ӕ)W9s8Q+iU͏EY-ڊ6] G7:2еll[BHt#+/IJiWPv|J^NV>Py Z)-jS)1:tDI 1"X/w3\> c6iP"Ȥ5HAP0KnC"]"q\8-<V"Ȭ@#gǃQndO]v{jy˪l8_O.l޾3*sFcC5ptЌe-dIZ"@+F̐ib*:AJpt\erƼHjC;Rz9zFa3zn*jIr 9%x%x4Lų6cɆ 5>EXn+Z'DGeH 40+g4#Ksl=Cՙ9>ϐr6rb|ˎ}NBȄ&ЩcT1fody#j2ːqpr&s@􄉵 B7;#k!:!YF$ɒkNxmɀ4ht[KI JH 1=ۏͻe2σaNɕ7^GQvt=wJ?-ƩNG+8sזO#DC "+py22&%3)B,*VcmN`& 3)@ܺ[뼌TOXg Z{ P%;.d.8aD"mpwުW4<ۉZzmk·?u[}g5v'3MCN sW)kv4t\>Nv+* o+|t5rFIeL<犃XTY)ZJӦ4_InCjť47I;N2I}N@g&3 (MпrZ$l3Dm )1i_-J.}vF 8 kS?Ji#ERH&$/݌kK,e1zFI <]P1J4J Z{VX A;+KeAKn\v DI m[]r+B5r$|3b1nI@$7ʠK2T%se V&l)%lu=^!v##P qLEVXJlX#gG9;,gy60*=>|ǔKA,]$p!ZZxԴB4MT";ifrÁ_|Rw/Eg,Ie[Ge-CUܒw(C z2K"=oG#㠒6_Hr`b%AM @#`dfQ2ZpxZ"Ҏ_=k9;"D[IK%i|(\.C-$˒2# 3[#cݖڀ4zY/\s26V(jt Td5kf+-}a=4գz\+)hGva"hj&ʝ $D:$̾q=}#MJyPGhG;䒒m6 #joăZP̺o2νb&bf%H' V-V +!?m!-,ޢB꘼b*cDAe`| 3*HY"zR20XXCt&ynk]+v;Hj"jXbT+e[9\ջt(ku}mg[h5e:*ksedRwQp zrS1 :rdEE)es+L>$Ji5tX-bw(r5[֕Wt {H!łBԴ8]}sQ,H N)aeK5d"fֱ زN=m]{֡wۺ*wbмxr=rv!:?-ۢ2_ lBm;X* ɦ0T-VUF|9wݓ*2yLC@%YhsFIAL":]+X 'a o*ϥet=2m힓ns4gr?h;lLJ[ǧRfm7+yx*:ŅJtb M'4Vm fM=gD; {BˉŞ@QAw/ jtL!\T S(Byg]OmP]&6G%u(|rN4\l%[}u\04%+ȝ8pH>nچҲ8ܒ=ݒ['SL28*?"&>Zb*Z'6 ;VRK3W暒j+ގD*Bdd=א4T+F_؁3bR8pU0b댅vƒbo3Y>ȴܫxw #v};*XĪ ۊ; <P3 [ }NCF4?fEʤniC:b)!SCj.UeO8#v:?&TP jg{/O)6@ڡʆ< 3.%+t}3_T)%b! 3+:"ZŨr[M qjqމ& lMp4K^c_Dψ8##>\@cuC \K65]:cA4ȢNžy$R΋"bE Cu {ԵL +)Xҭxm3"vaD|ȫNm=)uv%b茋4⌋&Wl J29xc@LN\8ROYU]C.ʯՌũa1<  cbI n~FޏjKݽo1>?Ge=Y8'PK|R"Ly| b_s$xؠC';]F^N^Z<{:3*b&f)>H]:v emR P`P Z4.Lb岢 no߽S Xr61HA@ d"T+E*Ьݴ~93'ަ;tqhd!>NJ!d'ěo] pu_'VSSU.d $6beM+S;S3yqMxkȹɰ̱&ɻin&0cƍSGvJ˷# f_۽Jv,EuQoiԀϦG]囲\DpTc<_~h,g7`[l0ΒY,K=-veEs;:wEV1"o]5ߐ N|Arƣ5= `ȁytA(匛$eH.ʻJiУҲ9Ow].oq=nD4\uǑZXǾ'esyyw6֏]^`ȶ=s-z$ȥƷ;nume aR]Wp!`͍a90J(/eYsŵ+loܔKl(DrKs|k|V~6;aQN&|[.[7s+䎸cVB^Xaw6ލ_]6{XۨEYmwtЦI~n&z:=7ʛw݄n0.TZOv͹hޜzQO1jX[N+ F4}WǡiS3Ehςo[I-ccqTJ5&+9f>T ɹy(V5^k#0aU.^L&m1v|y;HPjɟƥq*]po~k_1{mke%/5"f#jGbmq*]6!9_~X碆jkoąٿ??LZc+64ˤD;_.E/Fɜ2"ٸQ"s@g}Ko >}rzO@Ub Q/QkL%H(M1ۮ'O9_N?m7hu 6Flؠs\lƳb}]1=X烐s..Vk})7X[W*.c{E[Hۏ^Ѹpj'gq<'\lcWА6accG %B ^ hLO8ia;! hvz;`zRTSݹ.z3ߘڑZc/Z4zRU! O &H!$l)͗4/خrib)b^:9)Rd- +OT{ &#}t*?2ﭣdK&b,3WB42Ɛ}fLNƪ}UI./6A>F**drT Wb0$&KT͞Suк 2atΙeBʢRJbI J6&1H+2@JT֪ʲ{]sN0B'4PVFJUa~*2Rc!B~Tׂ+t\3`gׅ*냘TsqQ:om칫M]cpG6ø{p̼m@m;_ѝok0ees' {l|#F+'˥68o)5(Ve{;PYy6eT&^OTMŗbOY[yZ<Hδ~N`*,v)l-x&50eH 6 TCb7)+r Vz󘴠#.O(YCŖx^7߿m-ǒ mPݯ~;;'Xhf^o~#j,*Z6[i<+GԝCiw &myHEe!E>X|kӒa &vrDֆY&Ad#7<Ґ TYڢϦ8[WO9TlB%4 S.+l*LZܞ8mnU2{4WJQVR(z381Y}d`޶4zͺN߮9V9%驺D~vF+q<ЯǶ?lЯ=;="63rFHeTd^4Uir*}AiUi]e?ٻ6$W}tJyD!=;@/ve #ODIm,IJiXyD{{ͮ[T5~DAMi7 {0s'{\[]%c˫vYl8!qZ [\BBiJ()P`9(_jPѤѻ<^oI|: z~2 q)-v(E+"~ᾗժJq(eewe`Y-)iZNI9'7A /(Mz6I?zAޘ˗l@U gL< B8+=1h_1{ &vIn^jhy&Gtb3@O W b0c}99FS9g浛+ '*W?û# "L)x4 }*5B}IQ'ҶMf1Ș1Vd O ^gL fw1ʺdp # : ! 6'>а՝UOջgh5~pv{6Y9Уjzӟ?~y}o4LoM4DY; ~*XRƨonbrULĕUi%;Rdr2WU`\C1WUZ\U)t +p ]Uᠫ*. "iQ콟J)MgޠҨdH`k*=rH֫R EseP;<$tEQc\4|}WUJ-:sUQta1WU\bHZdJbT;sv̕¸%aMjڬʏuZFࣚQvnyV0O6*~܏)hA] , ٹuYꙆ^w"5'qicqA/~n5Ha~C+_7iQo%ݑ emVڂDQa D לٗj)S@֊fAW 4^-:ajyY8h9sGhno25'>~rO>8>M8zu9Rߙ:mTbLNNՓ0G=hT'iROV~rgDR>v&,㫼@iW}w?w)twa)=,-|w1ۨI|.z9/J%wzjBWK u˺e]`YW+X, u˺e]}wFu:, u˺e]`YW+X, 5ixG'.i`֓ Li(~LQi+s} %|ψ)ߛ-e:3ۜ@rw%#VP2%%Ld\ x%Cx5!1`&T-j,I4!x]PJghi EI`oM-š_}fWv"+aU>Qa;@,Y>r+n[f jB pзYTl@rz6&S1(#&QSЪhwGTJ=΍ Q&J\r.hS(δ XiFXNYIYx! bL`J3![P^g;5 y hZUBv(?|À9&P{+,mV -,pIPP5㯴t1v y"r!W/ قŀ wI`BMYBP5qODod/b.F&kg'SjGKCQ㲼bڭn;l\;tJQjRϘ H LDOD yiHd^ E^ A{p9D@=I-r jkĠ$U(B"JiOċ֡b, Lbt< 9 ҏ$*mGe&Ð'HdvDto٘fިjS\QWBA1<~:iE"-Н%?/t(8g20)0/tȾQw07oo={)wH(KĒ6ΫQ5FlIxPRͶ4U['JW_7iNm]vs|zV5ۇz8 Fއ+(0yBSx`rcuNG%!D M&a@QG'kpk+w?+gWw.ۼ͜']%ǥY~<w*2u%MemA6!bc|l+g{s|lVRyI~g3(fl=J.K>kKW&WB6 0sNW, :/;}k3G+7Y^d=$dA ep":4ꐔ oX?raM^hwg'! ״<ꈿP"iq):kSb܂ :4fhr2GK1|e,^\ke{G#LA0Y:i:NtDgٓ"dU,I{!cqzV(Aʺ 5 pA %քd5'DȠ(Tn-Ҧ*ӿ;*vߎDB=7 u41gMX1'+ TncN 2VSvi`Wh@/Xz%i ^K1æ! Sl}<$As|T|+w-|1O :?,Nlt%^/8^/?;D#)fn&$tShGu"MÇi,˃ԄDz\7p,?q*0Z9 a.\Χ}pٷ+Α`8jgi-:"ASVcl|o5J؍Y,?Ҧ<6"yG_'4ƷJ~[Mm%ߘ:R_iwmmrԁ8\)GcW.uNDK"-ӳAKP}Ia3|==G<*q_ѬݐhTsQM:Y{y'IBp}wo?7︰^wo~f,1ҤOa,o@>D]z5Tt|nWMqmI\OIvߞ&FtA[f69[Gic2YbD#Rk@ob#DSݳ!׽;lhJ}HCR\([RMK Yj vK 'g,#}ņ׾@9nW1L{~܇3p.JQ2Sj$Â,9E,@YÞN=./Mm{uj:qm 6NoM+598=dgc ;pK-j2#P?^d<^4g4ēXiykGؔ"G "i2!@1h *pGмÒWsbt9Y]5ZFGUC`GH.ă*;rUq;r=4Qux[ ij=GPGx֤(%†m9wg˳Rdm aLˤeF<9T"BtP)3z \dQe+o_ <"269:("D]8 OM=L8{ A MVR#hPՓ7K^|4ϪXlUT(M8Ɨz(QL: L`HN 1{Y-mMN)4 TcGmʤHs>HFjFz\VӌCPWB5`bᵜˋŒOngED> c?.+G1O~ɫ$r0 ۻ̭3'J1rtNҗb+)$ˏX=!) lJm!1"S+'2ZK݈fgi<Ԯǂz@STArrǼr,CH)(4(8$3@ 4YibJ*!3@^tHV(ȩԭQqrJEbtZNג&Fu֥㇄ 'A V(L4@t.]wĸAi)4i9 MSN??|q.^~b|g妢'NOtg˓d'{AD\dn*-Uy4m)-1O|ћ]'Uu ;i[Ƨ 6ѬGhKO4ohhj <`Tle>Pp52fWu;FY*4JS nZ'?kފ_}=0̭xfh^6QP徆MXoELw ׮{ڤzC ޭ^|IkbDGMԌe3.ufzk?ub@gMƋ j7 .{:6j:Fs{%}>Yӂj]NF~9) 8sW40y7:&6o7bJEt7c{-}gjs3m{8:,_q񊔜l_6Ƶ7͠'siw\J=jP/8Ma4 gFr>>!CYtXs~ߨJm=㉁{tӯ[]}lŁp:(7f>Y{UmAѝ5 Mjfm5)=O#z+(6NBj)g}{U),8#9;'I'{R?i|q6*J8)YnpS[Oܸ{wtbGy'])|#t/v7>HY1ڭ;ε{~۵XxբU/SV]M֝kEȭ<Ujq,iM\=V=[Rq3i]pftջ2ڣex;zټYsn-ݗ7UPje Gxsi(rq6-eFfײy1+O=LZ'WUI)7 좑L3I54wD7vLYg/^^ˇ@4B~DBABvv^Oo;yhy=,]jҵR4A5Сsuͷiƨmj+`VWGm!b~hu0zOvb2ވmߍejib00:>.Gw玉 qRy4nxxc֏b,8+VKyGS#}tyj;HIޙ8Lzr43[GisrAɌ0,Tgƨ=Y|1cѠsDH>ǘIUk#K r*٣ a- }w12$y_t4NWihh֌C(.I_,:OȲ̗7T&uC,;҂֞,zB1H!hY 9IDi"KMrYxvvV2P,2 ΀APe"J"V<%2MSÅZɳRhx-Ҵ Z41cYY`6[Dc(߮`;&Ξv_Pr >&)T.qI/x!z^XJ 3<:4Q-Gb CL |Z=p{BJ% mVqK@0H/Q<8*otV5OwDB61RRY;*uM 0 =0=M٦|*qTtH!F"t=f{#X@Hsd,efQ2Zpr!FiGVoEd㸄ы> TPƷ!<"F0Y80KA.'+rpJITV [%̣yUӐCrPEЋEXʢ[-CE ^.퍗kD[$ qO'&[]'=W I*J2i ɱ)Yǭ"' I9I7.HiI~\Iƌէ^`zޯDpCO`Kv*J7 Zw5hQWHc`R zuTSN]u%ԕ丳A]6qHsDBeh):4u(U2N&nH_7HgN\~ tQ)& 3,1,ʌf'ԎHUƹL7%n%vж%v xr{T[oMe{|Gk9D9D˕1[ A M+>E "mp&65ɉs9{ڜ]OE h䣗I>2 mSvr:GՓ7*7K9KԔ@TR]F;"ΗԶ䁴r(QRs*h)CKEHkD ui] J-iWiEBa&2c"oF1ko@٠C,KQI:X3>4"XzcX9hP^Ҟ.tJyBIM4&ȠF5, ҕkP}`hzl?@Ħ ` ee'MCI:3ONX!,-e'Zv/֢K63M$8,: *9,yf!X|.l#h``Sm W6CqrtrY!8ZȍrUio\Db ;JV.)u׼xyIι'S.HR\`P)1TRvSdS*eG;mjP,3!P\ Tڢl$x3MvۈAvцky))-b7"ξyRӔ}SLT l#rE5`ʰBbVxAJA+*5ld9tqb);tZw4 R<3f^ALW_̡qZ8?r'&-u| kH|l¿D$: {?n +[dm@E 8 QƆLKP@&jbA$%M'"UԴ L-i"21z^wf3/?wl$/3Qsc6Qm ʚ :0&vZC⚋)-9Xyu?TO fO V' sR}&|DݴYfd{-B+0vbVtC3Tm[!Y/?[WtC*jƊn$=Z++M@`Hyy L3PI|2x42ixgb__6"Cr8IXPIir a9I+A`m6510@Gc)N'  ڨ"٨h%Z1qom,$nV=$82*e4õgM&$~7dͱ3^ȘSBN8:o L}IjE-7Fp<'4O`;HH D QJ॓2B`G.`^i!i49xq>7 !"0"Rm2d yf1!s5~s4ܴk1rDP : .i9sH"4cM~_ܙ$ `}^A>V/b&w $8̹u1){\ymn~F;xwwvj<*jE죏\x6^͇xszG%k.o̚js~| cSGػMȺ󴽐L㞶8JrbO%~:T&cF&ƒJ}0Bc]v9:o[k s搭qQ`ii,44j-iQ,YS1DJ(F[QC ݦ'I(jrsy~xy鲃/W46G1zOeR zs(cEmPHsFb!)61F2ghf]S7w`g@"hK:ͩUB2 4Lz^Zs/5[[jӮt.]<ͺw|SWd,bm@PrHj0^o90 ͑@jz+n8 ELުRP*'4JO0 \51 BB}Q$@z>kq*vtiٵ.{5C[DKWbF,SgjH&8Z߈dV*xRB Qv"uwl[wG/~,{HI"6E͙T Bb"fyTiNbLN$ lTE)3$#yBt@-wg-8ɁuS x}]qx5_xu:v!?~ F BsSZWt?a1E :/\~ޕF|ש7T1V+'\yf:g/wAy?21ϴh3f(e1=J37?z6Exs[6&U9t@i.U(i;i$7~I m!1ayb\yίUJQ=֓r/.V%/m9eV*ugaQ8Lh*aI+> Fqzk͆[9{ꍿ^_xc%gI̍Ajvo]5Ho8(;d[n&3>M4vt4 mLQJD~?|_~?|#;iC8q F27: >(b-hJHŪWUz{߼~^7`q+p.ڙw"B^; CnǸrǸG#4 fGnnr+]oѬRݖunlhU^YIxvE%Slg <ØbIT}1/,R A; ÝN;Yg61+94x\Y[FH(򥖜N)bxЄu߆mWۧ%-mˬ]}(ǕAkR [۷{@Gڟc"bt}}p*W~k< Ea\NԿq\ \9Uzo~6܎;ܗD3͇wO Od^m˭ܵy;lP XQ@=4ϵtpȥb@ӗgxX'.8r~ 4G%UA><3l+ UkIQI)%3ZX:c&zGC<¾4՗AƠp= 1R~??~a1⢱' %^ Ay3En4B)YCšz[L*npZ06Jpa/ˈI uFc ׎GqFLȒ y]YR!ᴫn.Z`ΗCݦW9:[Y'UAG 4a,&ґLY" HHN0+Nc!3ZPBkM̨3h!ЋR*嚃;Kp6r6KJ?qQ] Xw)j@b,(yωF1O௕DV{Y%g4B\Y2+Mib%VR(]Pj#q$?8p+~  r nX{b%1H/}$饊c J+͢G~4Kyqse4GGЗ}];Yʼl$ȴw<cQA6 BF"0*1a/!rƃb(B!bcQkdJSC}Ŋy/%8 yp\LIY,-,xGI4 =9z MEoNK-'>􃮄VVh^V7EQ3Io:3)Nt?Ϭx0 *$x6[(ch &H(4{LK]3(Lg2өz4[UX)G 3 z`Xɢyny!y8NXs|R4qp !(c4(XjpH08AIABWzߵپG!9CzS;%|_ʡē7v_4uDL#?@Q!8NS#:MNimƷʶOdEbD1JK='&a7 X# %u iVCu!&/XX*#rQR [^#nvvt a]ʮ.~8s%ec]u貗V̄F̗rWk==]6:ZD9`(CWhSN[[!U]Y_IX|љ45=\ (?pr\eiv.|j6xe45 3ÞIhBE]&\hwh6m~ʹeψAM=ƣҎ-0k56eCgަj6QJOkjCfGNc>$c8Yrzp]RQR8AHȢ ,Z$ bŜp/#Siv͸`yفe!,U@vB8%H*-1aCT6A3ꦃԹ3Ѡ(ā.U1"olY1bh;GCq~/RH?e킯ݛ/s/ڴ;bߥ˻]KMՏGvސjiq1pmT]Qݎ& y: 2)W(VEr\ Z)~Uqc'(7 ` |f/5 =rK0o'V.pto- R-k8->_07W xMQ_$~ .Ǵ?]l] pYG`-'qhtz]=8ǚe>f|R{mVWZ㹭1H nu#R O E+@]X>j:~~\eVNy!ʱYWJ:}7+$c_v3)t3.yQ? r'<e,!]Znu '1beP妑EǸ.olCZG軄N1B VЩOŲcաSgM*]UrXK>VWO*\=t) l] I{_`*Ӄ eus(MmR[(+b\[@{5̎%+Q?BD>}e8*(,W| Rǀa\ѫ0r{svE&vO]."joՔKl%:k7ggtBBH-ӶQS!&("2߁<7\m|WK<&J2n=\9|YZ[6LJ,#H^w4DwփVIh~BIwfU/|YnỹPLS_OuZ4_2P姵~[k)z:(^y%pFKh&q7{r\7I0<̡_o=qKeI2wKE/G4VƑh(x9 :z@ttȠ[4M82lL2ox(<(| %$F +415v(joFI*NfRyy"yJ!nb au6r6:HCfQmJx J-3akAzibR8&(3: Ήyu+¥&LI5F.Ie)AJZGgrGu 2HH:E<%lY& WS-VK%%{#A lp6aVvyɴXőp.uXE002A!<JFc&=jQʭbriK؟\Ɯa8 5^" baŨ;YEAl.)%0=Typ{.o+]U9 n/B-Qv?gy*op<\v8azv=֞lUp;n O=cF@0LLj9?{V(!@}(Зާ&:k;-KVl[pu #/3<$sP$L+{Cszn۸-~ٮ~JdW/ʚo>h~Pzǫ?J^rϗW.,ڿ*v~cxʣ]!z6o])TGsl>2?8mF5qc*{|/Fow"^Z]w/6?Vl\xV1U-vAsIn+.xYȠkK1s]\y3mi[#~s?Zr\}AsUXC5Ťic^h&}&jm+4Spy3jusҏS6Ej͑lM.\иo"Ah!Wۼ*5LrB'qX=$ Au29O?/]-7}z\Ogdv2>x}u%_4wwzY} xaC÷iΜOxwхKsa}TRU#|2)RU.1`8~|]+x2k[WZHZ|k]-Bb"8',WnTq, e&ɑ>[Elwg ͛ir~w+s*^\\}9NTBQ%(bg |pP%g!|( [ \{V&S4p.Nb;gJsTv~bbO&qb˲b΅ړY\- ~MY*B8mPSJ%묅U6%fѵf2A= <-1"dxĈAemRl<|FL<\}ycc&&[/1ӔTKjH^8QZReլLI\sÏŹp2 q<|msHۻn=F>?=!7(֋ nY2F)muB>6^t북s쫯:EX7xvY?]]9G _ޭ~-A*I=֣`?e${E4QM~sw~\~ÁP?WuM%tF1K#?o#Fm˚$BS+cLZq V!Vfvղ^ҲS1[BZIyh}lϹ:IH9ɳ5nN6%0`pS[O8\ZPYghEm!'NyGXKV4NtKfLwF]nF'iλpy\u2 0 _pu2  0 \?L D~*w&\D5`qt:W]%/:E\9pAapj= Z;`t);rj'|* rI(j5W]Ypu Κ03kU+2 Z/sTZE 0OlzJ@;K_ W zY\023\p\ WW].(jEW]\.{6,?^{ջ|*\0,ˮ#vx0/ua*^;ئ ~\A!W]s?~\u\ 4 {? qp0 Z+sUW) NWZs̻tXa w)%v_gOəR)&zLP O/鉓@;\Dr]sқDr'ɱf2`?Fj={\u NWV;@`̓w2c6WPsJaz,:5 Z掫ʂĕ3ڑp{7 0 ޻*Sĕ7x`apapՆٯJ' \A00 ܗwZϮ/;h0iMo+LŽkzṫ'uZ~Ti\تjs킫c[U ZTqUZpuZg_xoߔL/gʉh`oq8P~*gl?S)6%"sky 39ܳrca0= ZQs4TuO OlWYUk(jJ4.:A\YQ>papa`+yu\ D ؿY}}*,sv+ga \Ap ; \=wվ!t#YoϹ_hk6p⢗$xD|?/\5̑|zD#~9BCAwy On U*_~Wj|6j+SQo>GcxSI ,FJeM|$ -]F|؇/7_6uW?V(fu~Yr9nRշ䠛J!ZF8I5L'ߢҳ\iKrA|5UH1bQ۴Ŕfpa!nNFۑWdA!-WqA\ TA6ׁBq-6T"r-e {rgR5 ^)ZJ&\I%F-`bI\ҲWm_1cB0|!w_¥gKYKF\H9. dM'e]cU> 20Xl1 E,:3i-kl1)JGę>bT⊓Ek}0=E`&$Be`6g\?W֤ C)azS'ާ\UҚw-";Af{E;$"yw>fe!X+OdHъOCW:2XEӹW^ɩ.1ڜΩ$G*hmLr-PZQ~j}9m"NX G_Gq 5=D4H*SZZ킝RA5 )"jOi4BB>cmp%'UxX'2l<⢧:d\ xq%dR13V$b\T5K1:*X{R1D va*VaLxp'[ b%- nEË RQ[Ή~siSo'8q'ʑS-^lJҫ7 Krc8%´ѷ2Ar'I&APP- MűnJʞTpT!Q30=w^: Jڬ`4,z6/ZgX@[WFkՂR[9wSm'p ZLjX*1%V*`&T pe/ DT8p.*dr}C$Àwfs[4/xXxӎ` #  ,=XAEgdG!@7Ǡ1nXD50Ǔ,+xX)lp[?~T v }7+IS~/1j{R NITs̠ؓSPNW Q! )FeHp=) :h_/Q+c囯yę^ Akuj`g N>_#9TrwН /~Kyme|ެ*^m;!zHΚqHМ'6P = x^wW5Ql`gXcm=#RL{n")H=c֍a"էn:N< ,*Ҭd :AJrI#J2a1'8 vc z/sPqᜁyT_V 5ȤuՅ*o128pG>A,TǨD>/AXE jbZ1ة` OӤ^dzi[\G$Y:ajd0uQ=Id3 .-zL}ӿ(bEmC]k Q$$1dɃk6P' 56"}˃Ҍ' 4N zE׃rh[S9P/O=U/-DJj]T`=@ AjЀF1J[$=|"2`ʽ{ZÎz$-ecW,B?v9+AISL1^@) w:|AW4&ezCɢR܌"1CǰG=@U +~xDmDVc- Ωm X{uAfjBct+Mg0jΤyAha3BBv ZuZb?TkBT4|ś"A[_%pgZu$k  VPxVp*@#-]!-l8Ƌ>A!W&E4S"qc8rƇJFDyPPkBz%bIPrm⊖4 V/+^#HP";XAc6ՔFʽKC$`Pr蘅G`trEP$9Dݩ KzP@Ւ[ 4r/O]gF@ޙGF'@_-Nf-n˥5 S$v"w9n~֠C,9q~ZW QwZZY\9?ZF?|1vvY)W "&Z/f_FGz}cэk.Ƌm7 2G tX\Τ -+]/bN%-&޷qq{כ+GaG϶ac[ާ8O&]\}2'Zk=]Jꏓ^G=]=pF=szܭ,qh3ҕ|DKk +v_z++Bkձ+B RQnFxBy<0;y=QQ(F)~j3|3Zt6_r/WZgia&-5rҿb/\hln ]n ]٘Oy0 s_rP:5_^՜M,f}WG*oUʒs׫:E+!d[$gEJ+Q۶>ԨgE_8bo:Liֹ3).wpO}-Y-B=D>a Y!Yפ;SD#} ~h^!cwMֽm9nG<磯>o5vtʌ覸IϣbyiE+NNj@sΥyx 6~s{))9tT%]ʭVr/Gi}''7H}*jPZ+4O[eN&]\mN%]Z=]J%+ڳ)JW=tEpɬB V{"ǶhȈqRbr'm8zvJKLW> +VtxIW LW!/ 91Hp]Z/=]K_bѫpJb97t46PK8J=k)Q~t84Cɟ(ۛ>ERHWS^ >tE>tEp>tvyJ8]t_zX;T޵ո̻^OEǮq<MN+/;#RкkNe:H9]Ր=%nj'U~\euw?7O:Upٲz\ )SIeTU]5wIZ$=pw:舘 ؿ4l^\ߗV'wWg.O'Gl΍ӇuP\YcmrW0mCYΓ^Qhw3lSiٕfdu*t^QNS!钢"Xt*ʦs2ybS?0[Op(.]dH:^ʩ^>n>lb|s1^߈M^O V?^]ovKu9.W >}a;\տvn.]/.׬ƼeOWgy+'ۖYZ*妺w]w ħ[-?'X@DAd{_->ށT=mڋJGyل1X&[Ju|i؃|'lw'ۿl52҅RRojƗ{~*Qa[~;xbA m8 hOH7F\=r(K/?|$% M|qZk6o^{| ]kP~z8aYi<7+](^٬::3Bs.+^VǠxǢxaK:,"Tl}qI9f`^ҵ'oAo-l4!Yk5s-)(iR{*S2Cr̗7|yhvny0w::A q:k){vz9ы1;Q{CBF]䃋<@l$Zj #zDY{??TRS<񏶳\No=>7'W~ug2z\H`$(ch迾iCxˡEwZ@ɧ&|q%+=~,>ո1;% V?]4|7^~}jv~1IO>`Y5"g+4Tay b~:%] U٤Jy <4&p#AsԤI*d~ۚ>M9N0m$SeѠj jITKmm;HL)働7W7!n}W4n x<u%SX$hqwhut,#ٷIqP'n3wZ¼T0ƒ=ǜcw;)mzrېf̉fkJy.א0S\&Wnkfw$x/ u ;W'Y (ךY +-;G]id|/̓.˳jڈholټWN-vw}s>ŁgqڤD訪nߍ޻icn:02۶<㼱U?2lmAdma^e˽6 >Lʆؒ+ZcmYF]L Sݩ'VP.Oe[0DwI朻!ՌcS @ zԗGme]W*\5hYjYwݠtm(]ߓ)]i}&M 4ymhWy;iғ{:]%vx)羆Y` e} i9h`z@ cTa x# QG+˳$ƝJG5& ~t'd R'zSH(c@7!DiLh*D4& 4P&KkITDW6F>Q&;*QQ$wAo!Z橏PHrx4J'PAPHraA-FO@=G47U1>j]+)cw:I%m~x'L~{|w5g)tHʨl IduŃ #|@ƷԓžRBrhaq@8Qk 8l :%*I% A42#~d,b/X(,|R,UK*kƣ$wL^ov2}_8b;hGo#ZRD&QmC4&+S䚣=12lr'!(fJhT&I+]: j)rGl75PPw j N=E8/#p$p"5@u3 s(F9Nb$-IHԔBF͐ZўGBFaǘB!Q*]9ac8 ǖ}QFD5  wr4{y"2*&9.~M4NK8 8cYf8DS!HbSQD4" $( X8*hI"3KG#~DpU;Ųn!u}qE=​wq<2<@B=!C8'Z'Q'I0cG#2ȀOŸP쇇Oa>c/]Npg~TQvKpzZ@wBGUSvNCm$SRSTǯ 8*|ܤH}ڨ格g,qɈ*z)RTAr65D4y D-*]j`R &N &T &eXGhk1b9wo͢tH.Xdzr9Wط{{oM>;~t1Q(^p!@`(9^|"7Vȸ(" piLi@'A0gƓu1J搳dd$IK F9!\ &8!( {C c@3CMP3ə9,]@q1r @==Nyx5_&v+mlqw7Y)«u75}<a&-:JZ22H `!Px l~y酞y4q\GjhR(^FD&Ec9A[?|d>gFg/ڠOSxA֡$9ni;HzbRelFaZG !+C e7v ?BR 5 P/f]a\0m)W"J%#S蹉h#zôx-HkM=`|״bvR]ĭst|yvfG$jttaHy`'\l61|1E]q(Y9_87KLfB@Җƨj${±\..ҕݥ?ӣe~d쯫?7UΫwf?!ubQfEL{J )q}hυ]u>gY9vverh;= Kc'6;<1~ycX{x)ڔ~h<}IaD~$p'=݊ncٞ.,طK%3x>\jo8  U cwOtsYJ\ô4r-}1эǞ'}^a^]8])@BPjm4{oכ-7/x>rtQiY11 TΞ\NJ;(Lϵͦ{t ]'& ZF$LP~'b ocVZ-60{}@KӽSv[eͰDvjݩfnrq_"ۍ]IqN#D)giqD%rD!a8/q80!y:&+pρgjxd* @6Y 6Jv:O`s!CgD+z}:y18f,ipnxs`Dq {?-菣R9HK2I8RDE9kV[䬼^ofyιh2YKxкkPmЮm횉NjZ k+ґC:~Q/4N,_jՖC{.E 8@?TmX'^m7UhBnqe]z1G\W/oWϨR7n#:|_ȃ PkfC ںj8Qm=K3yxփ<MHM@f ;(r%Ǽ)RʔRD<ֆe|MW%њ!ݍՒ~ʔB3)q;?ify&w) u\p nuR YR8mAY"RUcVIGhIg0g +ի[.>^˖-.TTfCJGsB g!*P|Y=9C,iuTMhk?ū#{ec2Ě0Ϡ>a 8Cͳu1y<dOk5v5ުsGt|/ڪW9 ^Br FzHOL$$?o݅(.5L3`0)?E~*jz< ste}̧މuH-mHnJ8z{nvl|s垰[U]θ+~ܿ-IڡY>}u|"?Ov7wamdV7}PlY&A'> tQʫO+Xy DQS҄8W&E2ќ +8۾&AؑryCq6.4:uPoceMοެv}zZ5&,D Ǡ\m71bu\%YiY]'꺻h|Je\'}IA[ﺑ ҫo?^ȏo'>o9o]N kOU;Vn2`aL#OHySc]U_V*?R:{d&Xnsv7:_={ݝ[Jy~]/vymW0+kkM.^\_J a)DЄ"/I6G5Itn|ߙzSwjn˸oÅȅa6xy}gU_+%\:O?_qQaJ;O/T3WRj9KYQ䴍[rq.(oNc=\:>$>t7;jakgA:F\KVtH%]cZEfHޅ3J6TIϹd>2 sfN\ d59KHkscN>وKdL1J֌8;L9@rx6lj*k٦bQX~3r@9eRqhLQH%3[O&q:d= ڰj?I*&;3`n9R>@~5(~]q㹹s0S4:Lv)Hd>tx/zܓloV)V Vd@S.%OV)m9͡>ՙ0bFW&^U!W-b<} g8 (toZ/gbB bMtG ʢGd!.FlX,;gX=͍׾'{%}0b<}; "X@}븗^; ԟ`z΢]8+Wf+Ia/+@ iI^9,%(i' /lAرjsX't( 05~*0:6Z02[#1+␩nn/G!fm c I 'lMYٯ]M#oܒ_{`o߻p;SΏ˝ߥ7m5=Ds;^^m" gk5b{{1GL8& Vqf2bʢ)hQfg@+vn{k{b4 3blH?]N]VVz̉7Z}Ɵlm ^Oo=|v{7amdYA;u:}p!|5%Z$ʎ&o'l& *N1D="FTmj7LCM"\&}n>ITc)aQ&(V{֤ DJF˂eU6X`aΪŊb_lb)i gTtNN4vTjr~l=PK&=$H}ZfFݟW~]咫WYA鳍c-`GqU0n.DSuqUwqqNWUnp%r^pUjjWUqW'+ֆtOVnpUb7QUcp+QԈj=E\y W"ؘ~pUWU-7+Q;W^snGot7DoS }˯w^/ҫ9x9^XTOwyoky?ή߼~^v|v?nO&V?2#Si,&HOE ut3Z?e*Ypp |Y-:_do'r>4{E;h`LFұ{4UќG#~1w4ws= }b{7̦gvu>lM旾P?w?ջ/i^ܽ&I3 T/>؎uiLFu*U/j[tUs80}:F4pGq\9yU:\HlYWUપ㪪d:A\YD0WU.v54UU:A\92aGrʵݜ]Ur\Y>a`԰1ݍWڭ,R頱+\+:t2WUWJ2પ:JCW'+R|$/d |]ro_ &{ک22M4y}N/q&MF;,Lf=0)<Y)khO@7Pr`\pX U-6o5TpVq:•pU>pU"Һ@r"9BFǻǕ%JT:䁫 ?.l6/ L%:覈 2: :kն\U[uZ)Ʈ^npUnpUq%*-{v+Q{ ZQ{Ui){ctOG"*ש^pUz:D%ipEz:Wޢ\&nm]-6v%Q2ИuE pEW=xoW"u*w\ejɴ遫Umt:,{.䦛I-{ݴG&<Y3:3h_Z4~#Qxg;П/ṯ=ڛ6COOU0n<*vsTr\Uqxr'ɉW"~pUb7j}t\UNWLeWVkWU䧈+ ddG=w+Kp ڭ /SIvq]WQ6 Da Dn8Ruubm{rk?npU3gP:5jW'+] 4~pUWjWU% g}sen .k6YWTZ.:tj"h\͝EjٺqBk@WZk۳Gq'0jq7ݞe4hnO[X֣i}ԨTkw?c|:ӣnR =9ߗ^634Oj`U7rU1]UpsNӈSlS*W^pUb*iqEZxH0#w*vc]Uy'XTzw^+K'\`k\U JJVW'>}Gw*^p%j4R[+L9U^p%jҭ㪪Fq-BWyܭK5Q{Uiu=rݹl #loLjZܨ2\CW= \&B/j\:A\iaWp:y0:y_hIkw,UMXLc҆g4- c_=xC7hYܮzȓy2zڭ,S0<s`%tgHWbƮ)s+k3vhߡp0ϒ'KY ])BW6}+[^Ur~I/h]ZROn(7CWM^'< pOLW'VW;Mn(M]*MOL$aAt-./Sm}[@W߄tm{Ŏg c9H_nESSN붖^;DܱSr<쁵=v &L 7q1FڴGh(JG.Fy?Bٽ(Xq򁿇E+ iRhh'E)@ϐ;ڃY;fZ8uRey^]pp1 3fhjoP ҕgÒƮD3rJ2-mNWrv;7B%]] ])ܰ+v͠uJgIKd9CmP*Rv͠tU86!Mǽ4S?A+`']xgOhd)НQ}ZXQ=n=cD;ˋ+cW;}vFW{ [ѕbdgV҇NnD $Ys۳7ⶁؘIDI"#zޫKH.dc$|BkL15A#}]jcmJź~qҧ.NOxrC8l#6uů/~uܼTj~ۨVK.+.WD^`%DJ_||<ͫ/yzW߻G]o -M|Qˍ_öHV"YdJ|P +bE߱}>R>;])ʭ%;t|芣PXҀ ABWU ;]nz>t*_KW>}"+M' ~]-wR`#])-I]Q>.m{Rҕ,8] ])\ K+{?.)z>tiIfi1wV{JQPoMO/ι'?;'VW4tJo tEzlӓ>'hDW 8ЕM+%}+EiÁ!]Y} +N..ۥЕt(YU1- J.r*1`%]Y[z] fw˵_ϝ.e][Y[ܕ֖k{;vw5wkY9ޫ}z1xrz /׺Q_d-tQ6꠿RW]YϡGmc`;dT;OOjQ4歾Z߿A9Pӕ=~6G]msl'Ws=4qu+Gg1^_]K ߜKz~sþwݹ7_Gʋ;*QPickOV꯵|ܣ:5 -ʻ+23p@|ueymf{#- ͑ۛ|ȟt[pZꬾ_}fƿ_ևzG9x)Dײf 88S:?B*e(F+_8Oא,>_G߯ОN._&_*WV(|unrvd$%qnSR6.p S6>|>حzc+lɅ)vTPs&7G%lmLr}?j*SM:ғre?.4HxP[c4"à XQg)$ Z 4Bs0ր O+-'d0P>д*qAY8![Fho Re .ͬuV6PŅeQm݄hJC/+C@p`$yc6 6J1z}AEۊF/m {g r 8lkS9x X7g\GZD*1+wOllPa:\d"dJsYb$8(.-"wC+Pj &6 Fیd2 \n^EzjPB]yB߁#7CAw^KC 2Pư)0m2 19: ȄE@i0#0jM~:lo- t,,x$PM;&q b`rg5W rs$%@ i PfҠ!\ :7XaSQE@7GRF WPf;!J` lAPS.0b_PEL.UuAT"ZzeQz^$W˅sGH T'%.o `3dA,շA.*\ׅF]bA |f =aPur 4tlpBrr|3(rbJ;t'"!pBBK>}m@< ئ2ƪ~WoF ZEGF].0H6#&z3x:p4aMȦ9E jm@hLUcXuL. '4X/:mVqV24ITLȼ>X,L0/1 dA!Y#e"ԭZHw 3hGx.pu2PtmOU;o:[ l;dzeA4D4X wW r"d&{Rj++tmi,#z,%I.O..1ρ 7D2RS}I*q8#(n_Ii0e(4X4k&!Վ a<DG7HX}Vvg$+d Wsёf nfƐoyMms(`Bx7:zq?6ٻ\ݞA{AH7?u9AgjOM5gkwG(8w6xY:J;5WZsLdG0癐Fh4vfcgr|QӭfidEÁwCD) H[N!U C{ wŹ7kD{twWza2mMYڃ,e( =`=\߼YOͰW6Ճ'~V$[RB0i;zYyTDyCaxȹ kUxKgf$U1t֓T,k~xk*mL11]ǀt⻅6vR;k֬U:6 |d3B-d_KOkWtGAц?joyhšVqdAiBҬ5l-k>50zlY?-Д nFII45`=]GCX4:%قn-U; …0i}5fs˥:pU.I!!|1uPDAp$4ub*ՌHzw!B` `B Ƹ;:^:O\rvvy5z/<SV ] ]:Ϩn}9}m}?@8U+vɭ"UH#xeSk=7>Sݕ`wtVsh͖]Tgt˛]Y'ɛtzz;,| e~Sc P,LSA"&`ag xrsca$bLr$BI $BI $BI $BI $BI $BI $BI $BI $BI $BI $BI $z$xe?¡응@ ̎iLH 2H=c5xh[6YN\xA௳?Vr$pW@o*L'}Wނ} }}Yk%4 L9\嗡 6)?c[w޹͎FR} ϼ/ɄN94TfA|.ܿNFyX$[& L)%lSZ㑯MT(%9~*~Bv;HAe᧗eJ*v7z. vJ7UA+yZ@ tt9oLFE!s,°@t$,&MFx^h-1Q^ xfUEO-iʼnJKsRBm"൒Mmgt jT kQV-5WΆC×?4.Ԓ fWXw[ͮ墜kSA+M{<$񵦤v^/ nJᦴu?]aW eEq]V[g:;zv ,1+$ㅳ1㛕L̓Z`4 8B{l|WWx:>GŬ3o7tLj C+̵|ϷK[n+Pt[#P^Q៻=n3Dzgj~P?ݒmE%io^(蚿EuTfyθTӢ누 lB y<)qJî;\O`G5tӗӹ^XVtv.+wB VِbQjXA n`6V)B8S"1qRg S0%Fi;VYFaZ*o!`1́@{ep0-{1*Eg͗a]fG*r:7cBU;pW(t?]tp}j#zGOi"[8ƒyGqueKb0D^K+Q)2 K49KI!̅A>.z,H=K%rg]K+(HQ{&y7.ZY͝pỷ/,ϋN].N%L.u+tH.KI[v$zR[| w?~D\NpΓUk3?MZZoha"!TB<۳ohZAR,!씚S.&Oh7vE3y4h6Cw=]|1Dt`ӻvM7bbzˍO*Ƌ42Gx %]RDZKJfKN<]R "gn- XxbԺ/Jp !#)e4F>2r<8g;;g]R kEI.)zm1d{l(>^]t'kᵬhg0,[bE+ i=_Aܺ1m[4w&ბT {xm\JqQYOWpWIzOoK;&?Ynw ˓, h߬Oq@>aǣ>4OcK +P19y8h:er $ L\7S$XZU8r9R;asT/o}w,_D i0RuF~>/VFzm/ko9=M+;pRmuT9/F;Hl$5y0XQmLdMI}{~P]x5^6?d@~%̕˺oYmkq$?Ϣ?[t _ B6T`&!tNjQec/Y|4|wt՟nigked{FnuS[,K|'ߏ&;+C2*h6z&(TQP:.?/'?ug:ͫW\P]ś^/ip0)le{ #}U=UQ]=U&oi{UVl-jq~꺦JKYw+\ؑAJBP難[=+h՛UD`H{^,+-TfmVY(,S0HQX1c/6M^Zqy[M{yNm8/(Tz'x&!("Λ, Mo:vSoW*]$+ym=XjqؤTED:]з<.<-X`ħFK,cVÄ5wT[-Jbn)Oe2VZb~^|p*|c4dWq7p1~Yq((\oehsW=I%.0}q6߇8Iէ+;^7ݼ~> 7/Quk O)~J1gpa~nOθkBZXv)it:KD.5S=K}a]l$穰F"e[xDŽG}~ZEZ3ea&NY Br ީ1b"8WX-|~dd9Hw6 "P7r >9S6;Cx̋AuYaUx 1̀&i孰"BH~;+s;++)lMo9%$ yF/ɛY0Z`z0aey/)ַcl'3p4R;h2`4Se ΜODcƶ[ºQ|(} 1u[}.CaAuxh!/׏.]Mh]VbRTB_1h1WbyJ֡ZxkJiJ|rA̮ʡwږ"ok4][ X6\m}Oim)9CB!q;þ6"8'$H+„v:]_NPm\4<{9Gp/A7t1ŔYƏ$q4'DqbpN֐C?IkEtI:"~]`/Rgڡr1XiL,◼J騹;Mn2tj#/r|Zn~ zE[U15WTHq5~Ry=< rc:澗/*M'՗qLTϥ>3j:%7dkz_)έX>\rb8/" ~ٿyO,R2swS!>T2inڬ6EL%-Ll֪=*~Z%2Z FLsnUYfZG2Bi/{ƍ%&i@U~[$SM^6/-[TDʱgjӤ!DF*U-+ktY)&{(6NSvkwdzu=6SjY>~ziB@1)=RZC ]rwM;jT X{j=xI9PR~Ŷuu]r`5=wBj]=ഥ"֥fZ~i=9W*b,/5FET"kx7Ԛtr&L 3_}(ol~713GWZ gJK7pϏΥżoIc=\-.FO\JSYӯf]jpՏ/ sYR;}HIaA]`[5V7;h[W>ͤլՏ6fo_W>Q|$~Y@>KYkOh!!\|&@TP2xxƵNJ. LuIz`T1A¨1b(FWDMcLJ\+j8A"C;i]WLeW-HWlZq,bZ^WLj+ez^7IZuk[]ѺЍ(ҕMЕ-zl荲 ]p{}Yb\ohþuŔ+c4|j\6TrhJS@GGQ)s%ۗۂiiƍQ;]L h9\]1RtŴ SA^AZ 29s^ ]N_WL誇+ g̎q]1m(uiUt#)$Cl*]13+ EW}UpA(HWll+1cWL1w]1eJ*zk3p+uZLi3Kmz7RU0v=v: ֪(sr rEW (HWl+(EWDSSwS.ꡮ1ZiN]-Yߨ\!(GSxGJ %bZJhq| hqHh6*{FÔd4h)xßYo3GFNk Fmxi"u M33Nδҕq>j(L4`oqQ6uŔ1]PWS HW ]1nRtE^g߻bܮ+:| ]9׉]1-BbPzW}P^k-g'#EWL /bJ,>* KTuȚD\P4\']CFb7 5IƵ(EWL w]1EW=ԕS9+q HB+tXtC]ѣ&;'GWpI&?BtEh1b\i'JۙEWU0" ]13δsSdƚ(iE[gq}+&w]%23|t[GELEW}UPHv wŸ(FWDUCLYVRW1+NE-EW1:TbJ]f□#ѩu4$-U%d63tzE#HW &を+*w]1%BUue g.䦢r]'=Yis:{[ĀekhRlu(N*)֗Wup -m5Q0Бްf^kxpFcer (&cܨdrV؈394e੏((IW LÕ3Ĵ.+誇rZ[A"`r׈B+誇tQNk׾Y2s]1)]I++NJZJ2G]Ի"` 3uRtŴ>{]1eng ]DW%btŸK엤3%+ V#.]|r8D܎Jhf02]i֯-zLy A"`c@ס]1mGuQZ튮z+C(-_*;%Sqg}AKJwROF%d\ RO5wJw'l{h8(yh{IB{_C&` r"AWliuèh?PSȻꊀ+ƵbzL!w]1eEW=ԕZRQNL "3g]WLSt]yJ+)zr 缘ʜWuN+ 6(FWkŌ]1m+K摒N$]1btŸhiC"JTUuZ#)$`PrtŸJ?ΔeJoz=R^9~!CpZWI^u{S"wFi3]]颫dž^{>uPX[+uN6uEF٢ )䂖raW=&#嶙/͝=,[\ВdZ7@ }AKZdY>M2Fh׊aZg4LMh`l/nß66I)41m_lZ)# 3nR,MZg?ĔXJ#)LiN6u.(+ h$]p rt!h9"Z_WD誇r3ހAٷ_^/fPq[#%՜*݋[TEѻxs~懿P[ό rZci=_·_H/OT "=z԰bzcpi/Z?k4u5]kl%m='g{cx\gĖ.|;V/OWl1}\Mat>\Q|^ʤ_Wceܑ GJՃƲw|z#S*Juuoo/Cd"=럿1o^_|,yEUE4g/[VSR8]Cj.k*!TWͿN'ۿg~9U VGb;31^MxGFoݡ14O'|{.^c@a'2kP5 U]sH1;K ! ]w?3O}KUѥڔhNgg_[ϛS}qrX[Zbg?-N/5&'_S?~zsaxL,!VJCdc2qn~MyuA53l7depQZ.Ooyg{Xv} apaQ]Ϩ8:]g}GACZa)MJ[϶OS[7qlR?zLl|:n>/w9x~*3M]|?|g)ʛGSpSß{>"TkbJ}ɼ֏uqﯮ/K*D\֯y>"5QfQ6}PrҐpoC|UR52{ Uɟ,=AV=mKf9AO~_/kV=XX"3z[~<,/b׳F(BOG\Sa;rF}8]Om+:zd{_P Är #P$bC4˸4ɺeq;r_BA`8_b˲qZRdW]jVX.磻)*KM"pJ!V&OyR< e)2j4OsS/,B {v˜V0Zv ZA^YșND401(@$IeX2E!Bs# KP),bY+tFub?ZmNWS4z%i2^EvveOˏkIrT(U/,b`ޛ`c7k+GPņ^ܽ̇Q?p^Rl^M|(Na"t8_,gѯN8Av,Ǩ4Ũ ,NN13=~BCXoPdp SvB) f(Ub2g!mٴx)bJ&jsGR޽[E;[0]1Yy΢zH7Er?-c5. b CܗO:p9pXNfC[G%eC$Kpc2uu0rX_r`Q` +sqY+ fb{ v:;Vu?Ĭ^RFb#r+ L+qYWaڠc5A7;YRV݃jUfvۃ9ґܥ+a&_p胾Sc ݡ3Թ4$ZN1!&H;nWSlf"c@!̒[Ww -Ta.zL\FB$b3HYzu^e%ƴ#OU DRTz,|^ߏ|\m<ן!J|&kN݆H >ٝy1c6} ﹲ1p,͑KAxu,_xiiT]ʄ ^y}H+ű ˥4r,A]3#y7HGM OQ/%鎽tR^˙^`KO*XQH K,GM)J 38ılYE}^jJ}^*KRYԗGOOŐ/m꧐`V?5Ut\:frqξı kҗo~|qz- OdZ䘧̪s8KLRL"a(we%>|* &|1^W47ף3lCX*71HQ'Kv/0nqV4`&~ @R`me-{ulVKQ,Bi/k^;J(TԼ&0-z"+@:p]㵝{/ZV"Şk-44/F)T^4|&yyBLF2d[_z9:h8 4ƬQK nHPEwtyO HU?[3j'dh|N-{ 췣߁ {-㊦s-y,~Qd3(!Th< f\-a-Yʔ|h0x<r.>f2$D N_?NR fO!K"x]=)+ l~~ <9d5YQO1pQf"\k-삣lb @!a?-/~7:%k9́8[y EH`!XĴ=̱g3Jj !\8%ʏ2Zkx*2sD1E0DF ŝAKnHEqEUA 8jr[⍐L"iQ5<B a2{b"PiS'қ8- PsU.(͙R | CWG9AmSW)9npb)0n!F`!q*E U?P EQ@&pC|J'AT B^&.&, w#h 殥˴ME2A%4ktL'7 86mf,ɖ@(fC.\އ-".揤ˆx ?LJ%2]<-5 ZrjɂRrm!8!yp~{XqAen?h7wyiDD6Ǩ%Uo|#Nd}BAjÃ;8/Ai|5-GuA vaW2ٶmt+*۰Mm "*WJ b1uWמQBK %QMJ#Ia@)>俧)Ev?pp⧀`\딃ǯGtZc"zW0lOH .GTɵ<woq ҩ#yƫ3[vEv W-]␧U1%)e6V2.D|K B,RpENUY|V1wF@)< i ވ͔!crD<え7isRq!7%YeTsv No8J)|n$ 0:lxd$:.t8  !kw + #sף&SŒ ܊gr+"r9$ '׸y33"Xf.|b /L,ST{!9c0(|(^!_FkJpK5YE>/f<%SSw~d5F$2 (aQ.FE~̴͘\?~ 184Du ԝ|O9 >0Ϯ 1*3)`@(s'̢[>%Ca4yTo K>\p;%# !ԵtE*9 o"!T+@`ͷ"5]1Q}$b/7}ED8k୸?zoI<ާW?~70L//?~x6/fҘlQ I`ŀ&$2@" %K@??KJ: ebV aYXU0cDIi rSƀR301w#1LwTLliňhOkL?5M#ih(9&O O0A7{e\f]m>IRHj'u!tqeT$f̝Y '`mH,5 Ze?$L*Fi\ yFg8~*52y*9 y DH3F*sv,Hw-"\Ak/O:@0Bjs"" 34;PKh-M&qRSC\ۀV2]uV )jO'gp;9tx*=p ҁ(n䷮Ƌdw;[z9"B)OZ$=Kt5s\OhKaR 4A(A=i $ YUJ!GMsyu 绅R$1;Ok#XN&xkѲVj2 "i Ȭ˅T'.(+v7dl+t$+& -jĆ#GwN*G) <M8B7j3mJfsK' R   XÄXWO?S$2[yJK"飂S&3߭d:Dw$b Ei稜l\>=c0V No'v|\}SXu {ѶS8e÷_"1j޸G.jx7zYbOwK]NK5s%OdY9lNˆ ]=NZɡN9.;Lh]2Sxiជ<2E}_cXNGcX)HqE?vW N#4Ni|8u5V3EːZ赺@EG6Ҧ,d͠)(3q9Qf;Bw^[`>KJC>;N#T;ll`/h59+ V撵YsS}?cgN)s5s9,9;8CmZi82慄9%h[EoLwN>E߰m;Bn]+(.5C_9՗zAbd.NV!R(: +3# ޵6r#"8V,L/{jkm_ؒe,ɖ,5`bYӬU*e1LɝYvuv'e'ϧs'u H(б(\_Rafab]O {qtK\&qº އ4Lȧv" F_ŀSB蚑TQVL#j(_l翿ǯ? <ɾVR(93h۷C/5ᾄC_Jl47QTc$g=1frXrN5zHo |A ü9?A @O{D2h)PtPD4>/h23 ~ˌӕS0i0d[f/Inƪ$ Bef5;>[kQ5bïXxHF=DDL#"HI "԰t4M0 r =PCfjy'dfg"s~)Gݭ( D8cU]Wvg*E&mڻu\@oc_彧ygt4egst^\ͤ | *V #Z#6êи#(i¯1ָpOb i  k\i'qkȵӭ ϿsCby¦2TݷPy+A̩D $JRE.d DjX ˦6İ @n qtG'$yyrdNN0)Z]1K㘙 OHP@$hktr@Yeai}Jdy/e8E:N k\O#p %|SV͉j\t$w))Bl&}6яP߾L([ k\ɰ8 ku.ZLQu&WhVy6͕Nh5C*r&FVse{RePUND`JKBvֱV5e_$jWEuA"`;u=Pc'ڥYMrIphcNh<"4ϵ0;]JnL0Ƈ%MR G1a[B픦KTBեfٽ7 V #DkHuݛqb.)<Ǒ)3>Nc]~(ej+EkP,ѡUݠQgTau zpaM?w W|(X)宗g2̲kħ]+ٙmX}Z]v'Y07qW-MGdxx{4ď$#O'u& P"40{V9fOgi_7Zc>-iQASe3Lif1*<$~ "IjZYqVOd(Q2>xW>ۏBIUgREUJ± #hlj=fV5.7gc.Jw^]Qݱ ΥxQKz&oJuyZL`nm{x5dE:/M k\[U,![p\>Zb=QO Yo'+%&m[1{;^+4Z=_ D:CF oozXպ9:/<ՇF4;y9Ŵo h/`eLNl{ԺȦ!Em#@WF񐶸~і k\%CysgDx銑V ZM?|2p\6X ] |v4 ws׸hW.s@O5_ju\1"=OvUkl`±Q dz]JEm=BrժȿuW 59"Tj7y;Tzwtp'pEh+Q{A=?MGE#c k\_O>xV&GcJe'u't(ᤫ:ޢB*Ӱչݶum2b=Zǥ>%+=U*iaYѰe`(ݔz-_|:2_SL1|e8 UmLWFYr(1ck: ~A1}'׊%~?H^d< 8~P;$j.[r߅Q}gzWPTO^{6n/(AIaگ'f/Q~zqh [jw?eU=0s[~1@*k\kS#|IqcCiG <$.&Em%&QOl xcԻj mx\(,{VXW"+vb.I~\5\pM뜀5 Y:My"e,i\豑c__ V_:  IF"o_0|j#:pq'3Jꖺ ?ExȷMšE2 [§ VS;^P-5 cPn˴jwk"Ҿ__}ݑV4&;\, Zbz.i[$1ŵzr8ZZؼߦ"y+fh23$2`0(A4dO$>[GcqExX kC[y*d"q7G :J6@aps֨ȅj9adb٢VoNzğʞpˉ)9p])0؜njѐwb.qhDVMӢh{š i:ۤHc??Y$":R…B8*{.!)u@toD3"j*gT@A_U8qU5e߽8Xi9߼vTX6s^y;?~ b5zHoK*hq: Rcyp"gGZY>vK @ABI̕a,w[` ꣵ<`u@@w8#37BL0<ϝ\^N zEh6&oT1_ThQYTx*Z1}HW}ҏh C\M9~iwzn1_ݧWŰ_T1D_F 0K]^>ApoZl]g޲bAQﱼ6=܈17?ϤξnT L|Z4DGeH:Q aQa=#:1RZȄ+exx{Ã_C1hH8P7oV'vϋ'B`NΥL?Qo5#66^qyܸێ_VRize2Ԯ Kʫ\ߝv7fi3ʐcSn+%g9!4Άet1ZhX=/4qegEw+x>>! ֦+bel5:inrR!ЎF-w,m k\i`b[a,3•Д#θyQ!85^kncF]8GUDڰ im#V]R;[jV.i%"286 :;;6䄖 %㬜%jel<O'@'eaχڧUtJ1Z@ys QPW2Lʝ3;:z/ˎ#E(_BN&iHj9G/B=8m4>|7b3aWz5Oiv ,wRk?zL Db~1Yh}oy{0{r^IYcioۄir]S^duR6ٰP"o]w|,#:D=?9`BO e4S4YBV'-5ǤkC`QIJ&wݿ݁dR`{n';.5C\$ mwo Bmmo2)eJtGlҁtUXBaΨ]1dފ 9<8>BS-H'Lbm.JsnZzc:D7SA<}$ƞTz,oWPnTj4UY؇a"bPRՏ'%#*7WPsWرiKBF0W/JJ=֝TVRR, r(Qh}|^J2P"8-_?<9՘.^ٺG#0Ė>`YҜr-= 0n6'8`Fl0KC;7)|bnlY ~X!q6)-{MRJ쵽G4Ttρ ! ɩvF9#"ه'N ^Hʅ< {˭5IZ&$p칠VgŻ.1"TאUy Ũ5P͔t` (SiqwIrdcٷхM2YLa5hfK9b>fU\bӲɴ|XDh8~?~+YMUmF 㬌ҨT/͢oxX~feWa]t3벋4PM Z@@VF6r9l`|oy#!-zױ]_tKk4!tA[N:-eLQL?Omp}IpX2s>6B)Ad!J ]#V !F⼃tB5p UF ![?I|xÂp #w N(v`QECo jT?,D~5ax峉ᒒBo'~8P~sM"W;R*bPSo*]$Y]|~ɹ .brL7^j͎kNEСMl4P!X;hgAʸv@+@iVǾ:ɸL[Ѱ'aƊf3+U`Y !lXM lÚb9943<(慩դުuNSs֖reǃ2еҨڂJC<dž5q[ $yyZyoX@FRQ]ZSC79"@n'bfٻq,W~Y XR>vvSOAIَ=-E6eYr8h'C"%>UF9,cU^Y)ϯ=u<B ijn|P%~C_Fer `cfmΑ 8C 6}C4v1NQF,|+)6&vsz !\6>.*<0=7`7sb1zn"}1*Q=WxR蕹N;( n@I=1(<(BA3 %4ֺƜBMA3~Mj'aM|<~9ǁ)W"d(ʘ$PP)TBcWڙ_hrL g 2vt4PhlG-+ Ǩ!(fg¡LP^i%4"W>QH+D-Dp ɻ.1bC~ _Hh;M{&SWJ-ڡk8 i2UFwW JYt@ڽUk 7tj`aL7WaJ"@`l.%'|4 y>$n`r4.RM.j(gn[>I G4S3!+0 Z % :dj(:"&: /z܄>u6mKcZa[F ːAKw-il4WhhUP9.L)/ֳ]8}ϔPceq>'ƒ0lGj QDTn4Yz}i>]y,$uSfz^T}*lWQ)LSQ6w\c iҪӵBbۨN YUc'#L(fMyd@-ҍNGnJ"ЏE]4㼤9$_Zl)*W>o`\L/+j7Хk<6SoC gM.Xx9\w1^aKEZRn/rM:̽|i0:| jC{# IW,{wpGV YcsnjlZRtN+o_``Я_'D?{ i0s6z~vƽDS} @QY;@*{K/9pL\\ee` [ Ew 8gh"s˃'>q=>osH>}: ۋG1%4>MzsPO\ϩvo9>wM[P fy3[ܸB$[ Ns`5 3Vn: 5Zl!nC `=ܢ,Ugea41KeKe;At7);1{Q]se avfk,X# cy{TعTq%m'vM|E|>-aHSmvv}~'/4' +O^Mͪ*YcB3nv-(0SѪKh:/:ڢ}@i\x2Py$!cQ]B죺O!sIE(m^.)yF q(1U_ax,O~}S6Fmꄆ[x-@`;a!& !^8]Vy=qXʂ:.D11vu[Uj7 )3=(2f>GrnMfT(š@"p==8!CW q=:Xhm<9)TφRC9U7U A c:]&1&7(rY/54sp88u5܄xHRYb-pBm"(eխ.b/eqISws/)[hLa\E$. BFF=>`D0982 뤒IR+ۨ񻻗AybᨣiZ[1 =49͑A\kqT 2;-h?JJ',Es1B\-KC-0)i\̼K` tjhBgcrL9AETpdƿ~-@B`=ԃk ]@4P)W*)̫ͅRZIPy kArky]H ×UF_+d̓dGѮ2&t6{;E7-1TSe{hG%"׫vN(2I߮B!X($Y&XWC qSO.s %4-F)3vhTM=8Uwm2݊]0Gݦ7E`lMK < 8TLx,47-M4-JsKwI|Q!Dv3|3CU1{r)rc8Ei &JbKJx>E'uR",VcϜC@g 3 Xvq4)A yfbY V ֢U~jxfRZ/)D(֦Zt0yZI=0h,_ ~ޯ.MD7Sj?"DU rd,?%a`#rΰ`P ѭ __2sk&ԉ"KM٠'lĴM7K!2gcq64@7ed#{q+XH-grc:Q:YO\̘'@1ν bq?a_'x=)07^X|8(6W _?~7>)* "{qOO'iсs8h/>*i[lI6ztAXsogϟ{S?>?O`iIS_?.9:M bO)Ӳ//~n\:hgND@W%sBIP 96:;Yιppe$PB#XܵX,w pV s-o_Mkyv6K޾Bf++J D <_ˍЕ" yKY,fgz06y"AV:uI 81 GΊ`@샘Y~k29p<~Ḩ؋7dԵAyͦjz/{d`2sA;21B%aqTBUYZ] j/fb\LV+ad ykq^ jk-3~T;9?v׎gδY©f#Zi#EѠ/CI@Dcʸ Rby{JkFfh~;ĦP.k:7)4k)|m˹FHA ecGkU>]鬏'ੈp_> L=dl{ԌΏ@7呺9&h|S%ݔRms7TBdF,.<:kVصR<*^:GE 6i?"٘%<]r$Zȵ%Nsیy?~ڀp|$ [5WL)*b6fQSK1,W9XIU.)J@JZRM6&wb/~x">)7?7yfݵ. AZ'z^8>l8PBkJ_AT4K;.c* $$܌\*S{\ɲwH̀{.UҵE9,rV[<]0 e0 )pF6 Kdz*ᛤI(kvx z%Ց|p$!"zFf&,l-1rHtk)3U'ڳ~&EXk>xޚVO; 5VbD- S Ǧ=;MRŎ^yR@JI]N#X2FIs-a1xW1qJPyS|:a,spe֞OXԹ%{)v J!XWr~i vv+ls6ɀyS:1&]m9+"WB6`i$ Z+d7"yItyN-}/Er r.=CK(I2 M%à3g`E ьnZr*@h,2 k"˱;[2!1w$\"ZHi.:>gsu;Yrշ¤ޅ%4u3C-' doA#olFeYumwʩA ~+m 2=Lo'/6]%%4:HP*[ U;&[dgz8_ie&-~ƞA&X; +բM IQr=ݺ(RVwSr )N|UVB_Nwo" :y>IJ\4ƀR(}!fIV{_lONpk;^|X H2+k=Rgl"wg-sn7А2{kurEY՚v53\m>۝Jcʪz,i5Dy9?JAcX=$x! }]d5 ݯtd /9~ɝ,βhL|1Y[]L&}&xswٗUf/H՗Wl:j=,''BA}n!fY5ˉ?^7ؠ8F#v6F@r#3KɘwᲷ[?V1q =\ \e3F I[$bM+"\g6M8mO4tJi+4P/gVTVYQ@/ǰCׁj>Y4v E4p0RdDH3̔ ;b" 8h=;X(®\71+v틝@͞f̚ OdTawkhq1C`uk"EQקdWY=̹zeO_~~wr~S/iO!%I:/b~{P;X"5 qa Rwsd`d=s m~@GHEom+V"B4`b5/{%sE\H-|WiYКxݡi-ܻkh`p‘Iȉ%-l}w>`amcҀ()/l;Ohzy`9IL"{w%jw%tɮVg~gf1g[K9&G[vģG:%zd$QxSَ+Ȥ^ `:/'k֓ ^VjĮA$>2U;ob?OZ|R8ΞDslU#=x kD0LXFMAD#N&UI=ΧJ< HosHnWMȶCJ>/np:96~́eA kpB[ tQJp6WİTp :M;bmL\iOV5/+. MUXxhÒ<ηHt|PH_5Q 01J} qnuKĔy!:ƹ1} q!'qw'u=1Dv]q<ޓ1R}a4GA8)>c#k=閱|k侢 cfhrzȏM{dh/qӓ1lǐ\kT^.3vCoQ3A4(*'Hh%p׮W2]sIr{WWκ^[ϥiT#ov+苩hnRΐ笼 Yx1B}’Adg-S0Zy=[*Z3SoǽEAw\亩9W=Ǵ @zܾwe -<{8=Q_x辒VLǩj''Xԙs2&q"^<,qg r۸頋1Y;h|eh+YpHžr3w2;h $6gNԧiV)HI(L y9loRn2},sp0T5.8Of}6B]7ǝDqfQzuG)Mq[8ᘇcnAmDbKiSaYuͿߪ%{aSӓ(COuP٫y8k1dU SiA1St "b$>ß?'oA8{U&g*.yRO߽vXCcշpv g렫} z}ⰸ7܊N}j^|*.a1YE1F}i.VdVJNb˫○.[9%~XFh,d?}&0'؍gXEOB꦳JSC,+42/iX^ )uhL9 ɴ|o @{q—L ^Y0޹#i4*XFxٝu>l{a3NJHdߓEZ5蓑5t$51}Cd4h;De%d x}H XfqkU: žTvt$^ǝ(?<'|\ʡ^C#[/èB\@xT~ >egװTY W~ 7A9Wz] 9MjJ x 龢PҤeR KIEY1Bs /q g:D 8F,κoCz7in4-P֑)ы.k"*2K xE] h ػLa,eV~]6U?#U?svb}Zش^v˫UQo {OV`Jso<Թ-5^R[[C[*掶2H/swwq^ϫ![mݧ:Z(+.$=K^_4HN3C JeG #&19[W4nIy[jFIď/~c[@&Q?-ns2*d D@d$2Ƹe Grབྷ;9 pxk1%=/xqKY ;h|uU 4mwe "X٦[ܨ+ʒHa~tsh7>aCUP*~PHAms/(f""l^J}g:6@.+ ʱ`R* z(kh.K˲74 \mlapvѣ=I.mBzՔI9i Fe( +͉B&$Ƞٞ* D&8egzg};Un=}-}d<( l*ft[C/|BǨO;$?1]̗}R/rg7RJ!٦rsqҏ|HρR@̙eeK잃(e5`gѢ&70`ZO 1:Ʉ)CPbK|֨X2dgDU= +UO&Re 2V\|ĵыh>4ITVbUX|stVcќExLB0?sM/n1}w6g2n(A"ǫűRxcB''\ ꝟ䭻?]/j᧳Ba駋7G>..N_zeKK^s]'=4EyO^\׼A/=}Vۧso_fo؆f4_ؚuOjߢK8=D9}ˇ+zst9;G5=S ͞Itz7_P=.6uE=:Q`P@bMQEs\|+lr+NR T5+si1{})dQ~2؜!2Flm@P&%9:IJӷxL=$WFj~Ƶ̍z_X;5f|/2{^Qo0M̈́P:v+\4y8 L'9r=i}&|ÀCX31w㩶cCAD2V~؜,rnf1Ց-YlH^ {rK\6@>d? А!tP;t1G޻.͙݁^Іk|eC=WQ #ehUDQqa=mO&0[[6ynybZYN/.ӧ`yk]G}{Vp P @n*dxr;JD@jooS<'m VXPpA ҾXq~dvZr<2~yH 3L&iy긗:;v5VZ]▒pE. 9ۢK\L U$FJMVFߜzI7JnubSq)Ϸ,zUͅo=0ڜ¦g.d7RGa!ɫT 6d H"EO_%= (s`XgsbAGz\:,waw>_0WSV j}[bǽ%zKubHHb3{Umt1uh\31(7$[G۩7Dl޺;ܧU9`d󅁒Bdoэ&aI VAbQΨM7acyʵ\o "/4, *&Sh!Tp;%SXk/4(을^ukx*la$aDZ0SdV|2ȯwC =~7ݰTibk1tx[C]ot&y8;{aC5rAR&%RFf`ņr7PF7`71*Cu.jCd,;dadl٫cTX\q/Ͽmh1Ra܂ jpcQh:mAAs}>.jc6Z'ڙֶf.LFtiz WFngsϮgsؔ){6ݽJEw_!R̫»|;8@7;}͕>r31Ox) sZWDU _^ۯF~0&*>6`H6CX+-VRPxJzN&αA*TQd%^\X3I pnDw{߆g@o_G=E<(59k)CbֆlQW!!T|I6PƀG&Ԍc% qJM 6ִ$hL 䦖#41YRKʵ!V|M_IU*Qf ADe/["GYp]9'_۶(rUoRw5s*[#V7Qx] )c7H2ڦ}=pR^rWSZ1Jδ8 ]ggkcvY_qao[e\28c15wx) I1sR L .Vm jhB0u`[&U8[!S [ (A^!gI b:$3dvV]Zc9^aO'~_ ǴZeei/~qXnzK%T~S*„r• ؟ŃW?#u/o}{\J/⯧zMW^׵RgrDMUc˜׫9fc?R?}Ax.'ZQ[+c/X= ޟ:Vd*O>u򵆕eNi_'w7ح`W|7 JȖ" lt}ms.(-2Ed8=R7v"$m (\Q@m|2̔$FZյz[olɹlbH:I)@ЛZ܎v#ak-z hHr賢0b'dbM4 c9%:[hE)(ҾL)]0}8MaT/&)N꿐|2)8 5GH16r)1WnM)E*me;"PH#ت: *gh@BKHjLV9}pՐ^Ԍ?SJ~vPQP]65`x`40J6>)űb%j]aHPg%0=`ldԉ-=zbk zN)ޢ#F5[ 98V`%{Iya7c;WvuRא7)y. ;aě>C*^A|˄\Qq^G mRZSF+ *Sq`UJF+Mq/({2&nB˅~LmccJZUмX B+љ:plhNZAIpZf?d*m|Xus\gO@c*竾#TRb=d;>}ӌj5:| !mœEuizאzuʑ(?%u Iy?`9)aTUB3+uZgBV)ҒV}Sk 0Ix] %T ARjjj_6_k~ Z>+Йq0lR!NH /yzd4F1?{׺ܶ%ΏI׭gI5ٵOzK(P&JGRHݹv5"o贕)+fsVR&.K E(Wq4"c8ql чnUDŽ @7QQ308SjAK*=mvc1Sfti#(f(@ۺv>m~ׯ|wkkj)TlΣ\!>h {tTzW,-D}70k@F28}44Єwm:eR,@%?[64ͤ9GGzjT{h?[}d,h(DՄNI: [vT9L"`vOZ!,іA҈jMevZI~yo-3jbz['i Ňb17j(g`M9e7ITQTՈV5B5 \ram9 lnyWD$hXL6&ra_[.,G>k$XaMw{>QHID+$&9yfG%L٠A+[$^c/I#Vf!8[ZV)O}owwBClڟ5ᷭCc_'?Bt=u𜪏p'$*[up/zNH4Pƌ<&D's(CxMզQ1K\${-Ҩ Qlڃ?zKFzhWQ4Z AcpcBI6c)V;=Dg6`# 9uiܒdiȷ.hM@4luuYϙC|{;{<78ynqqlֵ[Cc!pbT*xQ)N8dv7hpZ: 9r;ښF4ʘ$Y=@mY=X؛5}NӅ52(}(_ߣ0 TP@]AgotT,F\\pa}4#(Il8饵bo>+I4}Ιc*j; 4Fgo4NX6b҂SW̥揢1Q"MЋ?Le3ڜ]$YRYvaUB:KW@0J&FLG؝YB_ =^ \`{Jk[ͺG]q}9ׁnrqq͞SUy}ɋ)zC : -"!ZvCȻEiEU%%zYAS3>n*_t$)䓫Z ֽEEUǯ !O9%/3PQ\p]V"Hx2g\=V(%H[D(i60-ByhzyBZ8$Hۀ-Bo&~j^Mϐ׭Oެn og<4fy5YlF4Tqk+fdI5Ԁnp1 p0{)/y1^J!LꍝK>]oyj7p/lڴy)RS|_zϞwŧo=7/2Y ?`˸X y( `zJdLЧs~.ٺjss-Dsd⽝TW6+}5AeΧP";/2*K0څEx $ 7x<>R^.1kp]|" lojcS> gY|4,ع䅄lBPL)͎Z)w9Kx85 axdtuA frjxNr&=g;gb!BdxPǛTjEC(7̣:gNZOCHNL9Cf6L8j `OZȨaOL O+2.ܒ)Θnu됷Or[ ݵVL^UO6dGC{hPTxM79.%\:m{{ɠ2m \be6XHh#5n)XRz}D\'4i3&ns|!ƅ\g1 s$JGMqHw/FBd?IlTzMƅ[b2++$f>9V݇o$AtFt~C>ٟ6syOCW\DcHr!?;4$|U +jo p!axF[PcQf>qjˁaKtsh4Lt97'WH<_qJA.q@nݦ>l4p-u .{h5dJ|LA@^LmE*)Qfܼ-g_}vvYGt M>F| QTWS$bs6 )IiSð'{\} m1O1 'n۵Ube_/#^(8.5dT]mi7~S4dij Ai(=ϡKѱ3{'op8Zӂq0t@¤E(k3Zc: G=kLl O&Fz2e z?{Z8ሰ*rΐ$34َQ^sxD1A(E)4Bq`(t(b6 YG84c~0CaX&ؒe;R]5GK:%"ADYc ʹ3VsEI>~<2Ppt_% ^L^".BөA+.ܑ<3A;~{{o4)(o6obpy,&;3dk1jjzQLzV57&1pܑ tʀVm@x}[B}pDi4{ /y9~թŅ znpODL/+6sM/lv#%ݶp]B8ʄj?q6 7ݻ}(*ոoА ZÇ?Psc?v$R h_aAcpc ؝]l*ASʍ͊Jk$&"m1"Qgwb!U%y,*Z?Giv]9k1yri-|Ng:rŞwҊe=gnI'_wBt c OG̙М"Hf+5M8>aLnJo!S0єŒTК)JNj'55tj/3U#`9dfjLE|@ҍ#M;#{<&GgT\GG2x͗mܼ>'2Ft9MV|TZɗ5&ԥ{crhf'T#gY[sR!XE[Sgc ~JSezըE)ra$x`{0+86pX] ,T\ji>kmIfS%Z=pUm?T /2j]bZk# aVrcoR{6U&tTg̜HFRLX~o-9w33ZT8H0-D[mO^[!)T[=(w?I@VlOC7{W7JjW ׭pW[WH%q2ضrnu'N[S;IQ$D>i:v?m& s )4I ^5 KaIBqbP/m>'4 vұы.y&U\q;1I\PC猻ώ.F{_/(|$%Ƣ>YCn9E5^1Cx:@6tٓҥNK юboSbP۶֎9˜ Q&I 6wOSB&7wd䬉&й&GV <)wlSwU7t\jF\'LsD.OZNjݚ|PV:?tIM.7;!~:d aʰ9ՉWη5(YĶOaCUORO<~kw7wz?U_/˜J]>#m]?_]ݼ~.泯VznܶsGzgYyN5+qbn#ʠ Y1Z{n<0^jV^^0e``Z_CQIvlw a#;~UGFdGAnwsGt<0=u yv\>|'E1iD7[v{E%%std.h%%fI و ̒oެ<4?sx8, qL!NؠW #QD@sw6=,mb"*#%wMkA<<2zu3):|F4ǻwtƜxX*;2ޭMW?SӓtuEo"?Ey5F/|}#پk$nȐkJ !nҠ&P4RJB2r-̥(}'ڸ:QAM-SNlh82C~7vSkgi MoR)LqȁCWw.aĢ\'OKi RT%|f#cqIUQ4BD!d6ηؘЗ&4Ү^uZ 爮.S@EM5B uVc5%gRS~d]',RX5ov]r!($L.ygR6|[>bDCon=mu ̀?n<}ql37&djCכaIS[s@%fL*X{.΍MS}Q4.PXDKd?alut`J,g' _ۦs FԱ op$.4JQ))R6 |ۖPZUY%DmC[@hAd$mg#{~,|Yqgv3k}Ԯcߛve5&0vH楠wf]gI ia.9|X߶m߻Pcй:*jl,ew|g|.=}Д[>Qhpplqfǀ =c 1ϧ1r|h24m6+;r͙ Dq0t#(Ռw,o]]+M![:܍*Rj/n ! D\r3٫YБy3Ɵ@ k} /qy=y4Xr%ҙF2ܳ( HoӸBhtaP%<\x1'$$~*َ8ŬGM]CmVFD)c(S$k$0#Dp62b&Ad1 {o΄ybRh$s) ٩c01^xs"6'GɆ0P)c !f#/}@+G] }s -W~Fľ0Up4c/f;ҖAxtrm ܇}yQCFW>p w df[#tmc,{! h;S ?{#2;{x 2Hg'۝9!状5!>"DL(ȑڭ,BKR8f߄]DE&㭎y5 itja X bUͼfeͦ2b؋D.` IkN؛3aaB1f|a\ r䙕xU\~_@ ]#Pk@pG9s&yw:nl $yzP'Rg8wGrows$,Êz"4EC!Z}jI"7h |Q7XVTk>\) : 4˷np{nŨ0h]Uõ^#dyZa0bJaѶx&dg\ܑ_c U `´H¤Z_vdTq1RX[<Ēm%ԬI"zYXZKMC\[xp55#B=M= O7ۏe2HR7_ns :r0y y $Hׂyf pxP=jGUթ&;UjbT*`4X>&UGA<Ӹ:*eP{ȅ0>jA^;Jp$#G 0h*t[\}guyIYm]S=}3ӧgYE蜶i[9Lsb&ZJer/D`?(Dxpiaτi Q=Qǰ0ͅi7xNg9fMSt\/7T/:^Bt|3Y=B 0M?FUޙ}bTK}^F@C4|F`Z r䙬*fbfpFuDfug,~VԆ]~ޚM#%W#hdc>ͧF< r3W^'047Umvas6R܆/n%oOAxOPvr,-wc#SUGAnץz %O.Űv(e=-V/hɻ{Z\LO @V@s$x=rOG .Kp_a; Ց|<Q6,u01ঙUx1}_[)L^8zܰyuܣΊ0avr fM*NorYQ).nAvc < CHr{NI xJe*2!>2 dAF ߧμ[50Ej6./+0{cjZAOן]}l(\ϛmFEЗ6_o!m;Mڊ5OeWlIԲ[jI,&IldUb++ĪEt]PZp \3 ِaz>LJ-!cIr^H)xLNH}H7δ&eSҭ!]n>ztn>밃+Y}'dCYC4l6m@ :;6i+7S&f2kWZo,TL.5}R*SJeoFG2>S!eM&v&)>3U^ "Av#[Ps€Sb{Tv 7,:A#vxd;b`IAS-A'TҌW"{Yh$hX1Jl]m՜ ("NЮNi͍N s#csRЎ-5:O[%PKn3ZrJK 3ZQv jNhi ӺH'oV}LUÎYvN ]%(eT*n;=h>vFa4AB;P I foVZ ]v ы6c$t^nӁlvO>T4AvJ/9T|ٲïT@v 7Kwu"N %.//y*C*s|׹-dy=' hA@rcrpc_[RG$(DJ r ]?TTX1kA|^>>o4x`WkYYS|e`c3rjI&3<ˁgR}Pþ'Ff!iuq+j2ړ u|v|O,~|L!@g*"NEėsMVQp±&m#T®vJ1h{fv2n {VRB KMJH/8L[)r곴M`Zl)%b%uh}P͓eoGڭׅ1lHiSfyGmOL빠%Sq)GݏP;f}?oF]śQWfUuU5{A> f52W *3)2rjD*!:y_^_LC'J;i;P6wB(VEǗnr̒9GEiV-qsƘ<*e#ȄǦ 7 D]7o݄ˮw|[~r[?Ǭ72ݗјu?n>Y-_hi4>REPsWw}77ޗJ;a4f"0v[7bQIipoNu~":3Jɖ$a2W|JYe$]iBM.vx?k.lH;! iKN3:?_Ef(t/_oJtFo<SЅq2R ,C?W_ˎ^l4#?M ֛ x{n7:VN\.- }QL<Y8A*zbxTw{BPS9ޢG^2WXG[]4k_p+kF (r0KlWY.bʽRyOP@'BE%Pt,Z&X <TqӜgAsΌ+Xf0@@9U`xPs:0xS:^^d&Hl$<>9ߎҪE$ BUϿ+hnoTFh(P4>Lh'Ebf+8ۛȮge m+ԙ%a5yjBPG44ۉ Ů4_6TGiVv-H@m=e+FsP|2R b4c j? 5i(EÁAzeY4 j[i (7Z#rj!qRdž$G,sf~\vGM0fyZ-wPXc"lkF{;n kD *%JhshuV=EHiš1cb8yi5ѾQPiK 8(|K04M )^YX]bQtV.{'Xv#tZ\=W_A@* cz@EXՕݜ00z<;%mM\̉(#j?|hy h PJ@ 4TygJIZWG$tHx8%Qʫ+eЦ8%hx ?!ZVO: Šo;|n8whhr7Kswe;1S9erETݩla ٰv9d 7 97nLƨ"[/gn[j;mwͿ:elÃ0Uh?ö1ٝoLϡq FQM "ġiGc)+ [Ѹ".>΄r{'2'BRY4BM_טԤ(ӡD}<e2H eD BD(S2V)rNt!0.,'P0I@^JsVk*ɛk,{X"{@XEɎYMV5q~Ҝ1E)i ˸L1՚Љu9cbTTgsxWhϵ^w"~G} սjf|r|Î(k]i_ \Kbp؍-\\őt5+9TtC RD[?|v^y7i4dPgeݰ*Q۝(Qc_t=m+(l LRa.ʧ__kn[,ԀkcbWt}ӫ pi8|`>ʽw|f%sj @6lqa,pyA.;2bCo)c|)^P>v7eb&mӓHsT*BVEuR{6_k򯘶HKE9(tZK?{W֑Aj_ aฃ N2/ښP"Rc }Nq.7.^.ŖĻNթW=kg@ Xcy$%؂r+m%ػQ~t}w)f6=G_s4ZZ/f>mk ~1{WW3k Dn{L^ԌTmZVKXc|qNpAhCH0L0Ggafg0u̕ɍmZ(C ɵ!Di^ @dxMܚ?I0aQJA^-`XdR-E`a3(@zD S;9&"F G;]\sԢԫl(x*&/x|v˿g_™[cS{?a1_&Oc~_B}~≖޶-D?ch5hzٜK€XRA"YE0p:q1J+ giJ ç7 wu#e,>&DF1 eLR<@bm8}%4W_,YvSwx::Faw]|nyvx7)>zFx[Kaxu7̵Ců{)_~b=7s~Un[ JJa,zJr#J&'uAif*oJGӽjfrLRAj39wls Ef0paњjFn3фw*}QQҗ>5OۨT'@״kR6 2kE(A"A_ o}3HA}S%^ P4`-8%`GGE$SE j,gh,)}Q,)} Ye}D@xV#E)hV"l܃s| ]>|WIkI_/ iDڲL9`jVXV{28꠮h,r=kXj} P@Fؠ6eoø[|!nb ("ލrSnCge`y4F0H^A:nnP:$`r+"4pA258U F1VJ_T|i-i]w=ӏCR[ŢVypM m2A w`rLl=B$1:Xnu7ri;NY'ݬ<ػ,(*C.<)ͭ[}oG%F} TRoW /&bi]MS~LE%`C?AfbJȫd-[hTܬYs}|:cA 洦fp71a.?O NȚc0ƳfLRo/nn?NO{B#y2 w/N~qtGw ՜Z8᳚܈3Bp~ @ %1Gs,KL/8nG-+Qhډ 9Pބ#ZqeRX) Hi(ZDГ63NꉸOqZƩph؝!R[-e)m5rT4zEPW9["[PSWJIUOR}RUOZm Rc(AeV:2uq jZmDO5lfk*7+K[^Y=sͿJ0T$*DaEY Bdje5z#UH+Ep@/6`YY=>( Z$5{A *А+@ĻvV#~uALXf4XpMfpEð%QI) <1EF4zZy5E0JFy;{cV`Jn $0GGlK0 lPF5'puy5Y0Z7agKqc03nK ;(pzrh#5j0nx MCeh*E)QLD v>02upՂz9%W tֿG`:b,65I9}u}`%ro۸S<)ӊliDqܽbǽ'/Jm@0h eRSQ#xr_F8ϻz3[/IUKSlGM /osq#Ј#)LLO: Q> t\(ٖK-CY,AZFi ˨ޖq<㔕G+(DB"9>jV+}QQ[җ>?I=<\Nq9sQ<~ ty+%O~3))nCYw㏱o@,/W)( @˥A]6{cK87Qw fSlQs /br,v)we厼ܱ:J;NڦxuF5x# 0a&S T))1w20coE2rhMLapc"NY,gGŜcGZ D8!Ԟ¨gՀ+ZOz NT#ֽ4ԩVX/r/\|݄J{@&7Ң%}⥸v{K#P']ſ\zGXbEh8t2Jdbu:N "DC@YFp°"BL'vT5jRs̿ !\?{>XO9T (rX-MĴ*A ]3fE"p \@FAi0-h f"/t>JDSȄmg-RD2@Z,d-Odye6 "DRЍ P(؝{J(E[aF)!uTI+e C.2 L)n5.͍™a.tIŢ-8}Fs&bN$)`q^&\yqkB;%z1^(޽0 ?"05jsS풁/2EQ~y@j^s~(BЃV ⠞ NU:W]v+S^v&N5SD'0<2Y!=&kLv"`(|OaƝ!^F]n3JS+@QHbޥ FBjVNH_ôr/`F!VQL\DSEWÈLXPZ-i`GEP['QP)cI3,FlLOTxs Dvq=|^^Y>/x%1B25S!1-)ͲrV(>,kiم$etuo$UkO_(9w̕\)Y6ۭAXKo1θ9i3l.?Gշ/ t~1{WW3+ xƤkI'V )=`*s1@a;jUߥ`ԅq  Uvn?,S"?Uy:vNaGKGTv"B,ɶ^ry BzҌ2QsX2q<fپ.dh-S7Tydd'\60|Pr&ZG?Əc"F۝8hgW">tK;^r1k@3Id{tHx ~!P;Ò AX^ P. c.gKxw$"uDū̯2ǧ]<{H˹H˻|ݑbx< <|W|j/w{՗!$dAG 9lҴk?cNNͧ~bIH"Ꮖ f5E+M cXo~}\:&:-0aUI+j Uvciе ݟuP6-S&t-DPN:.ɒ$z[>Yuw5B{oDƑPbs[Lul Tݤ`Ԙq+gh{C^L']EEak$U x TTǠmjEnyZnW0pG3_4SZr%zC硧[YӅ=< $nX`[BDѨ1bkⱖ&HT"!=̥NQ]AL(G ?NiJʻ> crs`/˜%&2l K H t)El^dD4Y4A1])MS<,`fN\U v PUFs<Ҙ;#ȑhiߙ4r&w8f^|#?f/LWi -NG' cw{ G>:o_"B !sozn<3ju>Wӯ=C3c{fhov G>TNtoeDH+IYi%tlIrn 'EJ@reBIe;:̛0\ϣJj؆x\V|m[##c@Mf7AbJ$؎ɣAJ~Ȳ<,I'K)w<9IA,?ʐ;>MFTi}}3,,dB'Gq ZDZTmjɥ2=89&/EG<6ȵPaf8h3̵ ȼ1Kj8j!4SL!i°I@%΀!·QZAH -D$-x>C-G^9Գ$s,<AF} 81 &Llxm}O߮^-UZ yE}I״eXv@P)z*G#/A1Njm~h #AO+[Uq˨<^@Nfq9Iʩ8_zRN5Yy))0MΧ7 D{ko{LyFHK})chMB]ICKіUn=6L#Bs?Fv7sxm.]2u||2ZZ-TYuhHRa}!ŮJm៘mXAD*W&p\?:J_gX%Z2\-5‡D~0̅|c2ƓOy-ʛ!gPnzX*VyGoEΔOuO#>޳>VmXn*̇׮$ {DF`@ Gyi軕e0Щr>g9}Gz̲ЉZ5eoΞL/g?e~E⬷??>_ts0m/~=׫~ ߽9}<} q8OП7+X_Ξ]_f /]qTc6X~z .>?^xbm Kŗ*W9VË}>4jg0XSכ4v:ܴWo~&/ΆݪFTjAҏ-Z?ޑo'^ͧ7zo`-gՅ\@S]r1?_\jgK_Q@:[|. t04xr|҇ ~ ʍ?I2|a~s#wǓg(^LA_:[IǬ 8=}̋3sj, ~ reja\x_pW$rWR`8Vb2y2O:,X8o@J"|5e\]/gG^pqaezom9f6kdIq+r\>HYX]/jӅ[L op"`h+f:Y x=a9qy',bϽpvCz֛N&p"{~2cT iT$iT*p4+syd}oRjZ 'GبoU˃Q`?qFtob_F=PRSi4Q4MYȄ}0Rb`Y9-vٟ4K)`IbC?rމ% PUxݱ%yH,K,I{$,Kr%gxZ2FD[4IY8\[dXauN)JFKIvY>^-tla~ZȢ]º8i/^>3VCRVI ܤWn \..BEW xEt8NpqLS 6T/„Ԏ $v)˄b$)ƩFS2(j7҈ھ(t1Un y KsQ/peG)T'[-&v802J >!ƣ!xFk 't4SQ$ 3Xz|XxZJP"E (e 'Fbū ܛp>`j.}F'V1PzH=M>JԹL{f)2+qfc5PN:VeiF*`$L!Go-HY&f#Z)Z`e[bɨŜ *62zC #fEqڅ>LrݺMn(ۦ3M#({iWtl7PRhW67٨  !ս]6٨kǥ]nmtFsq w5lEJั Pd spW"(8n*)xvÛ ML-V ]*N#|>o&t!I*c X*My(OFmNT*H:^A1XoA q0 bjLjTfБ9S;rSM p"0IF-/Cs9 O5qGɖ`(Sz|QR wCPЗ\39}ـ &U/`u [mJE Ilі>hE6v2WIK:0΂ !bʎws;xGPI]G*6uPdTaDh3+Jw1K\;~)YfXL-x枋)8uSGsɅVسa̖F&;4;a]F^%'` 8pgBʌr=lC}SAqkM=p0@_ϸ c qʙ /}&̊oT#y| MS |g3p SGN5r0_H,GHzGcSA·mO'k h%EB oQ*-Q ۭ$SV*w3ns?I\H As ʐzۿ}&&\8kEAiS uYSHcX[= S4Q_H%4[ۺ6Vր HS\zFT` 띬@R: B=*ѰG ݀##0 YO lokz(+l"9E,lz+lڄ?Q܆(Shg=Nvr[z X,qZa=*T"h=ZOP&ZO U˿J4guZ܆4-՟:Ѻ`K)x&4&rʍhyZ-ciJK2?R&}c]>}HpzOy{j LZaƣ\4EO}ޜ߱(vZAR-;&8߹(ު;){bY_> ңRu",mp)RC 9Sa:Q8 rP1S"IF(kg6(*Jk;)) ~@XlNlz RB3Q65m ՙRv>%\Ϳ ѯKW{DWQ~tyf >f>?xlB\N]2hQLᨇg ^{ t,U>80^0+#e}DU9WA«=dtA.vr h'1Qf ikslw i#fim<\2V)fh`, _ %E~U~ݚG@jgzbɥ ^zS ^4џdp~#Jju^] zvhzl|bwDMdtl?=n2mYԨZn#ZSJ҄aMl:㙠!KEֶu{HTh#i ےQrk•+WN0i^Aὃ}S/wl *ߜ80C70ouⲂ@eXiF=/*o|yr8ZS3Gg_lE3;vAk!gH#ؤ-<@hŅ>4[ܝ0q1 NY)%9ĎxmzvdC;DPntQцDD"cmzK^l9/#tu^UTa< b9Vt0/Ʊ(AځU%ރX]u{ˁW{byWgyW4Xޕ}hIyGN.h%C+yZ~@9ֹkOlYy/fjop[F`R=eQue 0UG2ТUO$OV^+$=ǤzUC>r +zGaWe3lwEG`RHԍoZJ*c*E}:R6᭎fDo}Eف6~qY&CX{kF:('Lj;hjFvk sm&NZLvEX >aER%2a/1~$\nH C=,w=a<KJ.wb~WҊ{|+ A\/C9Vo]YoI+XRyE0c`=O=0򴅖%DYHRdQbJn""#3Iȡjg7n!tyD{1a3vcMeeg凲䢇%?\AŇ O"Nj|iYӃؾJ; *jm{~np0F<;wK;PܗVqߑ!-:NPq2&( [lJ%gǂnef`q+{1e3%tP斻 ְ7:q'K/)a͢&'ӃؾO\!zbWyO\Jg-dgf7֎0Oʐ#SX I5Ry<}:̑f1yʕ0`sw,`Q@%EzhnѸj܌{\2^*6ϓJ3P#iHtd TGT[˸ s^9ٓ)T˷/ Qºi̬"W,0) Cb=x꒷;*hk%B[kʊ@ȝ^F(fKwc֝,ܒ0Eу%C6tYN9\g`mSQߢ肖Mqg`eQPK9:8Z1"w1nLмtR51 ]3*G?Jd#~LVbOǍw: R뙳AY8He!r9Ƅ L%#|/gMh{߅Fp;\d@|4bD*inmJ{%K÷S'ў$ GޅW^j a2"w#g)EJnOZ'&r;Ns, -Nطik 5B^m,/5|`ݟBkI.9: M7{ZsEV`u9@3#&Ewuz(tk lּm{'W&=^`EF5 9 ͽDR <bhFDӨAW>bx}=Ã,x sF;m9a7a՚i2pk^4Ǔگ.ׯP9 0 J[vwѓ[dMB- QZDEuguyd;Aiĩ?m-ک<ҳ%yOgnojI>x?/؇x2'KߪyۻJsjZy&}~yw?.9oi6M>|hY݄/-OXo ٤˼%?|]T+/I%oxF_}Nƃ$#wd8b:s M&3`V  BekpP@!`of ' Aуd\ V$"H*PvYOJ@4\aL%@h1&\8H 䪳=.+A͹WϭKtusidxP[;>_3]Ӎ*ӲwLjR3=Ij5X1ߔy7&r`|9adflz+zx.h bwZ'>^kWOb2z"v8OT?á_"X[El7H!h :% KG6ގ >vS_Ak;^> MI y&M]“R!)^F[4j+V-M֧v_SN 11dRD4Jw۪z椿?1'62gŇwwԜ?VF2R {y7gO,*/h]޳7N3wPF b^JlyI>lE?{d@mi+˞fSZ XA Y 6 I4l}b/+ȼ,$;\ TUT~Wӣ,l G~V2߷Et2w(A%`IHDP $D*^7@Y>!O}q=k7SO~//sl#[U{ f|^DDkRrpl$%naX{#%Jޑk~|hweQ=B6"E<8]QF;ry:(HL& f,:!o32RGg6>vBkd?j'7;{l!;/3\㶟Yn$,(yB~_,"ElEjЍP#Bo#~Y{kpß+՗eH$@h$Go) FS H4.{YWҨ=؅6XӘ?(}hazt~b8aO >!9jnƫ8P'Fα`USAUS42jIKFSO(/M ,=h ңMi ؓ9'M6dcQ$6.$p׷4Q ZPY$z:'͈쏚H5]Ԍ}ѴjjF.;^l.y@?!cJSm!(x !  Ame{b{ŭe DzGK\.HzMp+gf3VIۦ?=Vq{\,=Ysv`Pjn^_^b pT N@& QB;st+&0&)U&Cd(.e?p3B,:K+B ITKm`܀B@ $&IJ}*^"6W FkX&LD[ JOpc<S!@GpsUZ,7.gP8MA1jKD ( = Uqw"WF%klM "Z^fN]MCئᖶ֠fiJÆV)VgJZүW,^&QBT 0a: Ĥ4fT 5xjZR0*Pw1ԯC(ՋW9!çN '&v'8q!`BTq+$ Hq=ڟo"Óix\vdz%x=bϓo60G#\8-1W?h?Yh4h?esC`$9؈G}XtJQ6z/$_q">R>ksFzTcШ99'ҘD[4vvyGZ|ЉjAVIi`Xz~E@zGkx竵W{P+_%B(.(T"ˡySo{g0%'m/"`qM7?'7祉ݦocs IGz B,o[snK½|TfB6rt9 TPP;ũhi. rKXTcS?śyjg4MWp4.*EE={0n]¿<{( kow$瓾?[C1RMu~l-@HsiM{ 1Cf?_AB_h~+i;5q=y\ ꄁf[!ǠL AXBpi\6TG`U]l1]ЫvUAVB;PUobvMhYք|"ZKŶvݰbPFtQEp= x*jZV5!!_֒)E:ƹ݊\wAEmTZw;ݺqݭ.\Dkä{![M( ϴ.eDnU)Fhڭ6e[LTN$oyDoGYNl.;7Fi.YhB\0~7v$S9PС5ѫ;&堉nAR0|RɾۈntP@#Z0nW)Wj^oÄƫmBddY5MBB,<+lػx9#P8PaPᓞPIJ:ʩ D+ˌy9ĉK8%FL T a64r#!1ESڍ?\pPF+HAڿ˂\| >?ߘO٧$0OS&iIZ4)=?|>+.[_oFYtalzT?;/|=A 8)"z6&j]g! $6ې8q21LûK?Dt #.3t\~wW!:Cƴ-0ZNz=hڑ20Jߩ;3l_N5gbWuDKB=9B@<9ӚyVSB)!JIY(/9PH]>w$PUJQmU%bg]kMר2çlf f,z@ZU[v28-r*,P4?dxѼA P T)-wR@qZ80 n?6: @Eˬ( d.fhz9[G5bou u6` /v-*_FQfoW!t23zvA f%d`#`f`z`21?)aoյ4GTTMJM0ϯ_>ŋv@sv?һ`?5x5ʗs}-?*kgF|_I}W.Ϋ['\,OU/Fލaoބ>!ތ" 4j\>enX̖>}!zDLs%6뻾@۶@^s vgctMX"s]leҩsA|<"Z4s6V)tIcfpa3X8' M M!ј$Cp5 [X;t7xTڸ1C_۵z}v#TMq蠚jdL7ih|e <B [mc2x\؊0ZjNIY)5\2RjKʗZ`JG  PPZa cɨU MH,){x(jĄ`j@^3I)I6wmnYF6wWmnAiaW`x2{c'77qyȪ @/e(&|:ͬV+Xr^$LÐMǯ/?2ix {; ν2#w'xO;v|P'Q=hu? ucyT i ۞E[`gS]'W`/+}茽7'asЅ ~;on5IÐN󠰁|+n6 z,-rMuc`1 P8x꺴h mnxPJ2wUݖu}kW Ou*t$At36{˓Ѐ(\iÃ#|`2':D\:0]w?&㵺2s:TK/AEQm* !9[7m,LZ_[ufCG4ƝjLJH:YJcRX ʋPͽXP-MMnhu,G);ylj EGϢ'BBsg;Y\4,B+~AH}ڨ}<* s uzz-՘U7iLͤYѲ0,0N\;E>f/.ԛ6<+njBBr}LImFúJږyr@e8"}l-ص>abhVJ8aqPHˁ TmN*8RO*ZJnt,VҼ :&~va)\|g4"ABZ@\Hȅ(*KQ)>(lp,Ĩ|8.= տ$o>Nf#f:.G9dnxXѵc ^m ̿v>`6.\DećËҥoLwT>vot _02 o,6ѠiHzȡWѸZ ABT=x^CVH֏ꨠ"MT ;A9Aяp R *}NP% !rC?|4.,Js".R>S7|: p!.V!h81$qLSaƌzĐ6| =rb4G5^9bV(ܡH4{apX l,|sqF&rq@|~O./nUxz^7wSH/.+ISEgRey;oFw quū{Q%^*6] RdIRd?ĄI[Kav) 95DE4G™FPגQ!:[+ y]֭U$=<Fa8|4\\d. .y>åh~8K˩K&Tgds岶3iIQbT6 mLbK7qR%qR 5Q=X"E$EOMǢGHjuue%r~AqHPJ +ix1&!9d%g&*&!2?<>4> !)WdvbRM{?۞<$LmcIaNp™_ !}$Fes`P$5 ɹy}I](H62 CrȯU3"ېq춗= օwũ_²yUS̮t>77 ͻƔHFVX9V5pwؔ p'Ee0î D?SkFnF}X'pu`pw6f޾8がSnV%.eU"Y>\zo;!q/jNBbM|A$6&fAsڪզ[m Nb/+yYx5B^M|psΑB |A1.u8qN_d \WBϟSD0Be CP#?@q21gG:ٶDY` ث'uJvם% jKKV hnCmC354 beίw;gif7߾H3J.xxTBonNHH1C|cvG,1\M媖{RXc7PnwêSDQģ~2mftEH0~itT쫃W=d2$72@A!`a@6J-Rc+-+.%8=Dh "Yvg^G&ͩH#wxmr,xM_ tB)ȜN BC`ڂC)M-S.#Hݪ B@gju4G,xWvFI C`YɊP|4͢5yD0đ:*txwhX愳kD) ԳZ=@YD Rs17SŜ1eckRs0av#%AZsKh8W,'Q KL%ZJGO4K#lO);bE3au4zYUk2-@A"cB u|]z75'*.^gVԙ@` \t[_rN0s$LoFmY YM?T*v3OCv]Z{-/^Ci (nBF aL؇BsFwl^]ف")4el4qVpdSdBy}fH]eq 1كD_ގҰ׈JilAz~aƥ (F$wM8G@<2iL.Λxqj_lN 1}J~|RD2J4{ΕfR9!\!R^[ ;xھho݉bd=)U]A5'I)ȹҚJH%>GqL -?|IUfO~L\hP = w,qN,[vZmWAUAi>vq \hIod"~LidD@Tܛ)-*JP4'l:Hri- -4>H+۹*JjczĻ%U9XvJql6<9Ǻh/R2!*OT.ȃ$)ʹ[ -%tiC`w!@qd|>OegD*В'=9mf4 o֋־Fy;j3o}AӷW64fggo`T79WX1'ǻ[$yLEOߟ4[ȀIcD{ZczH(/"~H0W_՟Aۡʅim>Oi'Y1>ɬm?l8 ,!z =8hD=)cc5/g B|=z<12Y%F`#}GaAE\~jݯ }COHxLSG D+A1`'rxei:O?oAE+~X?Qfjӗj1Jj,bJ~sMT 8 f0>qR`'دQPY QF0 GFF5e(J@͢|[5 AͶ 0Z#:I%ˣ xcr?dvl4]7z#N(Hqpn'TC-CcfJ#?4k3ˮ^:cTaƾ3*1paLb۵}'Ìv1@}6h (LZJCoG?)~g@`~V_ ft#9Al9f_')jL9 8J2\!A&N>lA@WK5P`9!ʪ  Vh\ihυIl΂yۏ k6.0kO‡nV'Om[lHqjW'%hO#k0!"$hc#7ş$~+ٗ3_3[FV Leֿ<'YRsFO~|+WXN0S!åfGσ7NmՉN@Lr3cu efE/ā V3fQ|:J>7 п4[sKr59[kD߁RDY֫'.¹>ԫe(.Z ^-CyϽ(kR.mܨ)qݗJI1'8U-~?KdPf$u =%7k*.!>x5.LaH%/!pRu|r%1.i ᵛK/!%J8wGJ5>F`YQڊ0v\?!(1u'3.ką-FrY+`,X\2MኛimՉ3 ϋ4.F惭/Q!Tn1 _2zgc>./μaϫKfg\Sw<3rGEyԦUK[!1Q!nu>ʶ ~͉m?nO4uCaD]2-e8GL3[9Z9=j2Z ϏùqΩ61- ̔,fB}yD00XS?rc`_ r~qoVP>#=OmAR ˞^\(O ~o~ D3y@SUcSUtiXK^ל$f!/״Kɳr5#ۦ=Vfmҝe!*[] >7I7*z[]W>f==je?9[F4~e߭2-o;m[7WA@"G hDE@"'$RЭ&x4HDW26!cr caDBK&o,<)&i$B"bk(jb]ۀjz Է%߂ 8-zԕ@J1k+Wh^0` %REw!hW1_oF敒pI+NVáM:Rn>w ՈvrYh{&d/~>k]19xkgZ9evilTZw^ ZLFn ^Q>}0t}^7(dZmor_F |&ndMK#W:Dd1l#(dk}} F@sj)NJ)I(DE|bT nf$QJT)],};> gV5CۡʽVXVªȟF+/ſ|pƭMon&v 4y݉Abuҵ۽t^2b8U̸#8oq- \nԧߧ.%;UmM<3ON,߾\3Tԅ 2v'yÑhF˜b %o)R#`P)G5Ҥg% Z w'z 06NOٮ,uj q&q`|enAY^(R `QJ]C|4{kJB n؍7?46Ν /*y2`qC5†,4PgIgHK*ʍ IS1mDeܶa{,^|&dD _Ras$Z,Mj},bI2$Ʋ|6ɰnSUd}ڞS%t%!˗N&@MMGp(,ʄ/M,\!g~T\]ŏ9q@wwgGյJrkz@*K2:{+ˬ$uN |*P;r~8ں!?=MQruV^4>]>煊s vmig1CN7w *Ͼ7t?'"v&Z}1 ;k^_yi^9Jf235}h͢ӛceG\FR+oM)oGD's;~2/c(E~UCިCȿtF&b#;llgӔA.9O~/б)9ߓ ń75M{*JNɘ9:+A:z7F6e!鍸4UĶTg"zgUlyM7L\r T٧z;15z M-OaEmKn&eְJ6N wg~/seav1ޙ's>Ҋ}՘oLQ̮y[tYKN02^?.M>M~*d8>L_d4Zx7.2,eM$G`<+FO^3)k*e:Su-`uv3HG]j#T٦u&91Fgy7{N"*&_L2H)1$i[ NHj N`x`„ȔTXa) 20K@!F$p* v>pa_Byi7稤oVlB1,4@I0Hko%0D< JEƦHGYcc:Ёs%FX#bGJlH?CH u,d:Y,ؚiki$z-jS5\PP2)* ;*VR^-UpEV4FBkɘ4Mi)d: c(@\b _ H`H@$U9b(gੌ'v V\pu"VF*g'"D1A&T\4EG2VY\KSw$;ދcmrpd DŽđ+ROMEzѳDIL B|TL&S ,zc[ɾ7M }ݰBG1X莵|uQ3Θڢթ_{robCъc J|[Zy1{bs#ogln ܕ+fh!LdžV[V Uu8ΗTYӽCX>r7evCDq9r-) L^2.[Ɇ /3}]AB0Ae0M`Chv?׬ lQt8_$Y:Й9T'3sf:;1$+=bl9Mƨ| 磔!)2 S#(*$H+@r1i"Q(pe0eִPe]4 C :5BB!֕6c4pqm3 XF7.eJ:Π 1ئ(Ѣ{^ZS`TW6M_-ܳ=~|֌-YmcΊǯ3+Qbᒈ1e#zIjabscf9et4A _qFb C4?tv%iSp{LOCSi^X" eˌ?\^LQJ%&nU:雷x 9!i~^k |U* r 咺~PXgf` N1HH DscG\z~ 4{PJ`vBP"c"/,b))fMF",Jg]LxvAх,lWM2ɖi~]Ω $"&u4> Hgʅr^ߢoKDdLF7&g'5՞eftHO`Y3{za\5?(-C_?^Ƒ p\G<1L{=)5e!&)QHT)Kb ŨH2m q¢\Fa4~3^o 4B 6L BK#S~S˻νK)4%8HN$Sxӄ3YJQi)zEW,a1Hř!^GUj!?21B*p &Dd"gc$P0|R1ֻsNqK~nlmmg፟~o=Y,k{Z1+װ01UhgWi2ͷ5a]ցMP+ M7#!>JJ}V6NE*@ddЛhlҾt4tN  9q+T ך;lT4GN,Fa9P~;:Ì0;<6}ӥ8Gn8񁔰V% xL?Փ͆9ž:mD~^KsR{8j;{ gPNv\9r:Z`1el}~"9!:- v 慰^oGŚsf>߫Д~\6a k\ao/ο|3sۅx L]sۼQ*^@q8璿R(#˔k2 ;^X?,)%U#?] Fdvko7r2w L XmDK-ѽb|2pmkc,\ | ' )c$T,e* ĩLMBϞX (Q8s;׻4KZ!) \zlڊ@N% [u2.q@:)`wR=jdaQ~ov" IM"r Wpȣ4% q @#wTFlE c^u,83l'⬛1##R $@BZ<ARK JBLTIZIg'!2<9a ,"1f~oaGgH=!iXjD a[d2fe[,4ԳF,=L!K|3ΆX~}Grdd^dY{1}mSm҂7sa ^̶ss7[͆9h<̴_-=w's5_lVGq -z{@)!Vo7srx>C)2&;M$'^혜W[p=H٧>R;0'z_}5>g.YM[o?5P;+a l-\\ral%  ɜ}ۄ2IPݟҀL6Er$ ;<>7ٳu p$NfR,'x aB-H <6A#h4"ByErw@ 5G {o/=b<Xo"(mN]FFbz Cf(qwGnJ:ԵT%G$2VǸQН6QBǨ:8ƣ.68L@0IP jV%\q%-qpF+f\[J@.-o}͏WuQ߾{Ug`oI^BK+U{% 4>8vEYׁw{./_gA.= >?+(>>][xtAcC)7N^6;}FZPU`q8^'g04ߞ 9y 4u/՗X} ׷L(3 Cq/3D8vYLտD4E|NQ2gN}2!I?2!V91Dovr1'y:9VAӉIɑy@reY)B"퉂/V yZFZŚ=y"-^~Wo!FYS eTkшBKm5 Seu\Ygd7"IREU;tFN ʑo2,`=*YI+!KT &6=b-!Ja(0Hy "2 }.$DG_x|E<&@ IcXUƜ8J`CX m@_!gl>5Gp,C,$ ku* @Dlh}C#,(:}DCL9@DBMpF96cD!" av϶3xCEGf 킦# KUN٠J*'m:ZaV0?$Ё>{oϘiS]/F .Q~A?h5f Οf1l5ٕD)WB 'lqz3Eߖȉ.}0̩U4^]/]veHA~k/aR0r8SD>!p%P>I+!ńD1&0LԄC`- {6 j8u/+ʯ᪁_oޭ!u9E;zĖNw{ %xa1ByO鱧=C))pОKBaϧ90.:qZGf~3j1qɾ >$%G]iG$Uap%KZṙ&ȿMh  1wmmKz Ύ=9NrkbBQ 9r$(\8VL5gKwuW9lCLE9`EMZ"V2">KMjuxE!bT ΫCRj"˙fXGHe` S`~ Hչi@&;=dMu Mx5u)&j^HU1ȥĊ 5n@j0>&A^E Bhq\2VٱH_-d2,AR_㮁P& Q9qՆkWvz)`oLgw0hw',_D^"_+ի`f/'~jf-&OGN3,tdV m,|dF6]KtwVMg/fmB)YCls г;|?wau7NsblBeB]2ZBΗ<ӦhğYB#jJo8;cZ^CckREez--HYk'v +$p9[yG>*t7m͚+6{"AHgԌng{f2x˫¥yK&j߷<ط2_31 mZ`INAwgzoêLӪz $qf}(o7 [#X F\+f-=kQJC(k\ew$!qm!SZP649eihђRNVF Ϗ7 !=#-߿0z4`feA}_Gm8ukpiY`??31"k5: *GHUPۋ}I5jLŠZ19b= '=(|reӁS+s,V#cBm*a1b%fNڦx=!m]Θ()7I\}yrؗ{9N qRQŪip=qVI&Oy @0Q- `!K B3voힾt_ k19d5{݌Qj\solԈQ=d WȀNUNxK:zXDjwC4u  >ȬK3zX"+Ma1. uA=aL%a(%_$at7(tU.?zBe7+ ^oFi˞ݧM R2ݝ4+{[&Y m+w4jZ9y}*_8]_@EQ[0 ؀ 7b %Flery*).!E@UŮH1Lsy$wD8/WOʚa񎙶/{ 4t};S*{ْ{=6`BMnf"/o _(|rf'Li9|{1(u'ͳ-Egs,3$sTڜ! 5C^@|b3ߴc@Jy+?zԩ{<~B&u%[k],, bz;3=t\o5_JOt1U#ZWhө5/BtJùye^Vɿ.pDI N+` JG˔ ccg)h S6]obr{ss7B x(JPL(&bmٹ. 18g ,robX8癋HD.Q1lTz&%F!rɅVc XәmZ#&A.)\B;碅`0$&GQ+-)-G])aL EID2jpǓ!0Td<0'DddjU}Wz8JC:EΛ$"WNK߄Sޚ>8aA^A %dp7@Bzr?\yCPZP7MPK${f7%̵dcd*D(D$ma>TcKݼd|ݜ@Qy9ϨA7- r~s{y3j o S&Y`U&cYr5Y0qyp1>ukm; Y6Н1;4g7 3d{;F+*;9x0F}N=ww3/m?fpb;6m7sgϞ Y5xg-"4 (HcCSLy)kjbZ} G ' ϠLihdyϨI~ ~&o0Q RlgEGbUyoUm̯.â;AuYk{0>Mp܀5{`Wj\ ~0~g>$5XxaXp-il-mN7O~K)U=wtW-}L1]p6Fm>LhgL1%| h`f g$Tm Mϔ@ĬۘRmb֚Kb G=u Sg=${|{ݓK|5B &-Ê&zj `&Jld9"Ă/L=AkQ?{ȍ=I]]dl^0x-L?ŖlխK&vU_n$BFVA xJ <`w48C@+ i4nYD%i1ksͺugt)Ѭ1BxH*1`˒ 4  HS YlE땓-f-8+LJJ FFLt&Y ^&p)D)6" V[~:/-% K[@t0ɣ1yY.F1If|PJMl1\iK^Xx+=͙()1צ#j$Dg"Y>LB\9ku̵fMT`c0O+~Z$zSg_x:矠O=x1hҡ5T-x/);9ϯ4TR1ڇ]\'AD§ L4l$Pb;HvLB$ ]%Z _%;%Tq59ik5ARv+d$T@;U\ O驊"xK\AG8ƪkeB{ (w5֩ebjSRrR:,k;\']ZrVf߫(:Fy-".W czc+۴x>#V`@vP !;,*0K '; ^C !AWGݑwJ'ZSw/L=/lvV WRd\"[Cz;NnN'PʥU:+ r{C>8-w߿/EA^n]+3Qżm>+=#].Mc!6ᢲ?x~+?O%HziuG}6!h`Soq ous>u(ZvEڢd\-h*:lP(˥ҩ~yZ僅FKqΟX! 2-ЋʪO=x_?zk_ݡ0a6R oN+& %=nk"'-p;dd͚:!RBj 4ܯ! \uwM(&{2d/oVxICrg`6MpKB u\t8UDRr`䩼0p ӆEo 6v*z& Pp ot 9qPo䊈] }LB>Q}ruvvHtARL|*1!~p7]_gfr;>'ZonH%X9<C;,u+1Nc>G|7tA!kwH9뼡9TuW?K'?*. {1U"hlεZ40cJFw[67*Og~M.6^| pQE-zo&Q}X26TwnGӓTDQ2p uqJB\J)2%Pq  sAP뙊g&d%Sl3kv>iY5|dP$_p0w~0o}g<{esvHAr ~ 1&VC{ ?dj?%>DMX\m:|) 8q<ތhפs0ˋrZ( ʀmB|zC>jϧfD y.;[.ceE%$PABr@GѨ($1o%fv_G!6,Dus_l.Pu fXW~Yo$Ē29"9<: BAx$TD{ڊ\ Z/EQ{<Un?P<Ҡhg~"C%0eEz^˖uQj1ncl%Cu"Rv:ůr;Yawbm` ZO[n/WM -5iSJr oMMt6.Ia>{dzǿb-t k uKTbMڼ ڐ?hFo ㇹY&&qqx]zNw^7׌q,gԓhͶ&$-pho:Ȏ!'}t/wOA[gkzS{,]xbt3g0}_Glֲ^ ʃ;ne{҃[pW/IRA@ւiᡝ0sM0U*`]k~z(VO98)g.Ϙ|!N4܆bQ>_F>~Kя-ɗ5 (z&WhO^^vkS X0 _c)Gu4#!KX*bѕwZjaX"7Z{f%ʭ0UMg6KF$#/(I߄4]. 1RinҎ+i2A(TA't'P̾V#( zffe װÜjǙ4"1N"VH^8QV CfNx|P-x(&0j ]=D}ٝsS衂ZU\_Ldy5ѻ:ۆ\tTȒ.l] RTkX4 T쫚h_ڞC2( % y'Tjq8D^KBҤvE׵+@aVEhU3z.gKjfצ*{S  z9&$tmjDHa.e-l4~vDSUuE"nNwXJ=ߧm\ >ͳ>IkB崟xxoKܷ;<X>v%E/. O /njJwNp ` &?5]3 !WA aL4QnޭSn%ϯw[dj]H=Z@A_`QY *:g.HE{C=XUs ~miѨ.)5pq:;ua ";nxӺڂDXy o.ӄR ; b sM|)C;v!MS͵`e>| u~^qO6'/[Ӄ0p˰p7c*BK'GPL%W|F'v}wvC†.j' zP4j#O 1fO2sy96"_ =\v<2Y+`Dpɖ_gqʚ4mwx[N)sCEɢz[Nb3L/rOMIR)@kΟB{C| ):IYũXTO6:Iy JcGU,dTɓg:t6cT=%C;Ż 9r68 $*;HmF\V :M"'k4q55I^LhIRF"cW4x$ ƒCE jAp0Y;IAD`YN(rL Cx$ VHb'2-=*I<]b]7Jb>2>,rZ_Jq;e+&@_XظN% EeՔ)ӊq`CK]E8Йu'\Piy43dc|Qjcw5F,^O1B+]aN"wX ΧDL5,ds=PTRJ,bMC;Fq 5"KQ@]7O]x9q=1|=i]oo9O n,,ce[xQN[|dGW8}=Lǣ2t=h&?^ &nDM/Ыy34Ep~;kd)᜽.M Rqyb%svȀZɲ砙j.ՐY<.EDqARgGB$fƈWNփh#YaA: ))E,Em/1z ZA&ՇLXrb- Z&Lg%I)\ 6J ,=|Vg@_6A'L  6! ^.MgDOc{LD 3A$js޸Y9š✛ oމ3ƷW ]`札IMb\ f;ɓ\r FUI '%UCm.Ȃ\_ UbQ!%T%Uٻ6r$+ LL0a{66v:ѷy^n%"$e= KDUEZvKb<'Hx T&j*- x*{?Yy]#:B pVP!@v7sR=\Lzםֺ/8lXKݪ>[g~,֙愴Zg E]aہ1Fc"\u)ѧw uBa$2%ିy,Y'+ ! %ݚPh}—ҩRBnEj¾9Q޽a2vP|َoPRsqS:^ww,u]_woG񆍔9A"IOpU-΅2 $?-7iڏOcU6P?쯻٨^|ښ3J]K. iQE%+UNc7Ԩx\Ie`L`e|1e*Z՜8t7PLr/qΗ\Ӕ39?QCd+JKٿp`4D-~%M!G̓#ToX/CQ%(Ѹ}rPjvVhbf<`{e!0 TbgJ;C LP٣#5wêd~?MwƙIUK‚(D0 2e5;)wN٠7:z#KSѭo`ؘ*!V(TXSzx('f59+a Z&@S/gCɂX)\YXɶ-;ݱj%i$TaoSJ> uLhٖJMφJއ{V -|HlӁ1{<=Th9|D8x|,q4[Û|+->zGW.f0{ Z m+@- YicUūxH{σ B3Gx^qw?*YPܮl6rѝ^/fscђpպ3pYPR<_8c)P4`=`f烈& 0[5]p+*Cdm2xls2m|z4Y}P#7+c1{x20F83:~wwjT,j)Fw9jWϴ+Ci5mxZwJw9:7(^RXX> ?<_ƇN7MOAwM[_KޞG{pO#rFDbJev\ܥ?<"gS,³|.:Bǣ݇՝_gU#A~Q0s;P%> ;p{t%1yS8H>'ݨƳiKS͆D¢a9tCp"BxH/%E8$ e,p%ƫ[n Yj Fc\ tOlY9ikB j `t\bjnT/]w$v i9 /8x`:Yi .+x۷rh8o`lM[m%9gP!DL]tI o?#f>]OnaCߕz\E@U`!GҞZ+3FI`kL{_o 8ӻX{uxvi3wr6Lm<.Ĺx"j*Dqy׎~KNjCҬ;BP\8w_{-?\,I I ch*EM9C L:3L*Fcru^~O9vc$j>Ri#n<X MBRh&10 vH"du\.۳4Q<:DD^K9=o!B~BLww^9xZyTk2l5JfIĵTsK^KSӶ|~Q1=gm EVONnRX0|eB sUTHOu t+@Kg`%yp[ `z ÿ &n&j;M4rl?eK#BZ&C|;56Czh- WjU:t1>=rB8mU,.#9H\ M=.2Tl=8NNB- B@ta_\V<< R lcZNw~ |3zO;.l:]8F#ïb|-M~ۡ>Թw_r\Nӹ]{O:-i`: f:J\zm!_tj]+Gmf;<}T cLdt]*Ewjަ j]W[&vYӥ\H!˼pBʂi! sO.=EMH+,)mVbu#Pc?Ng3NBR[^ .Z\Ԣ k)m%vhĖ3)8%B5k;ϔ,T0,!d(%/jVCQm &LKg#yOYtAb}:ͳ:m5!pᯣ'kY%{1wԼ.UJ=Ah DZ3m}gϖ.Rrt{ :q./Sz~,oqѹbaԛ_ ,d12W}(xn :2n JulU9tQT=B:Z$2 9N O:b4XNCJQ|UwU!,@̗BhDFQ)*TamE!&JT0xdb#,mh@F5;uV) h5BP⺃6|w2aO0mx BrZ(,-H0f=sOLѦ Hf wed–A VIrJP I ̠X5OBI*qrrM,\)=MfXd!)yMÁ1xu]M+Tk,Ӽ, Di*H).%p GLc.Gbؙ|Aa ӥq(a`Veog [a,-LIk[bsF[W`*LxGя;Ojc.R-2(Lw_(PH)z[FݷQ RTQ ??p6_Ip;Ko |{p_C,y1baP I/5?O&()5 5w^'ѐ˘Fb>l 8;jf}*SXt@0Q*qH=(<=J'X (Lքi0{N(SRJo=`QcJwr8(Cb٧=7W?JPh Bfs DK4Dŋn<~80QD8=}(ڗl>w)x<_O1^fRXw{odk+Mab&aA]$6^5(XP38s y97X h紊W뉥: ]9V8PD_7⫝̸ k\LLe{a1y y88%y QVb$u2:mJ'k % m ΄XjՈK:rsj/QtnV(TT +wZ<^ .)$ӗ 5m<צY nʀA;0m-ȦWJ{_g NoE&Ž&`^> ]ۯ>k5 `0T!\ 5 Vo=;}͟bQD'Z4d Q6,=:BAGv`W}TW>\Zo$ע_=;R-&J,:!Xky>;{O0 ČgKLۚ]p?4 zBsf{0G_ђSp2ZG*CL%ĴF'"-B8"''Pve=ӎ8{"ZFdqD9pP=T f.YXywy cCGecTtLrbQ{da r~p?=aj=0f %D%_jv+ޕqdB)VK=̵ ' d^nBu-%)/$%HJjzuWլ^)Ֆ>q H:YHNVVI0ʬV &.,bM-*)_ynyA[ {TME,p5k]NjڋԌK1a.yM1V_hP=-֏m<5c0sWWyfΖڏmk) vj&=O_٬KӄѼ6܁+ˏOӕe)WFF45|Y[}J^_^Uz;pewc  2s̚YAv8S$=8MZrJ 06cq6Ѽ>K^e5 'ƥ툆P]xk_yyyw+ta7I 5HZiX6F4̪f=^YL1Uɹy=*nD'9[<jpwool/T4|y}_զV M381@$cLLҤ1G.%uDs&t)mB/UVbrus>IӦtwym{"R`j~tQ)kZF*=VZG #%qޓґR^R4v9R?KJ)qohq^RCfc/+]_u#H4ޑV@YB+brU+ "<\+@WP3k[1zZѪigPt:<ڮ}Σ-c-j\tx:97TJKgK=VUgAVP&jRH'(RԦRӠsRJEXETNLmS3KV'PibcA dzί9 &ؘV vOdR: 8/BOuAQHd Ι=;uy<){v^O\܌)歇\Te I/^!X} k}|댂guҁDPut:85W"+QΝO•i#n/GҖwYXl}* R"HI- d|]֒!KӐ+&*(@\o|||xӅYOvyk>p= q Q"f@:xPޜWy#Ǭwԝuu]e7ŏ~?S2S VY~zs>.3Wd ` n2gK.}{$ VU9kF<]='1Bvw}ٷ=dpO)u?_ƿ;LoCJjia9^7Y*L($Q+ vDh C\{-1r F2Zf vY_Κ 5e^)0q |x X6~`sNޚ gY^?Kk C-oݾxkyv22rQZ%P i^Mxl/fg?Yfm;HCXκc:z?C+TwvF/kr-T𼕧V"ęBP7TXNDU:Nypt( h4 G^،R 2`< rO0iX\`2tkL Xg` {95ԅ3ô˰b-@5 S|~1! Lcw 7o@CF0!1s 娩BebӂJ)1G4 HJjQ 0AF*pQ8AqÂ$ؑL4_d%=Dk0b1XV kMdT UX)]T3 A D(<` H@PQ3B!oA-DYKdƛ VFP# VPljrZ;@;!r)^ ]cSh$oO89~ўQ~]~cݪ89~>Y~|Mt9P3ߗݛ.~[\.{~cw(t2կlvhͯe0$=R΃h)jLsPʹk'm]hqk]8z۵d-|vR6ޮӲ͵Go8ĭ]}.JnAIs@9[Fꀴw*`Bo#Pܭ~qr#Cbh~(n_ kCj~1T/H /U]^<50XBŘ-ÃqFXXXXX3jCI(~1HH ֑!AbB3 {Ņ, qD`b"`b"`b`b`b`b/D/T / ԦO/ tKٸ7T%iػ/wn璁shpYИ{8?wB)m@R--D FxӺ:D;V%g^!`="=\bq"¿]0W3N5Ff!>ߑ>mѾ鳦JʄXYs|O>kQO5XS:ڇ=yyg'W{i!p$r?73pnoF߹G}p䦋?N]L7[J5cپ;b|XY |x4p]Κ6$筂[ vE}jLkqcG磟bv~{F* Ph53?Ŗ6]?ݷZ} bGt 4D쿎??0&DEFքlf'{*raj8d,Sξvw -2c[{,TܘvՕnUpY4zѰ S]naw< { ?^n5WuĞ5 CL|":'7gk7_;#`]%ץ_iu/|.W;E֐K%ňZ{4 {1O) (f^^`IkE$RA{q ~\o!9ZHDB )F32,uIXMd)#! XJf&2C2ϛS ځ#ȍ8؁p"!T_AH,} qJP s3!5'K(8bLvXFBW 1Z{UAhyFLqm!L!h)4h9.BV/ A8.aXy♷ȩ2:n Nمr2 +d|-H K-W23aaXk0 )3(c{\xI+D(gE&̈9B`9삍sk8:XDL 85hO 0G,@>/qr ~'5qIz:qE8q&K"YrZ_:a܍O¬)wo\FDž!^x~~07c 2yrrU;^_Cz>߹6][;-`f|>s|Zo>C(kXG^Rz{!`XD8]~^!DΒ. au$8.qe,0xg."F:si]]o9Wٝög4be32$;V"K$q߯ؒ+be9#L0qXbX,%D_Dio܏9|/tپ(L^ۋJ?+s?:W ,߆?f12}Rt{W 9%meJ U/yS%  \"P>NoqJ/'ӿE>pnSEU7.TjO/PcƜCtp-ݢBor.^\-,bzuyI^iΗ1v[io2w<nqcvȥ`d}ф_k9`FWf7˾̓@ɼ࠵ʘeyw2tvyœ,E~;5ZNSb=PcD-00USAd8cTzWiZjӻ0pE!}!ޗ~gJ/_"ۙCtXR9k݈t-JaE]A-D(UC0*^@ 9`KS}/~o0x ~Jܬ$8_&|Nh{Ύ{Q}~iJ[ӳqܛT(=6b^Qډ}m |9D9(@[]\^ັ3?q+en>_dCU{yzi,rg7AR8՗n#^ٰl2¯XDoMX OˏV52$Ρɘ_|nߌoCtV-F8D! ŋ&dSn]LCǃWٞwP. ]j{OsA+g;>CHk懳4CB zw\CR8k炳DYE;Bw$rnAYYjW  1ʹ7Eey>zL;c"(Uz9|9dLeTF$>+9}:I\P2p2X/`$+c}4¥v3{O7(PҲ$nlVs5:n y.sFo#:,jOr%5R}[j@X֏4DN>2Tn?)0ZcF# Z+FMz˸ 8w[UXa8֔H)VEY8-p<Ȣ3x(x^NJ F0V<_a17&rI|N>g}gjXlfg6$QiQNS+WQO\Ag,<+w 5`rS>k+K+E-_zY4/9߿utn8^~g-͎n TkP{1!MƷJr#59eU^__pjUIi޼3#_YGubE|l vb-/ç 5~ҰRSݘ95jiǧmyq*틥&CGJ8~O,|#xvM&+N3 7MOd` ޛkesG[$XPxXnz;cnyHPrŽfx=;5T3#<5T\(Gu"pprjO~R4ߍoP̯,bvԐdGd~M{r/# 9} ̾VQR@r'0 ]5ۂqGv~*zNYQ}q f"sfO=кcFc3Zh~,S9%H!+j٠$2>'Kz$,r/1 Y0Te.LORA!.|sv֓A!) ҚSgSZ ;JST +l4VD>7Kn4VɼRpn4vXP]gtgR;r1~֢ӣ;Σ4$[;{Y2jpnV)K>8?|J.w "D/TC&3Yn\9@|3R^\QNͳk6!qJ[e~_+6]yU|eEv&҅df 9m=<4φ}6w_|hD1K7 ALwJ2΄`KX\we.lv!GzEYlQ~L* "ae;9@KG7 ] C'i|t:ZIj嬨#z B?]12 F5VRBٴ<DOl=Eˆ:3z#A`#4F~pPy$i=%&TdzruгshtdA_[(wyWdg bno@WK#9cw$"%te*D(2 lN29>,w&-Β?s2/Z1zz?DzzNMu|ʎsFIx'L0C;_Vsl}sH8ϻΩԃmtos\=H)O{4}섬nZFu -U8Ձ=$@UxNȴ@㫺]\nwJ=7c>iRKi;=PvҮiv#n-P&B^!Ҵ\Tкxi~C͵9^G흶sfقD?kATZt=ǯ`%Oטa߶$58wX ?X ?VWۢ[5S64ڠSP:.e3 ZH_p46}QM_@MwUgoBq8o_:wޫHῤqNrѮ鬽9}] ]s6 m?0istQ+"Nlem߻uQDu}1u 4# α\ VwrRKz7PTk2M\ 6qTT7Z`J煃)^ s|`RLLM/jxH]'p\sg'#wۻ;ۻ;nVqSXH\rs-]0/0 Ϲ':Y\7};; ڙS:nܿTm޿XT's2\ '>7$Gh7ؗp^2}ڣP>'H.47N`VQ1υ0^"E#ET]ghJ5R C* Enj&N*OVXž0ՠXU}!\;:,&-B-q!2%*M_Wi DA- R$$ɭyRPa7QGX#-6V҂YcX>@ jq9Tp0:ktԑ2y;u۩۩4+*T IgR-ر- z~ZL^q{LgTNG9D{snz3Z)ŅFQhuxh nMGc*k0ijzmA^IF<^3QJouyn'fIWs`by"S/窏XQl'F1!PX򢎈xZԀ&owi"6IAsPIB1Xa$y='cg)p((1&~^k*|ALnjIAn([dy_\jA$~^/.ZkT~|7IM( 'G vtXNI! %J°;"P$ 'QXt5XͭFh!GՒ+-)< RʚO0J1~%ҾRJʲ7N{(\!Q!=T >6Rф%%ÌJ1Ih={hxkL)|@ J#j-q!b!1 @z&uRSpFT:/T_&ݱvטZS`SI7A]#ڱn|'Bw*71`r c(iA EOq?%hQf{'%5=Gw@ Eac{%#4U ^%:-5&TAѪm*FHFe-ݳ%ow@jK$nT1e-f (r?} @:kӝZ;Sҭi#5G$^ȴ+> ˲{_U}x%y6j *)0eTs!213+u=E0'῎ -0ۊ膔Ew牍"~]av!pn4į>_]]|½\q 黍 Ir-s} o麺;ӑur]d*l}U~(t]Ǖ0ObON97E8a=7g3ٻ6%W}YEHalއ_=+$Xo/=/YJ'3էOWi:6֐1r,0[b\y=|?Vyʋ۴PUxV۟r8걖T_B$VXn޿l<ȋê (.ȋwJ ~/^8 ]p}F54m~zBМU=.w[X*9̅{y-"ZE" syˌc.28":@IdEe]uDSfm -M$Z%/"\kX=cg`s c)Q/J11$ZPd0{=h7@+͜u,]TJ8Zk) ju9嵿z{DF?*뚋0%#"Ma >{<=aɌPX&5AR3n fbMIA4+o . ޏJ%سoZ_?0Fr.6୯b*0GgZܨ5b>ߗ%`;pѧ|re0^bT|qi!c9_ Nօ HT6W5-SR.^`"zBX;bfw7w}9r+e(-+0Jeۻo'O0V\]*~Ryzo'k)miCޖhS&a}A)c1vrҺАoH:H5m &Ip?|`FF$xX{iJ f>{ ?%#T.l\z9oK/{ܭBQcsvqٸJh&3[ѡ{k0Pei15\1Ҏ5~Vƥ;NJM ,Gӽ%^.g|E?f7|cY|b7˨BxEhnn5K@SRv=U{]jP(hv_kocEqPW*>+hW\xy2Zޑ;Jr 5wYEkI ,$yY!@>(| sT ҖXUU0j 5RBpVz?{m!7.o6,y*źߎ^nUStΏ^>]WݕLWm/&8[*vi1?R8Wn hMS7+%pћ.қ.yV\6{nmiXn/*\ jBnSDCeU:*}wH#I"s*ͫE M""\,W2"x\yh:\ G6KJ,W^@o0YT%`z 5P(IBz]yGryӝpM'J(eZm;WsaZh*i١6|4%f>~=+rl2˛mo+G18kI@c*@Ywq(zBeZ&kV z## Wp"XYOXL?<*P\G=_H͊z9A+ 973 L(\$4::fI51WLK Rv[D}WZ ޙwhwkBnTc"&ruÅҶ+j ^oogaQ)۲& ază *+17(*axMP"nQFc7Ao'Ӈaj%P||:tԡwC ^Vʙ#TΡ S]og[JZnVG[ %TL~[>ݹ-]LAjJ'Oukdꁣi?䩺a58H"x׻ GA~wq:nMnMX37-JJԄ]"8-eq7b^3Bδ{?}/U1/oS|JҝeJojξJxX}h__J.듣pZ'^\뇾kKFo痟nsG%䄲z£8U%aJ$}l3=T"u!Qz *e)r҃:ejH5Ɩ#*'hBTT.>ۻLi/ה{L>v1BT.w5V5O1%ꋾGJ &m&LFCz t<(vJQ3Z9Xm`4Vܳ(ZrxCMdKT ]mR~|9H 鄹 /xfz.?.˧E䃏Fsv^aL,kj2v.qA6FFMcj>_~ϒ[#֞ VRG/>O\M| mZ|ya6.-]DaǪ~oˆaV PN5$mDxkSJT!Ɍ2sVY|S {RJ\LX @ irȺ~(!ږډ)wQ>S{9pm+?'R[aeûF/~F~rQPr~>\$qLjR>\f̋ZpsK!8Q # 3'dNF[_jb9'ކ+a TBPK_Ok_2ڳ _F9Y~>~6e g K~(qx\Z5:"D‹ /B$(G²P9x j3zp$KMKK33q$X]9Ch:Q#te$N9_EIҷ6Wu<:FrRs /i 4ޝqI! 8=zJ0]$8M͹#ulV)iCA('aDݟ##jjwGi/,(}$SO2w%S$p +!hz8+hup)Bאͅ> ]N˦1E1eӦEj8!s#\Uj1ZAN}`S@C@Nd"/fJsQW̒d>*ebݶ3&,R# W\9OebS-?\x r<^ !s \pGr !P߶Qa3+OA3nHZ3#RZZ wZ2!>aX~ؔMHpU@JG=>S#0q$[c8% զdZ%8Y#JᦘZL3٠;Wsww6}շ: k1ERz{.ܔ/.+rw?ۑ=8R!|5d5<-{gX r>觫k^߾>cɄ? }n:׺Gu:SZۃPZ  (~;wϛCY0 R &te8CMfc*,kp\V32+ŏ91UAU!Rah(nHзq9QሜY2Z`mI7?$=rnHj#cI69al6p\B 7;863%]5pN D8 1򇀧V1 *9oeЊ1VA=GޮSR f//r(9j4!b ROWN_LFXBR.ϝ2M"3?+Mn=˔s-.c% סZg1\dkΩ5th7"qAqmOFPZK2phO$LJ rѹɆD  ѷaf_=pv݈Fp*W"+PouԨj[vd>GǗ>M/y &3`ZYxDVbTxDn cPٷ֨둫+_ k{Uxji:ps4yzh[W3:(Um2Lj) W;窱 fD[90k[9~_ASi#&, :cLcVRܵ.qrO'|hqȉwS?ۙ`E6ld5_ߍչS??\g߰F+tt]Z]K Qr?ݻדDҮD4DU"FYV?׻"[07@Ϻnƃqx0ʼnb7 ڌ Lk-sLc vJ=w4vnӁz,T? nêRcEn̮<خ5Z^8zM!MmWSC ҞK5Zϖ]k ŌQ'NyWG"^-V4Osyx{(mg΀ BYL ;_ۛÔjژWOfAs( :W| dݪLre.lC]dd]AX^_0^_`.2L{~Z% y*DXirqzջl[SĨN9x=^(SM{n{#[UN5\n0[CDN9jy="4Sܻ5EZА:%3 v1~Nuz3iОw?8F5nWɫUj>PQɿϽ%jC]25+ bO#%PЯ"|g a#aOK-M/Jk=kVN {Շuq⺷Ef#Ao#hO_ªH]}D$ۣH\"aY F*B)}8B1_6Ǔ*j@Jx9{řJr)|9f9PK s,z!73VT %@k6vf\Á|2{=EL # f/O Fȡ`R{kd"eDkF!n @&]Gҫ(UGM:cԪr2>X&R"]ʘ8xΔyqg.Mn>ѹ^?]oru]7 x G| Y)2# vViiyg`JeÉ[acm`)y\eBfZ -#Rp 7#T =&DJ9βLZ)rҀ|$"qW)ۛn,ʍn ynm7 ya#(;Efd5i ol∎x t\( #{[ZN!@AQ f:pkocqkZАMtouۿU&)bTu(嶮[HֆpmSrWܚ+M=z\e{;A**g7Y gTZ?!wï4"vuhb܌ߧ..4!SP9O]&0NMc// Ȧ .S3,YL'T'YƠO{۶Kri_ڍTـf4cM1%)5Յy|@1Za|"O\O2ɄʽCh<h\ J*亢r.8]hxoTV1} 2nIr]찉)O,\KX2-$I ;o8Mcl-2i,F>.=brevI E>ILhgl (ba#q1.J8yQ]bxyDeIxiBlbqwӏHO vLHY*6wXXڀ@CY'?70:wk8Zm/Cr)ojÎ&Ģ6Mt(+MYu~26M ͋5iՇ7xٻe3^hnxYgw+CgL넦)$SdL,\i`lcڗ@)nj9[ PO%[}]G=ִC>կmdWu5e!(?ȳ,ҜDJcS~i2 "xȘʠoI7YaW䱡wԂYwG:Qyec B.mhIʂNe3Ӯ4t6`1XOa N=ө*l4Y:R aLaGJݗ#[,6_V<DM<em֞PS>?g-jm_0XRn`ȣ`0'0֍=Dmj& our%?}*XT uC6B1Yn2"xIs 3 p:@κN'cg[UFNmKf/o ]+tt_M D'yn@$=#֏Ã7l?%0A}b%ASy!2 6kqGρ zu+7G.+Ed3 HʼL GCJi}pMs!p6j;"Z{ H ڴZL> VT;1 u:YmCZwc !=śu&_ȑrkC!Pф곰ӚꭤPh|_l DHT RFt|p#\eJu}v1 QRQ|H,`tªR![1^)axS*ooB 2N௷9Y]3]Ye47GIN7$G"^#Ӌ|;IcX 4wh;4@CSEH[0&ӎN»X Io/a= YgeDu@YP{M/}M+xٻF$W,f10z c==y1 ]MQodJ<,[abVqeD$YOހ6%0~kҐ;6S.P}QgLo대77\FZf = .8yV!L[!ƓKb@^Rs]s~|̼XNq5YL}t͉8pE`ph2a87 b[|VOE/KB>f0{ʽ|fMRk }}>h><5wc>/!0aE|fLgi%poN9-ӶiG$sAW7lY=JJVQA&W,6;2, )aLIxQh.k$(RRL`UZh#to!8 >Du-&dl>RkGC _۔`8nYOҖ"|9-kf3rOqԦϞ>Uk!/?t;tfe3oۓ[\7C|G@>:'}ίܯ{FqDݜv?>o瀱7?ߜ_G=_&lneEaœ (JNǼqɘ֖)gY>d*R `ѰWV %K~玪gpgMvlCʯ݌7==ܪ|ѯdr ]t]#a7@Hc42Eœ5E۷fJ3"zRh`dI8gDe/}.iJ%"6񎙞Yߝ3V=GuL*[1ms^LH h=x".`u0.DV2-CvcBcHO?3@lhqQ'1=(u]pqa(<ĬdN<|w15)7t荛*=Pv;N!2"ю<=1 <93Axc9҈ MP+ܔrY3+׀fV`_RXD,*G!ʜIJ2LB/"K2зQ-}lDҲDDn|iZ蘼w)USIvYpj%2L^FVA}b^L.h>2 9&FR8GDd {.[KNC"ӄq+ᚃD-3%wm.+8S_bBO- JxBI):+!p1{)G$[NZfe ԭ<*{c &wΈb7c'bǛْr ii$!墧`xzw/$꒠KQƔsd?Z-,p<=v~L1ɐ|YCu0k72YzӅ Otrr2>&\?.b<>Wt歛@6::_jY "iJ~Sh)tGߟ𻦍A_/-p&x4(O(Ap08[xӫg|y9 f 2M@y){!? Is}kvS%J]`w[HeI^ YnaPAJ;!G+qt8,v)Svgwh= #}+m֢.m35P*2UK©297{טXD oV%Y- 4/3ƅ)Jgr8R(9JKl eْ,Wy3ɯhKepY^[ƗB8bNCzr )${ȟ9AҠFR:. @[~/f lI6vRd&J"N{.9 u=M^EUzb'SRJYԝb'3h]:fh0JށH)>2]ZH[n H\B56BNv{SVar8n˾YHw87_vVtk/CpӾqVV_ܟ[n CZ2J zǕxǓnIe-yٺnION*mvֿh=Yٿf-h2MW8`y+>e y6&R; ULaFrUUF&[sY VF՚LA*l'^ I~5H^Hj1i<.})ZV$r.:l+F?dWEgQdeqߏ'/3'y.QthOmˣmRDN5]LK^t{ЭQX`17͆RhōFYsє0l8Ps++hu{hF1f,hbnҰiuYSk\4zqN; 9i+|@H!?@x'_,"$*VŧgM!ЬWdpq1?U;y?f *>(ÝГ=(j%3OE^X4*6E4zI*d)Ne-)PuoU| YkN-^\Mw+qy `T5i਄PN痟ݬ7-֞g? VFs۶^LMF#>%&TWay*^NZi@tJzcygl-~zAww~SIo|8Ѷkşi,"ij6"1B[DHV4⦿rOlCqS[2y,%]䈟loʕy4,\ԛhPlΒ+{r|b_*(Ty!Gwϱ.@t^W· _nzRTJywDC[cn7u 7PCwD;8nvV'K]v$+wGGB-%j=QhKٯ6E~!3>t'xܹYx!+Aj2$AѪ*=&4*4Ř Vu@+.a~[70uZ6PƳU( ~6P.4_}5/zvG-i79uY{ƃr`W&@E+dcBHcn[kfн6r_mj*_e*JC*|ު%L1VF5 ɛ5vu<]{Ui1絋\Hƒ!rιa%5HH?oXU0@mnxyPR &C 1KFp #ɗXtR^>8ed:\jPj{U(6n^.I릟M RZMTDG@ݷO\޻Mpt!ܓ~s8\f.o&W7ɑdr 2hȊ#wv0 WhR.H)I\F ._/=@֪wx,]mLJ1#ͽ#}&Uia G8ZWR\B%{s_:l?뛳ɴQpEJzm|n)iՇ>kPhoOrs bhJܯ#BI,24)ICR:}6Co$ )nrjo|y`M *,rzq8)"Q]pجuATHׯD(E4DxC^Sښ2-"[H j!r8t]RO&ۛ2\N?-u`DLЖH那$Ne-Y~Szd8HK[~1 +(dyfFSX)Hj $9Z4"1^lB9[&{ClZиn"׳؜̫ep2Sü Di%I+Rߟ&pwC@ J$;$uq$':✳$!SHQ+g;!8۴lC)iie%`-\6\7@du!OPh$HC;mpچfD*,Wp/J[/o{lZСPE AY@pؔkU:4trji{{$Cg ~ziڢJE]`4d ]oӫ6T-:q5򽶡Jm?XxWmr,1`I8<,ٮuYc^q_ei0b3.p!9Ug$r>DA^9c_T ? y ᇖ271cIcIH5U0 o/&iU&3.J<ڲ%XFuN4Id{-HhQ<'%RJ rKT-H<6m]74sƵb5SR64r&Z^9üo:3wh.P;sv#J6%V9Ah@sr:kN((_OM Ib fuTgZSjUZc7unʝ-<}n oO.m K`4qZƍi;)hT/cIWWlQNHU)arrX)zcr00bY[Fi(_o'E/eCN rm庸`5*ì$ة9T9tDsyuD3(G _4c@H?12G!1ҲIOIlkֲ҅Jg2Y٪\n"5͓659םof~~AK&J:r;d^|-vX\+|p8xʫ(DOwoxle>Skُ{u7lDz2s.5Xj)wXGzzD{ч( ]~TD$IC=\up{ڪ%Zi Ҵ\%ej3(t6 Xܬk*{嫡i~Yy9LRu2%mɥ˒kfqMͣ0=&GO.9-$ڢʼjbء}rǩnk1 v8FĹM'!0Ts \Kuƚ٧eCǫdm~ݡVcF.uDѯk h~ݡ/lݯ[*wyw7B TІ}r\5 mĊ:ZCl{1vSг*uͱJf[B-ilƞ\㓒 cPFOW9ӘF lw Ahz2iKˁ76QfWЪ|V "%nD)hjIsSPG,#?cMi,ʀ@֕jHfC[ CfM9e*-71#NRo)s(d^Ǹ?ꎌ'<pYj۷b2iCj6gk(rbyPkRf;3G+Dž@~rKã30W1CH'`Y`7 aWrǨ[% 8P-@?75j2-e-e5tW7Q}JQ||;NR[MvXM4g2`ݞ{;tb_,zwi޿}{蜢ʪd܋qsv?ѽO:=G}YN1e}Eؗ~ُ,v`Pՠ{~ytNIv&@\́Q\ ̦yޭW5祠9 s@ #J:t / fV&ԪC`Ff҃ؼlEwB? O"dV;NݾvNJ`CZ\wm`ml>o_Ͽ}ؙvHXbٹD@s_Y'(:Y.YOjsŊ(Q"N< S z9kg/>*JrT(QQҏyqĢ+ bffPUdJ)hQ<'B^}kQn6*a^hcs)8:rT[F$sh`IAn &R MQ|jB.E .ti2)Oց9s$H̑~`nɯ/ޯ:tp?/KKD#ޭAVuu1T*#G1 †%+F|#E /}RUTzz?$gH~g+͗IiM~a?SE$\NrQ83Fʨ^1kmNh)< ut(b[:[kcE~z] h}ewgg~!e6|̊;J5JB5,2LD:zDX;a7S&vD YA4 \]Acr kSMBZb Oļ.;%3QeiM;nc&P DYoX=K).t֍(.' @-0w!HUFO?*9*t& 7U4IVIL iUN,54[]P@'c, He *2|nNH(;E>46Ml,(m't6 :"&P M~&)vсG  X'xP *YҦ[ΒJTE+ɛ1JBbIڟKbvМBD2oS ZOS֔WCenO?_"-\wpj IĨ"8zBDH4 ML3Hf;#[g.+}4̲v0 ħS˲` ̬ c`qw fWk3_Q tG LVuWI'.muY_b{{|M<4(}=ktZWzow{{99wph~ {v|c_~=M{-jm OGykM5Mk0tF#Tó[]{\̴mn" %Ƥ/E20 <[" Y̫i1)&?q?K邩W˵3B?/y/?4/|.9x9Z#;<*ݱu|nw{cz۽=Gh} W%O &j]? J1}prrrs6^?T#U;ׯ%m`:vc)j$y}q|tjְeگC\Q~|Lc|xuf0cPNN2vIOѭ%scFZԋwW)7܅kxꎷ|0#)8.ۋ~|ׅ.!)Bݭa8|Ν|ezƶP}0^/.w==^4|뻃E/??ҥ7׃^M#?uty}9vGbY:غs߶e=|qkr'rC׺F*8CPfqt| 7R'O_Ns|]^qxGX/7oAz!tԙ&.dqrꕍZ~B5P56AqͿ̀ȁ E r & ٪ >z=8x屾r%|}},.8=ٚߨ=LKz2避ީi2Bjax,='M6'">,DDm#JJ 6Pj3e-b®sB9[{+]! Dػo,ُ̽b dvi0 Fқ(q 6l3 "4\R*[hTqmڳGMru[f I-νnrpwPd_GlwA){doZ.&s|6u2(ByhƯx*-Y䎒H)UjfO=06C7Mݎ{ K{ C8H,|7j{ٮ^X>(A/“g 9~4ǀ\ŵIдvB v_?}V! /̷,DIe+}{-s$^ױXEE8W7b/Fi/~؋Һ\b(x&A}Z1SoB ($Nh@0S/朔\bo>8a )Gj D"V4RWOxgL=T;Mn W?+fAXEKRjٜ3W;!c5`012xΥs5 "p@:8 w.&rI)ka^{V,mfm3Kl4f!ŒD%Y 2IC{Ms=͜)ϱ8[aCosxyfi(bcM;kۯ]tii3wƍ$u-d‡烝8"8uχNgpk|9E)%ݼ7jeڵfh77԰ %G3;efŋ+O<ѷiCN}"i~kMF*1Hifկ[2N'_( ~hC4m8xzڰ0m%ULl@6gp7j|slg=& p%'۔hk\UYg`9꟒bXV cbmL~"Imb{6ޝ=&~~ϥtN鸐LJc]$7ks(tjn|]}.n疝xh %,O_H84cJrQjָPjP8(.ƔT {cbQY:RT4d^]LLߞ\32#qx|^W.E1jHBt c<.Ii|&˨-#`<~2 ѤL6KIDQ䟝 R6;牖VzM8WpQe3ӤP:b:2^Rݳ+3g2~kLe;ӼAʂ}YQ,txk[#7}qq$@2Bkٱ{b~L0 "94ofujkZ&2T^ * 8m( Q֠aDVŴ. ~#ÏƂ7p|L 38 ,w!;j1=ً29P4ШN4РaV/bT <鰌V/hIW)S8}qx:྄%sS a d\FLp_[hIkeO="BNðEm"7W'νOxPޫu+~h^vͅMQt!x{\TP1$(1'O#N6%Tt&ȟm9/!Z@뫛xdTj~4}‰bDh%cܰhc90wRp>pu^թ-P@Eh[^=[WOmΑȘ&i{T_JD2ɶaPdΚFeBUlQ;G얲hfȇ%"hUxeIYvNb5βY04Y7,ЀJɲ=I4LOJL[RLtS2ks .Յp<~{ :+rafޤh W&NjmS2劥z˙3eQiLcj΢! ,' :B1t7>1`)#x|}iau!>15Ӱ[a0:3%w\r x*@(;l+ >h25*-)Vd4aoRq g7ÞNo<8\z`eͥ`MowIkIF=}!_:}cޣn{͕Ygtay}~rYHmsqɰkuw6b&Q5Pӈtȧ+E\rej|IZ[}R%.8Br-d;Ézݧ({P/%QIGas; =@;ʸp8>LΏVBt_xHWqNŎ8Ue;Loʜ-CŠB Cރ; GMg#]FFu ~/a8zP #z>8 v<SaӶ'9%GhMGG$AYwdXzd! ˍwO!K3)ܨoIO*@лꃷ Qnoܪoɠ0pH#}%-zǘ8fzr~1!bgϸˢj{n}1k(6ǟVMW6a Jt粩77(겼]qDž:6-`&c"cQ+9ld/pri$~1ACLYNqahpTВA"S\qUc֣j{֔N PrbGv4jʕm\tPR "\qMn?!B&XeςLtm 2N{R r#56R8.YSNF}"z~d~ R)ؓC`=9;'g1=9+]|mϚV;4$12 }Q .=rK:#d^@bH 9!ric nw2WM4:}:73"k^ڣ BW~% hayB7o<-`8kӿAl_FOE MލcQ{?/ǜ峒%sFױݧ"%@_df/OV'zߏ80\=ոw9Y[?z}s% lRG|Ԫ[Hzz{MgtQ?>[RlKu.n] ߬Ct (JʛiU]*tNIZ`@[+wq7fîzu!nW~ju 2a_aoww^/j/x0lc ~‘ @ȷr}.r}{ؽ 5al]6F]5F>sEUֺDU+mٶYeR\ _4 s :` TIGL?LE 0"&aS "#kq&ΛPC аOm APB,XX1Hؖ8 d`0KWǣ~!CRzf`Sֲ *%*+*GkZ֬IhR욪/!Y&H-&UhX`}|\l.}LR_qCڻC]?9T7?+κv @ {'Fp(g0SeZY9"7]\ʻY.f!ZlptdR=ɐߐdT{ lW#D7h@?\2%d@L=B2Pbi!ўrdBdp)ҟ˲k\*y$#DR&I JO垴[J q?dz]`e(aB>,?FEݰُL$9PS~TcnZw~z ŒLDVkDUgB*)!T[B*sNXbp UW'a2Qߓ2ȔEvwbow(;vl:'+tQޝp< l@zy7fShN}kOxQ72`TN}t2tS# ǿ^]QYW4B$v,؋H \#0cE~87r3@ZQ;zM;#Pxd}Қ[ddՐ+ \)f^|kzd \hFiא^dQ2q ǙpskOq7X[lb5LP+}Д(EFM"@-gaڪZv*HUU 9֘,0ƪzH٣i%͏9!Uۧӱ 򫊸L.ӈAhL/D$S8ִ "v/й#eiLx T~=C.r Ĕwo~jn(DD-m.߼{_P#{V9)jZh=R4X &G&oM.hUږ}MhԢt(tKJRBa T{v]?quֻO()Kʯ) nNjmtH ;ض J-kTqͮ)ȶ\T TU=CiYr=wR)"RՐ#A:ǟC*ǥ9îT= KȾ':,7{d'`ߧ@D{϶7GZ̒=V}KtJ R\o74QBNN*j RTgh88SkoYƀsL8Y>xfi,u۷!$Z(WR@D"{-~@$" }ă^n$zw ȠTʌr+c O럋q]ʐ{qr7hɼL`vOV:^O_|8\ݔz[ړM\&]d;VJH־ _܂Hv_<(`8SL~X! .ߘhRc5AS1Ȏd%Iۇ_mGeN0[.ދGk<)~yə'S7w\(12SvPVWus3yÙLY7W'۞{CdC|֑fqYHI2YUHG{gneh禮P/>p@[]_\x͉16{|\ZPG_$':qYxJyyOBFo;_8凶p ̱y}{u#[+w ^kpͶַuxWku?۳߿?I,`sҹX\tkf*63kW2;XKy x9mXazt |WTf)]d돛x<~&N ?ǍHc6rKv<A@`%sߗ\Ԓ%l9c{]>XuZ^}/?A(?z'$?^ovl  Xmz]?V{?TR, ERۆI4CY]*tcu.Z9)y%gVRTu&r6{kI8Mv(`TE$7 za1lWǏGv/7dQC^ugf}e6h?,颛hzꛪx`4}N9 nS?>fk}1Bi/_4->,+JI!m[Cg79;",AF,`H&Y)_ɬhq){)(&,;:K SRRƻ.SG>@r4 6.BOK[.OjZY˃[ٷX9F9A+5!R2C_f̍/.n26k emx+=ƌ^J7f:1mM&%Vx@2)ifDV'+i^蚡ק\<( B4E.18MPCy?\N6{qZxSlS3̈b#yɛp 71fPgvB:MYތÙ]hjJ`G-U7.4ӍJ  šf1kz/_^Ջj)ל~_X~g3y_?ZVufq}t;YLE-bO)mk~6ƜjjCŪ_lBfd7'? (ʲZ(iVAEkZTIb* ;ES5:ΓvhФݡ/Ye .9.MUCSI3,Mm4ؔ c\FZ˺fSXa;dju^qXvuθЅc7㗂K۠' _gsCLŅ$!ڄȅQKtu+A2*"Fkەt54Sa%(z]~W Q9E}^ Q8x.&@ߍy5'E砥0b{kS ӊ$T.PV5bD]8}˩bQG Fאlh [O|D3ۏcV/47HOyzOls ymC׋:"0}kͻg QB{|t}Wc{>yJ=ѭ0ߞ\[n ?Blg[;/)2Qm Iaȓ֓ t_f|Pߞ2k6Ni?uZPZst7A/th>?߲Ch޶#m A3Ch!.C:w>>h+B7x- YڸќP1KM+ƞ=X;K^s'.K]W~k/Q5!,u.^rYџP (;wTc#ޗͣgYP@(0R%2^PAQ[y~ǞCMz-ryaȗ7h,'Q Dcq"0-G9!To#=D\bZt'&AIAF-E>։÷Ӡ{'.,N 6~;Wd"9[a'> O {ҜL 2V0E?X{t>`'@Ygrd6 1|lq22-FcF'_ cA܃JgN%#uI7G`jBgxbp *B$H<@`cqzg6nJ w"m\`{TVP672(>!+tpYj9@NYvFf(7 48,/U°!UeLHRuV />v gΌOElQ Cȳ#Zޔ41)RF}zEEz"YCf=ӰIQ@1gW7KdcljgO1vc1 =n{"a1T}nO1Gn`mTsjZrNI-B᝜n%)6y6 0n8Xrc,;,k)م~"tc@.q;!߉<1|pf{vȖXk;I) so%b) va/3, A.m 7'JBJJ+Rs:LqZMq.є±LGouvnQ 4|w'b\b8_R=R|ciTSa cݛ7$-0H"F!N8JBt" ]4JGfݐR=rYGlhFHypեxW[,=g}'g{S ”Um2`޻ #qd+#!2h{&TK JQG&i<%VBo?ٮH kl:@R9JH؝OVUЍ 7?m? NV?B諭([w!,VH/PC*˺"t/6\L }G%_BD`pgrժsFMG(_M"kLE #lCV2_quٖm *HӲ=Lڸ-7˕ ‚x=4JA7*Zana| k-X,cf+y9Nxp@ Dc.2Q;^F[OÝpyK{ΨWcm@+&Ç1iwŘ +U!PR Uae^uh.v!N5$1J+dk6֓'AqH p ONTK{]~v*b[1[/RW9W?JhUaLS'KVdto(!. Y:x' B^}]N^xTwd|6a%쌄nn: Փ7v)#d ؀P]ӭC?"d&W `V '#(hwA tSЁ a}t-|`cYs[ď;@&ݓ!cQ\8?%gAi)ԅRIsJxshY̗v4!@ٕwLbπQ|7ߗ'9!&*dd$q ҡ}&@p #-ijRoMp- =P𦵔j`Kܘt0CJk) a%wl|m: _pcoT#glC͙sSsz ;_1F@ '2cAʱߙD;ˈם0o F쌽K' :1B#{^vwD_= 5HX-cC_ VpV:({LPkpUNjtБ>Z|-D5(v<+scp]T]I[Mٴri 0"a0JR>L#-0Rn+7Tm#P lf—eAH$,Lj@X1wDeҴa!:(gY2 sy.8~G[n;75vԞD0{r;Hq/ ޤ4'E@\ c:bIhɁN :cg'\N%i FHs虱_Bcuӿʺt^v]Pvv{LC7ؾp`6i׎kg}d,Gmy7O iLGvh l a q ݣ~/{q^&~Z ew{4v7H0k$&{q?N/on_d66~*sQj35uC-IP0A?]IMR:xϛ-Ρ F bΧWp p_Kt zqnuGx5 Q15>iw .EKy/K6zAVY s4'+!Omq}h*_<~8qwfj2v%|9xalY,<^K1z?\ ,=hw.onG@ӎiw#-ϳ\p{[bܫ^sc< n>u\ej`B﵇;jYA9, 7ލޣy<ݐUbV rp{P3<Ә}E;xv~m~d ;s4.\BqkG>6 i>4~{p{j&"@! Z jW_{ן^ bmFyۮ{'O!$8B_SF)N s^_O:YQ#s 3yn?0dE(iX'h4H m3ug|pM(M< N7c g#[v F۷;əLlƫ9-#.} _{g~ %zPQi]|L21reä/'74D5g+p%Rhq4 L1l (_u\xb#/D͏0 J\A+ij&M 3d4ۻuq>Rf=dx*lW3PIWʴ  sԓ@3V(:`i[\V5fP-$UZ1Y th 7H?Ŭ24ǝ*dZrxŤ .+(NKDdӦ42hehB(*P0Dj$@HXrI0K<*ݻTwU-hƞ0NT+`&Ü5)R)b E )*;bO6"p0G5[H0H-V萂rpI2jAjo\N\/][,2.dE2yb%V9S<󾔇lc)e(@^.yE9[|C+ȳ\4mSuBu~}=gwj)<?~]!v÷ {˗6m5 I2r^D%כ/xVas}W4~P?xZ 4ArwDZG LuS2JLӂ 8ONkuQu&&&T4[zpy5 MS;>(M<;j잢+:1UGDs> Y|^a@VuQP6EUG#+ul`+e+j 7M{gghapGJ"vܝjN.l'ܧ5dx>Y3O,Ov݉9X-V҅ 5 p X%4  ʻ;+TYG8gi"pr%2e\I6bv7*2@Sei0qPHFP*l#ms)RMӅNu5-MQ~i]OĠ$%$ŠRUAAۑ0E SfRbXQ%ĨVdi"9ѧ8Ii S%X8H]Zj4 aSκg)pjR% .MS<0`ԈTB ̅#uILoczӛĘޤ[QH˭\,viuԚX bA !:"nz%q( [EVDabQ!wv~q߽cgvZa`R\_^ͪAs~J-@z;`>TKRią-y b>x֯pʮn4IXsےe׳_:hRtjTt4)^LM7\ }/CKSz >wNaS >"q*DM,d`@p!R`Yw0X=V_Ao7^ׇ}|ìzXV3(Y$Xyz)ZjBpXm(YʳJ1((ࢰ@7gL옱xQ^S¹{}NϦXf\r6ʲ8㻷n; KyqSooO` G {ڽe߈C.kkD+zcFhuv5ؘ\vJAXh__zUlֲ"`|.;q:R|ǿ oǽC6~({A' {`57{{E&^y<90}xw5ɱK M2|B$R|g$ԉ\\c_*Ў'4a<*-BH L'O%Xۙ|q;ҬWZ,cl{RW[/)v1d` eV Z5VrN*Yt*j^;tN.US䨂sRJI܎srlh[R>w4Fܵޑ[x?d F };݆t>'sHJE CB'JyXm<&Y]eB)ABmnBaL\BeE=o~b :oEF"B!H<ٌR$O{^` Žs"cޟw?Bo>Jc[JQY7#9.]WTw}8H H6,K _%Ҏ俧z$ gfEF\gjz3ku41@/h w'0g 9 N6A>h!`/qO/?v|NW~}lHv=;ۖc'WiA?Ϯ/o䊯0W7>ɢ͏>f>4H/: cA=AXt̓8q؅J7%ȡ8~dmY܇1~?a`p&w `&Г>I,ܤDq!DL rI4""X"˻RK"gE0 +."š1o;!5Ljaz&+P#ӆ&jQLZH%s#*w.[./_c_nNJQb  r]B`@h}q? Hsۭ;J[6{v"Kk/g 41F>_@,@(WNnU C~V<ӜiIY;۶_ik-̏dMjɠ_.tOfx1^™Pw> w|܌rF,'xκ]l|3nW]H^u|+yrON;9+~6KP^2xp֧OΉ7٠nklP&Z%kDH8d/HDmIΰ$fixo>. &z;˙fe[ղ*P^?MVgV_4O 6 VjyFj|өvj8jf+Hb3`R%KYUB9E)Q"L<tVCd5&fPMĴ1/1uҲ:XehvVzue0zux7=)DGvh}^J䀲 y) ǒ{CFcT:H6c/z#z#$Cor/WV`[!03j9ϞcDt s?Į9O#=7GFF])C`^ľ0A$R0sR$T̆}̢x4O5?]U[BJ]stL \sϔu3 u/T x ɒF+(꯴X#a;QAa ۣm)S#3mkyTDig[g%Sp(C9JBh)%(+|?VJs EdV(㿷Pc xFqh%0̫IG9Q^vˍ(A;qP_U'diT!EkWh ^b8@VۤC7tj_>-Ls025`L (²_TJbъb#S4lE 7􅀌~I6As6LšdQ-?rYShzExYF%vɢUl$p Օ,:AaAcgه,@bxUy6&7ޔbmtG% ͤJڴb"Ԯl­КʆOLTOGmᰀ @esR״îB-N lutM/Yf2\)C tes脱/{U9! 01#OyeZ3R)k,Y&/LM"3d{zکOdWo €:偶 Mg@ `]yI hY,;QSuoh8=pA/%%uNs|D8,}ǫܤ6L8wu0ysS:  !T*Kڔ1d2v!Ōf˜i?Xcnܸnd/Qf%憢 3)kS˓)FZ(=s6dcUE 6 c C #a*f{+sic\*?3?i%̏NgR,Yd_A2; nRq;{ͱ`5uzؤ剚!-r<|Zr1EJ\EpkT냜BXD(Sln3`cCd lxBP7E&h_Q% F *T_0O+oDxf_җeū ggE1 (LN3#K:S<T+ {ћ53\ 8]6 SbcJb;`W2$, Aωɪfj#)L^mPC]Rϟ8  /f%G:O԰Yٹ^דx[p[);yi}qYvPF{ReeIZVovWX$Tg%s\GVlG8$4)MNVkhnJ)Kjim^ ÁZ]q{aZC/2z3Ix.,iGˈ`)QVLF2~)m݀vS:X =WnM![vrR*UW[vm\$UT8뜱mI[oliuDW PB(3=E da;N*+"℣*O_Ҏ-QqPFiG8"fj[yd+֛F*t mIɋɫ9tpOw3綠:]Xwq#ބ"e qmQX)5:Me3/`ߩg[&gCG&ufɒqtaėr:~4lFCDx"c:3=L{C8g՗2}ܑ5 NjF)&58z9ٲ~rJlWuCSGp4S^S.|_fcPcԯ1OI% !nWy8.v1Kf|{s4Ѹ ?71#49N-W׵\ǩzO Uk٬$]Nrf}=w X_W<Ԫ;|EHh>HL'2./VF^FCO9.ٸ2j~')$w$N!=7_BE [)*Bʷ"K6oyy\~w6v ;"6#) c}f7󌛖!MAzd"sF }E;A{V!r^-_U7ǀƀ-G|$9mwsp8o94?r&kqol klO/VrzB׾_du'BY'&B(p*_Y.L+|v8|`Bc;cm7ͺZŧ+k\hd-ۦ##Jm@Cx5FA m5_qS([@=/4E"=h9iN2 EP^{9\ZGVv L)A/_Kb NXiQduF$GTވB(y gi_P G9\GC\P=*CFZͰ1!Si4#a[pP"UPW Zx%]uPw0݁vSB`gbav⺇mId*H)VZz=F)Mە5j.A{W6r J/VzUJeSv푼dKbHʲ  ÄkׇD|LstѠHdeν0AX '%F1ݴ^[Ҧ@!trfڳ2Q#>R+ڹf5vu5+B& =W51 ЕB'JiP F$$Xr(BnQRm;9?Xtr=m72Sz8^_K6&hqTo* JyiE_w.6O^_UWSNL<*R36U E:uhhˤ(a)5D2oB$cWm-&h Lq/*2r|^PZX!Ƹy) @$r5MS8Q&DqmEd״ R8csƟgJꑢd*k'-U'|'qNJrIK\MIDyN2,3a[ީ2ި֔l  $!Y&u>rtKu٥ͅI<IA+Lv- ֻ;BLK1gvv M0 n`7+zְ$b5qիφ=F-ֈ^}E.8F`( (ASRs0[adteZl}ɉ[mjzN) ,d)fa$#ZHDn]U(ɧң^pwToOvKEgNGp49̺sJ1Pnjl6~(zb&}39Ib<+;0ދ^X`ǰ?<8s/`//ڥmĉ}qrqkܕ]_ށ./aGog7P8.Qac u#ڤ~ox~ZUf0}L*du[uoX9~QE Jrrɳ;MB׽7\%NŌ}W>;cN}(U^muyZ@_x_8AP7P|^n|' odnUdqz Kx<vp6?y!&Oe*Җ>8Y~@k DSj/&N(y$fQ}c?c>Z—ӆB:|ZWD{&h׹OtVE7v.nEuzQR_CȽ BZ'm8(VU> ] "[:HL+#DOer)g@Oi+~Kը\FgIP lß6RBC3c\:<%c#ԈD3[k2LJMAƄ6W۬vT}Z+=QV`wv乀M\}{_yЁ1agPOH%z,huP ݭ(̃4?h`@E-kB;`L@_& )qF&1"%FaM,H- WɢDX >LDV^fIRؽaޤ; "KN* _G3rXbX9՘r XaqO G$9FJS%i *C16:"-E0{bJbid*"[cUA֚"KWجϺC;g@Y"bjQYhC0/@u_gJk g!A`R$Q %jJ0?w$7qO Øk$dLFRUɌW*ň V+9a8WqR]r_h¨Zz -] 4A[-4XSYh +gĖ5VlݖMЈ#]n`ҰJ .P22gDw$!Si^6PJDʾTWxd񢽙/7@wn?'VϛC.Lz| Rה0`ENcH/*8kUNZ/b*3#C>Oshg8=եl9QY?1tMGe Ai_ސ p0~0!6!_w򤃁!> x7/?NGJNJ YP+ Rؼ|Qbl,g[H)?,q()LiEv xwNvD/l$;a17F$1)G\PEbHb)6PMfQ;0z~r)9f(Co*(ԭ}.Xa5c+u*0O u}:uƱ]34g yQғNO^;=yuQO.[WBRPM $TjlS"έLQF"6x~^MB՚~&`sG(%z]EgNƋ` o*4YNXj%BɄn(K\T["NR*P aklL N;TܵOk Otnmj[ ]`M($fǧŇM D:,zRZqaIQ.X*VΧOh%.;Na/n#wvb?#J }a!f>Xc|KR/-]bj`.v0tS!Ł\ nE铱#*zf1-~0vq8Mu)4!! ? |t|(<1퓁&Ie(Wʽ|9ݾuJY,:b9o%{sqP;ǡd-ΒRjH3¨Fo/v$TXP3I'JH}܀*"w yypSA-!%ȁ"aԳ=E%,$R8 KBtLDERg ejЙ%-., ɛ6Q+/w?A`-*؆~!תň0H3\P-3Fĉ'H6 i#}S]&O?"]y#y(ZM>.$zC8L1)Yg֞8G:x~ 8fR3zԎd( uE~@zy&=)ZAlGPbHI&KfP/ˇ.N50]2I9Rҳ1ZSmɑ6rfIPh*1 VMQP̧sցw' ş/{3"ьNLQeB,S*l 2" xƆG;2⯛lM5*]6V.qn '?iPY.r毃6yLyJd)Eƚ5iz\<Y}t5 dJ1,\=qy%[yd1KW#ʦp75c0$UfY {pfcbxᐇe9 ""D$|@g a _~FDĊJ[c>X :wGS\ϫcwKs׺6sV#<#Ё cZӓk' {ǎ~6)!'{?I^ 1'MJ$oƓ0\98WtlC!x]o P8<d5*2׵e^*Grl7Uvߛhq _O+H5$G 8 B3O~u&)^``\ \'ʕ(??̟"7k$]BUY>/?CRsx<=FnͪE%b XS)h,SSRf0SYf!EID_fBs}1Š݅Iz1P5d#h2VЄf`GZX-yf+ᄦNmyN~[~e}fq|W3j$\F<k_w|e}B 1Α{DǮ%?=N&ЖXY{>`ޕK)kot oici\%\ͰlWl\߼mmʢNQ*PZ~d#1Fә;*?W8c}Otm\iv>I"L!CBJY:Fe/b9K`ov?FͶ&%v_g4yӐnNQ`ֈ]|"YEVIe9Z7KZ`,n^xQvcS*GʶƴNBX]g )51+Pӷ]`it{ x"dn(r9JLQi#c%)㈯Z".Z~+NcJOLn2jZeN}w;_`/]`}uHc$\ \#3585|2S꺪HMp,7꥾(Fԗwaxj'Bb_TZ"ely ;GvE k !ޟ|u|%T=ytf 2" C|=v d??ڞ]?ṵ ?l|ㅜpq .[Fv_VACFbu0(F`6| fñ ͲsbBStm1D;?8X1F6WF~C'u~2l~կz HD?9[5'^5X5}\6[<2} Mxnu56iB. I= ؉Vff_șZ~F^E4u8o|>O@.P 'qO(խ[R8 ~Z3G+*Z)Eեt@*mW VI CFx+(,s[Å%qILV> #0Kä1 x<4&'AT֕F(W(TaTվSYT`r(i ]D &A31 4 {@5j`N#VĀJU2\QgK0êeduM(Krch&!wU&A2yȬ@&1@Wj+Eزb{kR4 yki 3S a$DRI0(~N$CMBJ b!I@WTeT)VlZ1QRYS XX˺L @APi`%[Y`%0”'8=IXwJC1t>˄ +GYQv `Rqm}r.%N!pQkWSƌD8{)`: B HP7% ,-5l-'* ߄2ȍ;f&;vi8JdqQIS"n0{*RRQ+'xVk&# ^)yUi5>*hh}ւCYjMC=Ks^$ʈj[AxG෸hLzigo*ۑhOp'+8%.]pP08As0&4'؎ߧGK%P|DQ_jjqCLjŭ#i"eSJ:*g~;VrG(u+Tg+c a0"Fd"52Ԍny]vN*դIeXuNL[#='g[Aܕ\Lf&rs ?[_wSv?^n74:K͛|18Xss+hO[[M^,X˝UuD2 =i\EtɻxcR1Qwn\g,٭ N@RlsçQYV( q? A[zXҋ" =c-)7*ZNnzяPS9 (6PvIkF a7?t[@5OZ Z/' c̐cXsPDaF Xo׮͠#vB2́h oLi篻^ǮibwQ>ݍ`FϰQȨ8G\bdY=H3"uJ:$$Bٛ QDB+ǹJRI^!H%/w X{jnlAu"w)U+$Bx_JtߙQ8R%u88_Gǵ(΢Ky IoK4=&|wc,.oo?xhxLj!e7fvCiNo P97Med#&b seްNY.d{~1{Qqou_s$ѴAΞL (} 1(8P_:{BO5jE#D^?7= O]JH(6c*OJJƁֲ3%u.ӡHX0DXRe~(64/w)2/ϱT&󛻙ug`55/%fq@RُE%J#ⶨo^N**L@#oI"|-w>M u淋o6kڇ .\Fهr6-a݆޵'}I\%BVW8ϧﯽq4uStEq]:4rZQ=T('57 [iZ-zRh9QWv)b-vp1,):QWPDzEB-4BJT*+8銼|pxwfUw u]Nҕ=)}rnz=])LHw҄=yKYnCU]r3F{{ `lI ٲTvRoU {#YHe>^̚׷~R8lN%ɞ'G(N_Ӎ+1nH:dЌA.ozϛo&Ȭa;{?|I'l&17 on6Gh+׋e,0FM1G0h]0Tj R3Q*iGRa؊ySoM |4m'g?  $+ml3;`%byȉn)EcٟȅD~xYᘏ͒Ws,槴E%0VnfSWWw/NnfͿ=j}>x[../^|,g//|_"]-iӴ^D6c2F~ Jq)ZϽ 9UwuHOkl=2b֝mU~ jx1YߖmQfE _Nֆlg)"vL[dajy=4soS/}{5X6lB[u:L (Hs~7!Ԉ~۲G q!_kʚֿM0f_̦~>Lyӫ ADkox\a13|3F3 ɿTw~7_&XtdiX i'o5OgtS:¦X)XDŽH3aK\\)RtS4An-ҼkMVT `fkcY Q)RUN`"QUJËt$*-fΔL.nZ;<'C{g7cD'TWQCD lՑN4B ,_iN*]Z'\As։PNb> ,[g's=baT3޲rO 5leԒȝ`(9#P{rH,YߤG> "*&Zy/l)M`x&6aݥԈ(;I3罇NjUG}*RHHp $ߚزG[Ϸp*]-,@cF6.17ZP+nLQPO5V- ֕;2t4}Or|L^֢'!{*'fDyώV \qv#,p4?A`gЏXϠ?{Uu fAY0ϫ LkL=czڡ}GGG TCM] s'#aŗM3<"߷_:X'Ygx}\ˏh,#Mڊ|~5nca|l ,Ww~%xWw~%x^ VOM!t4ԨbKuI 28դ%qXYT+d&y`r\OVHe^N6y52w33䚷~h9Y\K?MiEU9|m6[ Yl? r0 k0Cd0;A(?Nu931ڮh7@0`Z!X'Cnj2_*; z0oU 4#ZrEw`HnU8rX*9p] .bQ.ŮAvZ_<-׈TuvW*@ ɡqYvsr&cc[eSm%QܩtLkw/ _{fn B1Q=ucm<{H'qos{P,O[̓bӓc]Ȳ};(-|uP(?.h$8$rNEjuf1R .9DYVYc5,6PTXrx\p4DJrBˑED/7#V69$/gjĎF@9'IѮ{gVq4MϫRg7go~)i)Yv~}(81FPa *_C1r\g^nqX͛L{0rxڦ|̫߸ϷW!tP3[[|4,Ahx~B+o)z*o܇oyk]v4>tJ^7~>+KZЏbZ=|mbVwֻK欿^g-Y5\Up2"Բ4u,DT+9EF8p"+B޷7d 嬘]"}}˛/IEvzˢX>i)jR^1 bh 8 Iћzڮn|"\y͟ $=8ltY:7F/:GȁǣK& Vhz wWW4TՆ ehOBJ\55ۍ*сPZ!INU 6mjE!hh F+$K^v@ !iťR & ’yтJ>. \A [=5$иj!Ir8<5L5fR>YGs ;F0"+hk j[K|8c\i(I'cʐBȈ):%$I@vZN`5 -=\ǑHv`!Đ2~Q$億$zAYM|$_6 КiU'I,#5i*IJZES&A,cd X$xEQw㠚EZ ACnRg-a)AxiCT<R i~]Õ B]>CG!CߔNa#c/]cUމ } 9r"-(hJv,F g{?z=*%F9I:egyu3EMd2ٴL6OVܻu5,>^H,+@RHZc^@|3z~sCߝ-ͻY+\ڭ x5t1 ڃz=;ƣ`{L>Fxk|`gOVjEG[.NmyAЖ 5[f{b/zlX qڼOٻusrU^L3`+6i18fgHĊ#d7DhuԪ:ٰ* \n{+Ro19q5}f ߭4_F=eB0BjH!vO,O, {&K0Ѝ}ޘv((Q"/z;(dыZI'=/:x R|E6% meFEl>TyѢ[̢FƋ% J`[ EҞ%`iP%L,M1~ )F ! 4܈RPQ0ͳ]Kyhn'z%x6O\V*lssnN. RVfFH'wg ykeժ`F xD9Q;#IhϙSQ0V1Ę 4ȰBJJG_ˍVYA^O1=&HN spu"KjA^}[HpOI/UM"W VK"ף zHxF -%v : R&S b_(Ա~:r+ԵR61mH )WprA[[|ڲLŒ.[#ZsU_-/뻛[_V}WUb3) ,F5'qa Lٱ1u b<{?Jٛ+/b9+d?~|̊8}@}t緥PbFv^JkMd2ٴL6O6_ԂrV(g}fĒDwYnrh777ټծ"ǥݮ#jAC F| Z!Kn }y,h9$,l顬9 b2(~Z{ؙ!GݔMhi7sߥY?W`)Q5Ҩna=<NaqgO3LTBL0kYr")RA%Ut`aozME\gOR)*?2P\_OzyZ!GΕ}?zMIcK_Efʟyuo,T FsWdZ\iqEuWd9B_TijS+f\rG" !?'rSϏdV&^>ʡNzcFlީ~HI?~Γ6dc(aJ'st3bBdm.%jIΖv::ȓiX i V-^ bB5;(Ati{JЭ_x򺡷z/yvwuKCLaQ=lUFF4yKkKvW5zk^+_zٮ_jcoW qVij.nZ9)ErP嬽H2g7N?{WF `~݃n}vClvK`FT,f]ZXȈȈOkՋW\UVc?y-MlBLZًU22@eT@Af]JRΚ}PJgs*;1UȒHy&\>d&=ڋsIFS89YuÍ_b 3K̼Yf=3o cު Tgj4w_}9O\~R;ܶiYoZfO㘸M8$jqkk`xNV!Qe "l $kt\-}Z-&ZޭԤz>pvB'{ɗpt wsf5>j81F[A+48V\a;L5楖ktsBd,pux(̈́FZ쪙SͰdm>-;@FSvO>^#}8}2 A2%X(N%>)QN@c#96%J *[O4|v iRԼ(xx6{XoxzfFE~jMku\ӯ~]]ӗoolY6"}́'0GNH%j) JR݇O-uu}ɀQdxC.QͿHaw{p DYe>G ?fcٽPK#u1 /ww~Dp<i\ybrCr,[?z#s}N~t=yQo,b:?r 4$!_fɔCb":cԱn=J۩[~YDք|"!S1KiD Q-vmDD"%&ZVX:-o-V;[EөW h9j½#8Sbcun:וZlQD`-{7ײn4ҟ(\䳿hq"7zo"^fb<xP!r閸cO9ہ3C  '{D3lJk PywX~TOKBJTtPҜqmpbɻ#gAeX쯕&=o"/{8AhJh1ԌPUcgͪ}F۱{>lEPYgDFyyd>ܻcjDd(u`zLo=2AYf'sh;~--Ҍpy K( ^lL5\)H-#4/,vL XBVMH8,/s)ҟCy/L3'JTMg 5]lᮙ5xjQVj;EwEh)iC?]Z1,QHUv"}dgP)b c)ѢJE,C( ЊǕS${C֬&8e9g'$6+@zpT&P WثKD -_Q 1]7~]lTÇ:qA±R#1+ϡQaJڤ16Bw" _?j)h*XD/׉"^oɠS)ܢi^$E(0+ l1} QIp V¿.Djɕb3#Tؕ2Nv%cË0>('~(p³a>@ )Ɇ"tAf/I,3)2IȭO+lY7YM~iDv[:xQ]BMK*E)U-JZ)uNyVؗy1>o쌖vih)ocoA\1ABʦo$R'5sQg9_^G8v}>;HI $.P2".x `х &R£Hz}hͪ2)a-q{{O6pkzswBXW9> cQtg =c|Kvϫ} "8b1&Nչ\kD !ڌ}&qv$^$a\AR@4bZ5q)Hkp`)xp!{7̩F{Eu/{io㸿H<讗½=c|,Ԫ݅so)NyP>9&w`VK.N$\UnIeKzoޔZ)y RZ}Xʸ+""n= 9# >E!n'Z޼)]LXuA'/ ՠPRQ \"L0 JJWiS7X*l2D~DC >CvS:1u5\BRO VP_cuHp`AJ'5k҃iž!R*_q>. ]l:h):h;ЫѪ=#my_\Djc77[ύ]G4A4`Ādk^uZz]Z2-gX+dS"cԀ0r:Lt(Ks?Z%,tH6vYG֝K2 Rr58QJ,8gF[ ^1m1Ʒf"Ep@c/3êeΟr+~M+ J2O %ҔA1&Qamf D| R»P(~;k9@@jNtR >mYz; R6YD񩢸 N SJ1`K56cP`ɔ/M鳀Қ)AFj %yF99KjXd%12l`$B*mu7QDO%^Wls$WI!9p^Zi%!h naE%GDF;H rQΔ#Nk'HCwVB$uHn˒ܠA۲%5MauYc,R 5XwIg c!S@5 7jDRg7͗yX|`^YA0E$cǹuqgQC EB̆Maݑ G<C h$aN-A?gd \MQȫ8 O5\/oݙu|0{dW8'?K& ;(y0,?/P<9,Y؄Owd'T(DCRtGVF嵳ޭd4,5ƋoGӏyH!um )c3q[j(%Tbk1}nS3/xsh3sg39ɰ}r8ey-D=1RuB1ͅ F RjAԤ]VujU#23A׺kQo2 : {~v%nOniVwf͂ -PZ= tH5hN4&:^6;S^:ZCh{:BZe\ !ܤCdI1ds ȼ&tb¯VEHÝhaDw£,LW^0aUjöǸ0rУ<˧;>vG9j}m1˧;EvGPC=3˧;MvG!>5I裥y ч[CV!}BrqM0x*m+etz*8Oe:UH$o\^:MQkjNZ7Ni_킣z4G'h v铬QlAꕎe{>jR{,!l(*ݪxg.gr:I5F4366lL*ٳ;zr9OwآdɬF6[6h]s=irj[5$XFug39Mo\ɰKJa]B5׋8O` 3[4=X鏭I출ZOg BknM]~{";܈J%Ph,gߌ^oQJ6)p|bQz{|sސb}֏`JhVYQRAkQJ5 "/*ol>0wJ略T#Ip$Ylf۲RyP6IrV2<# ]Nbf1:h)`HͺP-l֬kr\%!vpsR>Ԉs݉E W:{h<}g9E:=c<.IyGvI۷KjANi{xx>m$:hXU?͜'&at7_>dV8 Aڊ2Z#nF`~)L­a8 X&85y<%2Am\ˮ{?QEUfS4%VӀDi6.q1g_/-[H{Xz_?~5jO>G޽}L)gww[D7" Dir7/h|ݛQOgXHđ0y0RP/F7luXN4X q  P;.Fw+xyL]qbߧ1oO^2E,rn?~;_Tqј OZC*V/po?4I%Nl%-zA*~s +hQuWxZ7W\OrVcoṊr% rnי  h4!͕zݩlzƉm^xEjm}}JW*,N5,O`'-hn0ShnG.`y.1&bW(dƈpãIBi[>,(udk}j=ǷN,kaU-b{k@km%`Uᭀ~htj ]Sخe0})'qX)X/]]t}$I&]DSݧi,pgIP?uޠ0wr# w8񧟻#y>ݑ|~7(Fp|F1']3hOhwG;'B},>"Bpku# FmFE(to*FDQƀ jXjIJUJ`F(rYrjk *ƸW{ t1Z]Oo R"[_./R7ej/N|ɺ\ݦ/m7Vʞ&lUx{ɾܦIj!l/GIY˨n벤._㓶 c\nГ˙sdՎqt!wo}w{I?.G%˯rgtI?s]']$ȫ]rו.#I ;RS*(AOwqCjT&kŇ{C8+pҠP7g#LS*(RA06]PGQ_>I#.KAK/$?G:#p#}UU0p$~[To;5ŀ 앏3`Gt`kb3!Ο˿_ZƙV/e6c?OхYo/YeL^$}Ӌ|FFeL&Ш6r`rΗwSԼtXT_,yYʑFBs ʠ s)TudΎ|(xɥjJt_I×"L$2Dd﵈N̙_ex1_eu7]u {L@09dIq"*3,֌42xm~l?9k6'SL*dPxCi„3e P̺c`]rubV5v_m1 '( cJCδTᙜ=gÏ{*rJzl5N ;*Y8鎉cx퓺C"Ju|5ir KJtқvbukȁtNej9ÊpHE^ʽUF T#go=vRYyQ/;8J^'^%JDB}Y4 IuLvh?ysh$NFci`ChH,Ƹ*cÝ{CQecvI;,(,cFc`wftJ붻?w]FtE! 1Vh4!>NLxGfv4fXpC֋}tMM>G>baSNc겛bc>Kϋ\@,.Z 蜎qJjn!@DXg₣-e:t+]- y y+mw^cs#sRُ8IFE!s(Fy2\ǃ+Yk` tHdIqrܥ?^?|pT&ϋh8ȋg#G# ׄ~O4`FN_Ilw mwv+% h$rq遠;3e/m$0x{atMtY#WmֈHi@q#nƢ|N';r"]ZH$fjGMTrZVhkNk6_ex7L!8˵Zev|)[ J9hf& rE;WTge$mT.L$*$cKH,k.^uklxSڷ~gPt c1{u*ռc孏),̇PcMD7bJ?s?{oVI0)s |ί[̸)r{Pۗ#LxM Q/]8:VNՏE؛IQ*UhxkA9K ڣp 5/:p&LHb8uUga*IO\oDȖgd:(.JDEIS]$-D'MőoOj#IƑ%@JiFFF;z{E{%7 g]y0E 'ٓyMĶN/ջAooZ pPDiolY3S|#*4Dd@Gߣ[ Gގ9!k;F @- r6!AL(0YMiR&ދ [|$/vX!%иB8YAgquϭ=[_,!7EH;Jrڞ! Vvtxx'.6|ˍ^$ԼD4-u>;}xQq=./ Ͳ8ξWO/4FqiLt&2]%nNךMR8CN;؜y`˙;2o!*J'[GK# 8YU ׻/t'ˇӈ>˼|JJ` ,;6H4(֍r(sLXp~jN7^[Y-U:Kz :9ay{>78@24)p1 cFCHRgzaǫ*9R-;wtBZv%"[Ń`̿k/~7 3ޒ DӕJtET-:1di^'0*o"qz^Cc?oaXO+v^O9DÕ"b ]a7M,#嘟w zqOz%؊_/X_e+v!-L_g SL;m8|sT(0&^lA̰$N"d(ӐrN8ٛ 5TL<w;^vEwAR7"H0]odoNBgFN 2cr3ژḅ3\)Γ^)RɒB2Y¨HHrI2C q09++%RuRI5m̼kC!Ag) hݪǜ!ttGH90f̤532҄T VIk̶Bv{gτZ SmiS~δgD##S<ъƬN8$wNW!Go *伎wÉeǡt>ګ+! ?UڄncA$3: ]vԼ bȾO8;)6d_ z!o .JצKĐvtٵcPŤo}U^!1gWg\갫pP >B Xd\Oߺw w@zE :'yLuBh`^ā*t1XcHS&"2Y?:/>;sM  dzyt411ES3rG>=K AZ yyCKA>K[㹳Y:$UD4qyOd߶V tXN"cPvXqҐ cu33y:=NkLW~_I\<~mYwOO{1+܅Wco .eܒK;Sc|#^D?}ϸ +/݌n#Ʌ .w"q?;7Wʓv?(2WhSXI.U#4HK`J~t^ġm1(ӌ?=! 7!= P\~rq Lg?Ĺ5GYpØAx2㔖%W\cizg"4I6+hAܑ`[Wˏ}< [e*o 1wm:+|bUt2}lu5]7AVXb~*_Z~iz$˕6U$Zf1 _ȯ0*gi*1L{O'"Cen.IH"PN !יF+) 5ȀNuJ/,_.>.v9ɬ̮/ݳJ]\}чOtey.bϚBfXh_'IǯbIyd?ว=ĝ[;Qw}|xZOv|՟>{^Ir'"NEޕtzLYsYN-<ʜjjmLYaR,I)aMɐ9[)K) Y"U nH*iJ[(O5D1y.]m#\ L ֦z2ʗ9'$N,]d7^bBq/&?D-rɡR}ҚT.⒨+ЩXf@bu5T!hiErHj:JP@2V֚m_Jk}K0+P܈VVIޣ%<[-wUΖ*g]%c&VZ9ڊcK#6 L s3i5/.//j۾a>>>^CʡW+w裻>@jfS#Uv?#)/hQXK@br$ׂS EGۗ}ˆ<ⶼGv1K+ 3y3e_ omDX@v9^L.@*<bNcQܾL}.S&%\c!/djՐYP & AJ!7bU39\3U <. < x/'=4'_/jRcA;Y 4Fj\LQ  NR1 dъrsUْÙ(tQaVkCk aTRT"]YMK8"l"e/9XbY/V')-R6|gQFseaF^X%8^%L,& ?lZls`aARUbhԔR G5?nO7w?n`淵K9U=j^?=O#ߝZ}krA!߼E+~/u!M M~(bwOww`" B_-?s_+yXt ;d0-Ji_1p~ k Yä,so~o2w{ۻe;z$"b@˭&n xŌf4q+&b 냛AC5iXqT⼀Rí#sD1. Z] A@En +P(TnL:kxrzhvto?>w})@Hܗ>C &Q*Ro[np}Eady?/7'Xծ҄}B{/ 5IEeS "˝;t~טc-S"PzYb|#?`*Dӷ|Ja;O;#HNv?Oc%Hޗ̄F8SŔk1|(\~{ /Dϩ/^-ev#g̛.N[m)\x=@6mavu +RX8mA/heF/|>R=wHNSG j,7#Qk,6S+genHЕz;x!n# K'χjq!C6in?Ȩ|a8vu>˦Edt| k <ԣ5m`+^XcȐ+ɿVxv7 nA?hd|ib`*>-?]:c1)$4Ժ, RXÐVM΍q @ʀJ,=lb@RQbh?x__t2@f5I`¡7)C9V0]8V0U΅9)T֖E*GҊ&єPNH $T֌h}~C6.p6#Qs7qV\jF\nٸcӗ*ݥDcS9N]$cD75.LS"uwGKlv4~:Hu{S׷?JJ'ȔRեf<:ŻݸxvSwGPJ`P$Ω JpE`X*(SkG l0y ؾ[횢9Gbh[ƒZT*$%A9-VaLJaaiYIf q{ 9׶ !zB !GQpS߱%E%3[BlFcLv>z|w%u Ǧ/cwu=i9r?e|\2X6UUɭ"AxSPKc("z+\q ?0Żb21(nVrn+s]F0{7{;{=n닫OZi^Ww?xJ]QǶ_C"~kTAf SqoQ-4'[› CT38+GИ ?Hle,͞gl{ ɠ}Qw̔'[ӷiY1IA`*YZ]w#!C] Z\_X9^z_lˈǶvbZ_a唍V읣[OA0bhm{'YHmSl<X_2#>(hj~ R9{~>VːrQRCbfHyZ=;SΖBV흻P Zs bmD{Σ'X~;2L!a=W_n\𼝄^פC,R_jhRFOfRtx2)|)1HFX{br(ϓ>:>{4Us:@P-UfDױ?PN&1^s4~9'xKVoͲD*u|X%85vZG-ru -?.2ט0cr#'1<=D*Bz3+vxt7#.vrzG1v*7aSbU2cUNA)l|R;$ΙDDaUQ)SG&/Esv[V*ǙtbY+-{ n˭qkD`y+199KJf1Dݢ 7;ow/`滐fa1"C97-8V~oT'Ab܉H 0eJػ)`FlMHaR1WF)ҹ HعVAU)F?Qk '< {HZk>/a~q'򖼰_F^rH)\S s:六]~ܲ;~0W$Y-J jC29a-yvdCί߿MovVasZ_Mf'K̴Ec7jsS6wC= EB(Uꟊn!k9K N $FnUOf5 rXJx9e|R.0ds.@pA"]i7}T]|A'鯮ay{9f(U܋6Nj 4. ku8Zf8G:^TڥdC%o g#_/VI24X:Ei3F`KYUL|^2Ѣ1{$IT6v e\t1g|Ĝ0D@pz/ffUׅrr=ߝab˹;;>/_mj6;?XIrQvLToޒHvcP,CV.J.o[EfORCCQi)20aD1-.b@hw _- yS4dNN RKW# jgL\ =8ƪd߬Tt]{P??7zα>U<s$1tp84"H"1&"Zz(9L`B\m *Ռl, ,gW?aE(z XfzՉ\N2hA[#!#}<εgY<W3 Z2C,Za9Uu!42g;lJikoՀ=Ɨ7/;\*Xqr#j3|j4i>/wۘua6^61V񩸻5%w{(sd M tͻT)i5!w!b*^_SӪ1H5 &}>'C\i_snʫ+YNS^J[ЁJ?B&JQwc{N PDznIM7tWVǟHޭWOňJ2vї}[GYVR{IpWƈ͞aXa,t6TM;PvaQxT7>I [@W-2ulm;t[)ywv$<$oڷ{>FNpNƯIϝ0mS"&$A^}I!9NҌ4 d_̝Ȯ?@$`ZNhr)yc'NAPk{vA1Wh9qm wcj0ja&++% *mE8"UH[+,52Ǩ"45:IҩBNr9ZjA})ŗY<9J}v)Hv[ \C4V7Ph3ԍ qJ*WZktڡZI bXJkCKջE=͙&}Z{:WuXZpoUt)9gD}o6AjV2WgWY#&+J3 zS* j}|:+9&.#EKҰg]qcc &(/A BS h;@l1O,pCu '3DSh % 9ù*fyQT6Fk7*^@$dlXG"!#C{n邎c1PNjw #A/r9d\8M>Z^@!j7ミEW{Uq+)(bV"D5q \YUUa*idE|ѓ#b kQьLjr̥@St^gMߟ%bZa -r+-@DLiDTJ;GWe%޾ ԯDk?{6=gk%xiI6Ab-]X[v$;i6mwɖlech," h8-YFE 7A 皼m_(ae%"+2J kRm XPڜq+53Ӫ /ee*53&aA0jf}o{7 ز Ju&?I/~-|!miߕnLsym_ 9݉W &&%[Wq%{kruby @Ig|9U@y-f *2qu}o[-F7~ֆA*uyU)PR5WBLU9y0x/w&"׶ !pUmu0ʎWQ9_\89dF3Ma"4%DBsJ[4o끴\aX n'{ jlf=}9SQk6>9AG=f  \db:y|3ÿ7lhnO-=m`9s/%}!˴[:!#6%lC"ɉyؚIQz*#ZyxH<ֶr.w=@D:tD] Zd *;~ؒ1k@^w]v#tcxb0AMJ9+zsY~ԓPx݉7'*!ossNA'2B'0Lh Yv3e2 hTF\p+ ,A7ok}Wtkœ-cTFkژd؎Ɂr}P4Rr1h[򣟡 0#leBsԏnhsemXE5W\;px4}Fqo)Ril`wCm# 6zg4"c?ltZҟکdz36ϑ8Fk6w@C [GmKsj~uOijq H/={t?h6.)kmugw߭y'wgo>]}hnaLpZ--F-EZNÅZ(d(QK~^i:p3[B}&ܱ[.ez-7gq@V݈x׏7'Wg?]|a>]{qw>}>xrKǣ!/_=>G^-狣n?~!r束 #/l~];gP*"m˿Sr<xV{h[ .ӿ#qVNYP6V՜U= =6ژF>=уqv)a80C=&4E`N0N 9RoAGhGߓΓӎepw>A1c]pO%Z.myjy@ ꀢG밨4Lx>zr&}|qn'ߠ7ۭtc N?'{7_>w;n#ԭwPWe,GvqiXᛪAwЃaEB@N׉qaӽA9{w }g<p_0tAMU4z`Yš9CQ_?aC?,\_M6AS2x\x#<+ p骀>$4c)[49rA&5laŮS\`A.nM{ y_Q *^$GgL7Sn2يo&#[p|xvUV,nD!) ڨ2/Aq^HsZcs q_I@E7MYn2mF~ ot3kn'PG*Ƨ56ǯ~eLG.Li^8$lIQ.kL mp5ju@w؁FL)[ȩ9Eη94lMN蘘\|".u^'3rdsߥ]lr#y0%9vn|Q8<%yY$hIMM,; );nKKEjݠ׊~IzXԓ|gUz*nZL2w8l{}`f"8*D[ژ d °t>7~7H{:]}wyշ>̇33l0bl}U]2Qț;yĝ9Q<1z|JkV=X^Oi`J&aMJuHSpl@6Q׆?=`ƿՎR!A\E*`Aݵ%pGFMh4 \&і2pU,n`!{M!!MIקt(ɵ׾/'Gki))^$te{mm+GxQJ5 'j&&,Ԧ4"R17i.\y@;Ε3/|K#c#نŢR/#q$LܘqB ̶j=nq881m?~ܺ٭ز fpV-:ˠOy] ^7mV&ń";[e~fUeҤ~.OZu¢g)dKo~N XP5'{^ 7 ~\ ;4 }d.O*MQ6n=9`T/ D ١ysͱq2q2[ m]s<"rORչA= .`dp,t6TQB(ƸJ0^j>W S\`l5xkKNVٿ9޼)/>:Fqbr],UPl+N XG;'^M>twGSİi^HI~L1Sn\Nv%we(BhlpPq>NhQ# I/(upC5S#4KTr #g3Kn5džIܨl6[vv<[vv2[ AYY>C̊d`8-OpL4co\V(C}|x}f8Il T.%jݮ [wӳY"/LeÕ%*Ơ=}zr_q;s{XWrHtٍ"è)YDg6n[,O qGnEH+эd*u L`_|#Xu~j`~Mد ōNH:6~zNyb_HĈaY sG6Mp`K.-Qa3DXX?dӠ2*SᦑFtgwaUУR ?6UN"2M@hv$KJC31Kh]8N):Ux(u鄖,r&y#H?@Œ< v m f,W$!iho6, 5_WGaGQ~Fl)ܼ_ ͏^zźΗ(Qހp=(~k̋<g 8嵡$ѧJR0W&fP̢Ub 0&M(#X 1UFh#@vs 9@|9&7 `'1d/ Xծu 1R{[=%k r*2 $Aە ކA%WT#3G&d* .(g8!8D )BX1h(ޚo06 ] e TPmMlۉjm(͂J) *ArI SJNJ#ژ5Z1h+eTj mPjަApvZ PRgKد:%s` Φ4 nD:ФdĬmcL8| F4<- Y|T 8VAr"pm$UTHw> ty\jTa͡ά=0ԊQ1'6l35CtOƸ9*ns*QsTKj'gK]˥  > 9GN wB`fX%ʢLG 82ĢpMB$%aĶCZa5 }4n$)\cEYp5 T=| TGZ;9ͭ}ec)n.@P E3TX<_G8_qڪ̚=z©"ͪǽaz'{Fr90T|F"7JȦz'B0ZGn!W,lΩ@ ယY)5bnIN9Re=%|J순'\JhY4uFhN*o^H,&(h|W%r V$@z`сVJMI'Z$#ϹBTc֌fM\j@6 ]`̰:|n-7(B5zcn?K Uh>PC؟(!N|=5֑cA6H.5ԱOR]|)1-ۜ 1NI0 Դ8P@o.Vm9n oӠH*&83Ncn=4 ܞ1@sizz$=ӘOvY9֌cØR݃7"ڒw9qkwZmEGW%٘U-J@*{|R+Կ~po/{7(p{v 7!}>^eqm2d=\k1x(W7ڣ'dQqD$(Hآ!IQD2GÛR-*=[mԛf?M͘hkW(X K@W;;eNV`e\lU NBmq.Ok%'XT+c)qA{ 3!%HL G3oORt:.w]])qD1SB]d`<+]S{^ `(B`9˜碭VI*N Z`2=/f4B[/_B ]_nE%No3tO>%e"/`njF S8%a۟EFlֽH>IRvzj ׅhҎ8H̋U j%`AyVV Hv9W"PNAmiSg949C{d8<] Cvįz4ဍ2m]XB6Qݤ[1n&_fmy-{oUU"p]9nBEDTXbQya6&dङ\rf@1\'i>TLqBx7}C<ֺTL:S@8 i2Zd5DI 5RiR|ql6/2S}.aX tnh?쯞f{j<$Mu!|.=\믝hݍ{nѢos, :68ڛ~2[t[C~v{?Wv2%1S^K.իQ.zg;g2xk.w%C|FCFú?{Vp;m[ v%؟Ur:y@ S +EČ [9ꘓ@z>#M61n%7ңxICG{fQd[թх2:ۨ[q:,I~#N;VjH}yd"q2d uO吊ߌ1 _W&*6}jOUBV oρϷͅG;6fgm qMd߈[j耲؎'5$IJT<;|2x `81Է`+v9BCzwXr |a?V }M+?oiK*()PҲiK:&qqryB\=eX雋וux'-5S] "id'V><0λô55޻ɕ' Tr=T m $DZ $榴j1\ TTݬWS%خ {ےH*b~2~)Em E.}RA[ BO&K<UF?YN|";f'C KN_o(:e7N@Q(5TY@VvVDqvчOmn}ՆoA~ŻF~n=Yw,{^_~ŏ6,?Fva'd'Z(hpۻE)mIk6aD_a&IVƁ(ﻑpۈXhbRM\BT/E!͈懓G7oiT#i_#Q}Ї.H#]=ݶ6吔g}kHMAykNt4ٚ ( [irJe#I)O%.X-3c =2G1k%HM{KKWz1/]b Od:SۣHRHJ˦6UR\&%?i/+G\΃"0PE]r(1({)\sie DnYAumJ/il UOئQL8=B򐲘FQѽ Xz ;͖4:z)k MZeSILv6k4ziK}}߷ժC*%H3Ybu-5 ^؜ƯL1X8/nMߌnl*(ҷ*egt0rA'׻c>@1]v('-?r<9 6\ r~wu՚+B3Ǜ{;tb@,~x7w|Dg74OZA %%^ps')TZhtaPq^KY,Ui}khP'-igÇj|S_C"]qVP򈮪2"#Du(k.` s:&#Kxrκ18 ̉ cF5؆gH1xŸK\-)#}ldWdsqEIA$1JhKI **h> C?dc$0[8\+Z[P:XlNf8iPsK_;SjP>F!(}겕I|:W.totRI]iZ@t >l&n-> Pa3bqv=E+mZ ӝ-)tv KmBX2+@Li/a6!-kH +AFczM |نw/Pt{TWJkwSb{S&Hā }Sfggv_ʟD'̉hu\hH榃::dT7k*%BvU^K 4~qKBty"Ud,!"3H&u7pRWJdWmX_8%ݞ`c:%#7ddI%^$?^^ =$Һ!:'il?g6pm3SF41l 3ISݗ ƻّd![ Ԋ;O#n&Y®>Fbf0y1wuVڗZFNGK_ 79]$ތ;;I2?j|tb{t(~NO.CW. 1J3fj6,7 *'4"_P, 'f=roܠV 9ŲVzc3G;Xkg=1#(KKZ::;Mww?*b-r>Br)t+ kZ.IbV Z$ƞsgxUj\lm zRE_/E8\kFI/AK)=茁pUY; 6?ͥ}| nk+bSdKUWR}x _OI蹐axcE窅P:QRr]w:I9_v|#Fi/{N/ tmH)(ΔIaHHٝӒ/Po=~ݔOփy7 JRiraM"GaAoMxG%D۳4t: 0~5H$OHaʜfW%]H.]ΐ@< GFr*YKʔDɖ*>cfNL")#Fη3Lbfsib Obfۢ$h2YY$ \5ޏˣ$Z6IIz?@[TQy}*pUDCKZT!x=ij6ώf޻.KZaKѵz.0k"HPѵP.՘/eUKG|!`ZkrTZ(G2xYMѠ/p]){a66]]sN0aXBo +_z0%g!Qv K0iq.8F*.XԴ<]`|"/*eX&:ϬМk ?9w_ݺWO~U2xWO0x Z=Vi0Ty-Um|6 ^$^\ jϽ $./۫G^0`<̗5 ”F_6Q)Z ?ϡp(mC:)sg#S$8 :q "m%1QFkgHkBR.@zA~o_gHϞmAx =}uv~ _Gc20h4,!'HA)ό,g8;hPn˰Yr暋)3@Fss5gr- ^KbEJm)kr}&~o0NAq7*/}e/pvySܢScX\r&1ZQ҂eRDSK p"W`p_k i6* B9 !TRG{;p%?{d P;}]R& 7B9ch"2< ms}c%e7suoPETB 8!>dz瘥,F͸1(WVJiuȁ?){^?{!a䷃=^qXۼ|37J/rxW~jKHC+q,¾e9iH0>zSK]s6~H PRjHSJwb Akbm HZ|>@1P<5WjP!p{p3)]DP oUY%2q#OA:pB(qG"!0aQ>sˢTycU "wLJhICu}hm%%F,#32!9m D+#B㲘P!'>wX>g',woCY_FRD3@fg& )喂'8k&chԠ4h喁v̙蹷*(DnRn/%tL/`h:wsBeJkht4zG~\DB%0,mm{"<̊0+ìW5^qJMƁz3⃨0R{4lxw|%Gyxkv}>ĄsGg~ĉɌlVL'p4~F\k0|~wOr Эjޝr;}iEA.4R{O8ೊ{-0%ڐ+3*$OH?&?55i >`}F '[-sa"Zo N9@V*k+4"g"bkX Gs}_$o pK@}8#UEv4~'fAfg.l".{ن8|Ezfg?!'Y'9TY Zu#s j$NRDcDIpqU;n͐Ä'&x}Q>͗ITNo,$7',l$;Vӏ7c͛pRw[B"ƻmBkNTí(]XasG;qvhڝ9'Ѷ%&/0I$1}[0)F adc= K}DJ$k+ዝS3vcsAIq.Se= m'h0aidɡ]#J='3ծkJNIb8jX1Nm J 4GYa";Ntc)LqRKd 'meL&Yz>pA$N;|<)Hv>cs$R)QG-py^7h&??X̻bM3b+zz#9˨ ߍ|x@Pl{[P nCƋBb&Ga#9IdƘ3Tz否%kl-C{ . -%B,,\3o[[#sʋ:6i"ĉJ*s+̃/FMZAAl*g|B]0,s%'DqE9ōeh8m$v DM.QŒ Sg}|1M*-5:N c6`H],:(W@0 YJEX25Uc 1j(*+ g׸/=g"JƝvsTIPL %7UANYlU6^lU6^Ųy lֻS3Q,#TsK9nƹ3P^9l*+[p;YR(" \ 8J"~#=U7>6>&<+ 9'׍Q9?<&,UlzXQ SC`vkY)z\X/3CTx-W=s\(=EMrg<Ι\A X&y{Xr#[hR Ip+[F.8Ud$JE @TmAӏs!T =WMKYMR pU(E19s Tc{2iΝseԺ%<'\]'ra3bp{&s/&OF_2Wk^ƃr TCN^jLac6 ̑4c؀2" ,\VcO"&_D>!L{_\Oww]~LȌ^_ތ> =Cz~`nzSw[l}ct`0ڟ)uN nXjL,5]x%%#CBZ2zs A! ՋKCiNO\F.o[D/θp9{A>V:x1c1'^ܣn'Q'V~Vجng63;_8ٚ>P DM ͩPJf193c 6 К-sm[zvH -فnJD5991BHDy_(q3;z .VzNhLO1ƻ2DtXSje>+j._? ܂gr>hEL .9Nj16 f ]Ci S3زMcMGLqn<e{L<b ^n:Q-OJnUB&3w@~`6|5IQ|\g?x= |ג{Ncv*ѫHV::8v%51#L4D%gvE W+!zeJPumOmE$qK7j)G%7Tp+l@ZB*$ѡҴ/ˆVQFR\ID ئǃLzş=IOCJ=`1p<6B14G]iHȈ8€JL&X$*ADIjMcv:WٔU񌓟\ 9mtN "[Ɉ]d``@$"i&Y WkA:HsU70^V}6Y& N#Dn:K-αGgR01&`#2KWi辥ܠ0.y\i-l1El9Qa%~$V;I@W<ӐGָkoz4fwfzͻ @5ysEctkXOxp74ۺGo =C&g?4.91+ $̞N~pU{o? /+77ssW!c& 3Ya\M.PiQtی4d]{p*˄DYTxCd*{l=? 1"Cr%Y%MnQ]EU5Q%LvCWO.#Gpgqҿ>W^VGEm8;JQno/Ocd[D09D^5(dm.Wոʭ2L^ QϼoZ`7M> zT8xͷt "8{ap32Ǚ dXTqDC~O>Dw.{4~sЅ&~v/Q949ήbV^@/y0gqqg`HQn3k$8 HrcaZ\C[ 87dI1Lsܒ\S"$xsb1bu9-&w=Y.*X{{+ ݪ#Xy[cwgᖸu^=IPŔ )IbiEAp 0{ 0RS)[.N]q-jq >٪?~~Axf1GT.ll"11.=JW.;8+?y+{Q0~~0`r1_G3.5jfjDvߙYSK,6¤-7DDHO_ݾ Ɣqo~|qn'@yi깛`143k6a=Cۄ=7a_/o2 ]£'I^΁3yݎBR0,zQIִZn&Gr{%UrR >CּBlE &͌cd +(eܚ!(8c%;M׼wgߞki!VtSɔX駥\yX RaVt 8bRP`NT{wͱ gϞ_՛BPzq?k{d/k֙ B1sxT8$DDr$H9LKXnqDjL}oi0>!$/4Ojb%3V"B? 2Ud"nsNp(f*VY AH@޹-\*`6zc!%h~76/2dA}Q@:\$Z*N`jg; $wpKgz<3^쾫ߓeЬoߓQH|2|h }Vܧ ^5Lo!b0 #^4Јo-(Q{2nC4kU䌡-'\ ju Ef&# tGĚpPPtآT@o4H8\kk(DwfVA s:Y,n PxߢӈC:w8hBVhk=|͠WD#Dyl6?RϨ);_\z}ȅTFS3 i9s6ۙlY}eN 2DVWbe99W)n\͸CaE.7cmAi7rg(uٖ<-z 9c_Wt²@)uNG2[%uE02RHa2Too m:Q=ƫIb ÖH\ӷ@Ú"r?q)qg+^U H(TFǿR4jCsw% QvU$b.Wzh i$$ p҈qag)QwNT_>QdwoJ$x/6=+>4ؠv:$6XK8 ٖ8ppN j zx܍?8Ob㹌U?). ͹KxS/;}K!9l2ccflzJq,;=2\,lid ֊ R1C?xr+L$⃉AGjp W aSy4c 8 x!"ڪ~wq͟ak )*1xrj( u,9XZXRǂ !,g%Z"?j65CJK]7 ={4C4y4;& kp>&kξNzjϚ"/H!ҊnIg+zer!,^"gEmGW=8۬2qp݋Xa}q٢9+Wrj ¶=rC\*徭w'zݍΐڟ_Y :~l-?4rnukd*F_?\|AIK6A+ެJVL'yB#("^!ڛV_^0oMx'hlI1f|6Z .LS͒%n;m,BJg *bAtrSH8S8t>pYCׂJ t*ݿ@%YN V޴QN#fh7A k@Fii" :NhbKcBh΅45hJPJgǩ|I⤥*a§&f6 56q")JL8B*)P"JӆC%%Tp ]P(XF hBc ڑɍ Kϭ c`!.2{L4J"5 MĤ`VoXpD VI왶 T{$x31K0mLC05 c`E16A!C YHc[UliKSYn)!X,s|")0{l% Ǹt͎2p]ܺ|z">99?ܯ 0k^ "HcZR7woOp,1/F!e4hӠ4B?L'J?Mgzcތ]( 6߉+Mγk'ƘgL-ki.)T2 lQ?#)%wi%rN*E/CwqXkwgV?w|^QnvI7wq&} vR{s[:&P6Y kǝv|J5Lm\uhC $=|A:9p'wt/k-a5GW#v M#vJ}{ˋV78#N{xsҀ ^cR]E4[^磯w5JA%@'r_{?Aͣ 0y?卿p_OxÏ/< 9][*nYZ-ToZ\&N[\ܶlq-[0$2|U%D+a-!VU`ebgDWV v(;/qLZ7Y oh}7Cen_ R[Q3h735/7.ӗyjNC\駨w3w]~&ϧɧ76}վng^AwӻO;Qsrr=\dr_o|YX jn9QXYg&:šYA] %!+p?iE,w cs)GeSח Mś1 /a;ꜹNtseٴI2 'LU5M \o[0z hf5npP^ߒf+*xj'ܘ2{'x?HsZx8ULMVE67͗=.u^JW%gfiԏbǀ&Mo_zd^zjwG_=5YS|qeWuzaMƁ3+w:OӪSuQ}uRhtP 0|ާ4O4P [iɴ3DU*kPƦuo0rݷ>7?Oc}xqK3LWȾʃ#:.m=yCN_JKp˅QXy a[Y[x'? 6̔,GH~!}S4`7RSą:jM׀?>|lСܓs[ڧaew銅 ;7ry8Bp,^Zy"di9F]E9-=ZE%Хt)ęZ%X#nDY "kBʯuXkJK`&EH(lh%kE 6Rrh8@Lg-BOgDi{(xFNC&HxN;.l3gi zQ<I 6ANUy{瀡}o;Ŧ0NGܞ<^X"/Kj=S 0᳷.u+nقw&3YeH22.:vwVTGm/6k98Մl <9&?ħi[+擢TUZ(s- >|1t[(8oᥔ{ˬRoczIR+-ŤeeQzUVҚ:qH's |ex$ڢips^bUǭo4T{ZLI8oKkz`Joνw%h{ &A`q]w*w_8Sζ@lxOA=b?Ph` uYFzZ |5ZLD.^JWjx V40[vonڷ xqH|p0T);^oEcLKK6cMcM<6rL[\FcoܚK5+snͪ~niMr-qkRFMMbRЪ9DWթ.¾p=+sO#iPj-GߝwHJ7lm55m̷^ҭ߉ٻ6rdW]%@Ivqpv< D|$9`[ܶZvw%Ś r[Wd8$ޣQ̓S $|#1z %"!"!<6H̦Ot.P&qy._̐Z(*v v<U1R%ѩSZa%Q+u"o\ qֈ+9dު~0\ڤ@fALNoԙ4Bi$iS0,"Wb `7A Z " eYV95`SU ,f":E*\ZRQGGjd%kB2Rx|yfzolw:NWtZpwmP\۟m_ӿ;}Ig)Bv E:K͐dJ7E$݇c HK,;xӛ7CvKJk0ڌ>pgI*swClz5]+x>^MiIHbݯD*1+6 ~Ku'*CjƩ[W WZ8S+{P8ٻ?)RN>MUTH{* jCt%|mzVF wJ(;{ sAUhѲ{Pcge1te=O/[ EupZ{g`S\=W< u,7u/_Ά8NUl,IX'2 V G81(\ie!V ͵&OΨ.:U|<ک .X0ƂD0&*ֿ4iR횟vHk>ZMB.Z<4{;'yn!ne7Zo R (T+ ba6^D >>ZZh7hx?tD.SӢRD ;qPbDE;-c~H'x{D)FpC.Txa9\sQm{jZ.z!P}3;0i.+m.>VYl<[)^VB= {qX/F_m4yƝyttE0SA$|m΢͉Dp:CYKtՄ`=I ZXz?Ff)]udL盛J*2>r4>}SL ^|}} uau63P՘Ey!+PqlPL4i`4yN5sCXz]8yU*W)@Sf!3 `}> (}*Ą=fwJ=VasD?.kv *A4V: ݕHc"J[_XoI, #ejB̕(&(-"#C}gm8+I%yBS6 \F*Jb97S!Hݖ7̞ɂO1a˞6YAqjMrxTh({2(D4^jF, 'IwZߖ+0l\2݃!dE* m8*ab :pYd#u&Q8!06(%Q•H[h@x"vtcg ASZ L&Vz/`Z"k @Qz c pgIeW{s1ю\oW @ibQre~Y߯鉿Yg:Ğ{+9֌kMm~}gh)l5xG^x5DndWM~3pխ2fӫsIN~5_1,AMJ|XK:250GOp&t*>4'J~M dV7dJEӿn1-|~s=zC)/ªz3BSɃU=Z\ wn{\g)efw$Z>F3YcJc`UHOia&0^2rzZNkB3 3Cer3roGz1&(㡿٭&}yD TXׅdj,Fk &j#4ޒ{JJje";]rKb]R8$sRjL ISQXQ0eyaEHg%T9&ߪ-4z| cRom=!KGo 9@<,:"jQ8`r[Hr;`֌2Akw$M2zOקEBw!XMoCQFHz7f=^?=.F5fsڊ2w4 AL#bƔoz>٦G)- a@٦F'GejHb{1 )4p25<,42N0Y V$EʋHsʧrcwF2VU7$aYf G7d<LD3=2*`g0 ѠWL%Ѐ&(10L+1a]FBC<>sx)T-sؓ9S\&,US[wV٤B2YA 52t c'9A=QI4 2L0wyZy}-(1rX,1jQ*V v (f@V++QLՁ mam(dHh +7†k B!Men)OdcÜ3OSemǚF%BJ(˲yjס:ix#bfBJ:`˶yQyۭ@8<8hU^EjC OpOjN s15gٓR#⽲ nڈ̍%>,EBJG y%Vz 0xc)ăL:kCfr+{kLsVl HRZ0f=@PAd0ޚ;6i+Ӿ|/Ff0o6 kn NyX8őU';ᘑ|۲_n(e;fw \2X4U e 0X5&y4zC*S[&"2.V>EHd 䞈?y|zsyB~&C!뛏\h8V%OɽYնxJ-I֏r"M=~r9Eh/oQE8{TC *{}]UxרrUƞUJ]j] =_|jrV;%ݿ9 { q*? }ϸWd!+8ӛ%'F9q3dWnϨQ N Mʌ4 >PIZEm:S܉{̅ŪH}}cgSW,,Y:ZŕeJ |Ab>Z&v%#j .ŮP2Kz5]`JAR3\b9Y,?7?HO;JQ'SK΢cIurf"enGiݱFJ_v̊zF39cat6yۺ y$Ϲ4:hc*u-l`CyC4kLUnY=\0,i?}F;؋1!NkiMs&H!Ǫ6Osdvǁ`9N mU@8%F]&i㠋:~A84U&ét:-S4zQU u,DCXz]ONwZw-@gOkC{5p.pj8l;WX 5OPK5(D3W'o5Rh*ܬb"*BKi J~j&;A-etR9_}s4,N7UoiÁXI # Jޡ«R KMAZ  )a #J&'ZBW$N1S qV(3Qi;p>!sF,&!: qP43tYJEVeBkXb@{" V;4/8'elJw]lB#(#! ,Q @ .G=]UV&] #Օ66hS#=^]ïj݁avAXxdzQXP9vPkQ*K&, "fo8Z'7zT 8XU#ZȝtbH Lʦ|Yܑ3KTU?M+_K%,Ey7X>{XzK41:>Cg9{/cЃ]O6?&/9޵q$ЗCK"悽qH8߯CICryrh0lT=/dW2#^` xH"k7/"XQВ"^YHLk g/Bae`Z+l Na &V9f9 ʘ>(X<6:8Audx91Cf{m&F{6aga$Ѿ Er=.I<;X^寏?d.V"}L(|7/ G3M.׏ 'Q/nxŨҭ_^/w´9Bƞ~ dqdaW̠09%jF#w2 V'Dז)$|7?svULH*NjI[#t* 1S:Ih( #84vI"Io@1yG+ h-Mqn?6[=+&pͻMݙRQnU>~OZ90VN{$\xRESea_D&2eZ@lQW96uΌC?Kgmɔ'(%c^A9dn-2^h a 5NiYkH1B9` f)€YIl"B[>{nnϜxaڻ!;h; xNMTE+^ ݢftC?Q4gܢSp_+UX_G{nA[2z߄BU*qB l,~~*J>Ck}P/1ZpSU+uB:I8FRP6Oi/30C_s5&t[ TMt5Zb i72Z'[޸mM~k?kE4H=$O*7d $ןZ,S꿖ݰ ֚1ׅ޷4.Y)n@Q_aK ۝r?&. g rv(՚D!Df nMzvq|kA{3[;a7qn| k܇֌.i3D9"]-m$oi- Nx~%-)FXɃ/6 Ӵ_Zqك?4c̖_cy4!wWw6;ϟ=^N_pQvool<&svGkKWpi,a3$ZfXXÁ,ꗽsj}y^buڈJU|u ,&cGQi$dGIgR>w==o>òj^s }Ph*@䨮H"!'P^ï'B>t4 [ ".̦{.5|K3xKwoQ;G4ãN.y|()޿TSv&-`AggP$1c+8U2.R'DaI4wK $5Yp"wت ­VLL[}qYizi5r^H4O}3*Eḯ~6F[ |:P5FP N*A '(!6$$,Q\3V\@ ƛ ʐݷ63-SF(! ֆ{nS7D+Z*AS/^t؄IKeiq/O݄/f@ |ָ 5:q,!"aq; 4rc$9;5O8|xfRbNIF3D;N D%)D+iEk!> Q aPƿ6 Kr\JэV?|,*Bi~]١ICa%݅.yO6+$j/K T܇{wҦ3nG_`Ö<a!gĪʰ 33~Œ+ x@l9YO^XZ y4YH(*, 8Ew.](BÙj?}ēcБG$IPwqK&R;͗5-2,r$rj:Ɋ,&4x5 *t5[<^!tNaDfY=ة%Zjř8cM`56&t"A\"*X{ul-Œ`b!?NԹ}_h@# CMJ|KUaC1eC%*W1ڞqv)r*0F|Ɯ e>i"۞C}4-dp9h~i#CbTɺzbդ9h mEPfȩaj$OAޅG~5"mys3\Ţ¼cۮ̮I-o.}sh|xu%FXMB_.F܍h׫)`5L)xqܡl ^`&qF6Vj7I[I܈11D ft_0l' F;6Cc"1HqZ og PhyPL + V$NyE%hqml~.PScQ]ס6lnM[Pܓ(b-X߶`J1F/Y$33qysy7^e~_VoRVijB)BrU4nOE5MW48ZQlHA?TI6o/BD95,bLHƌ3>ֈ kN25;Bq\??.dlPQSPXW#P\ _,dzp\Y`uaԪd=obnip9X͇46u7-u~Ngwv4Gخ_cf^x(\}aAWv\lby<~\PhLLi.LiV)_k[ I5iy(< +ڒrm$Suhv>vkA4}Fv<'xʰM5oڐW.dn;TKΟ|;A<#9ۄMho'@N1q3gړD"O 9MNZ(+`z"br29qpcuC3w˛1C, }u'K!Kz}<._cw,Ǐ,Ռ\oJPBH-O& Z<͆ 2UmufFs"<`%4M SƪrPmǡkhy.v@s2+cuuy)hNdiBEkCʐ&J` zFi茨^^dk{ZBG"p,U+a]CēKbwק\2Nr !v6MJ7vY Iy)>u$H 1=1cZw*"Z9Ze АZh H"W ַ&b>6-M1 Cɬ/8|\̈w66 wk%'2 X_M^I_1*inx0n]/R#+o71۸ uwo] \׶.[Q]T(\v'᳏ x,70ȚPѭv:] &( `jէTU/ aѥBX Z!p{_Inp~Yx!U%5 ϻ!}v}au.@3L.j2n:ݦT.<=•$7YSKbSOq盛 0MTM]DqQzE]22"Lwj7̈;2|d e6g7:t>s nH7r< gox kI)Tv-z(N9c1T1>+hw?;61KFeR !a} `.糁()m?H >!lIB &+F;4^'<1BhG I$~ҦLIQI  x!$M=<&Mq>A}/Oth6eu I:ٻq$Wrᔐ,$EvN=[TD,^,w{$L*R=o\~Jڡ3Ռ1U`8ʻ tzp 5DBcEdNΆCrQ{Fx΃DCu^"OV`pyhG \!@.*a@|>_je XyQQ0C@ch|@&'Icf=Bq{|>ef/Dc<1=7=RKf-h01WTzGt[IT5DI"b(Rk't KuQ 1$_cpeX&lYIAj3Yv|lFwe㧴rYg#!d^#:m䧖эZp4D_a,:X4F<wZ}qb^H#^ q hbQ`C#X0eKFHh©iVSĕ _FP1՚̦pQXEcSk}x;+3V.hFJnb6Y@ 60w"nL&~ 1\Ӛr]\w>\\/RXbyw= 1sFDL9 I ,7+$Hk`c:Rt*y1ۻ )+A=~$j$R+p]Bc[{CcNk#mS D\q9|C)g` T:bT7rAu.L1)= #\R1$q,85o9*.or,vBZs62pL7[Cqqi$_G1Ҝz w_w=f d[w9b@iXW8.hECw}(Oz-.[v.nvm _ Ey~pݭ:|,62FJqwY'Z*{ }ػhH[*@Ctʥm 4K*Mqh:PJ*6i r&MFG'o7eSq0cq:&pjL㚜树P Φِf@Pr00؎Mh[һX,K %K ڦ ^~` ))a JBֹ`vީwUخT!H{hԜ D_8cZykkP弭3/W?~lHf7=^샯/$7~(z~M Ln3#S0J{b1A.ԓx2G)=\gٝK:ś._pוYZʻ2*2%!Ғ)$T|x`͗AOM 3ފO-c FH 85`Xs41= ɽvC\j.;keBm{q}@hÏ0R4+u$#2DKt,pDgT3R?h J(Pfm=21sP}qiI0oFC2b2AqIzhbA[fy̤d5b:mh5 At[ [.o ~)f='5 YPqӠ0A8}HfSgz^s,V2&N5{c`[@MD!pܕ&5`ʹmFv)Of]!O Z,1j]&Rs}rwO o/x,>ۻ˳ޜ-Ycj ~m~z L3U1rqqw hseۃ -rlÛ}t @Q- ͅ|Ǜ7_uvxQEPWqQ]eypy?N% 4K`1`WRVF_\{vְrvSx],]}/k;mٓlljަp3N ,W@Hcn92j7~󪍗It_QPGRn%U*z sЕu_ 4UUs5 pp.= 蹀EY,D‰%ZEa e4LkgmvSsf&77i>y稱Uz$do](Md1Դ;]]k,$٥J~kr)Xإ\>_v+dU.IF&цhCdjCS޽?Ym6n!&=cg0bOEcØ+*qP;[z!Z+*fҹ*ӊdsiуTTED0$p1Ou$ +Z1ff'>-&wdOU.6PzcwӗKr|{rҨ1jRy7w\ 1?B,iEH?2n/#Қ4q;çע~ȃyY*fgaFlt{lm4Z}0#.;f]a֢hHqWb #WYx%:iT'kt ,RSpւ.*+Mu5 >ȼ\[FS4z8õ* ; )IvTL@dsG=I; 3QFS, uz!Q:X5Hm])/d*eNB7 yZ/)B' j5{֨ǯsS @JpDAi R?IT0Ң-8B2-}YQP/uSZN+ =6RBWKM$'szwdžXh߿PƲLzSʿA5z9(c륐3ȁf85&?NG1Rc/bL}h ~?vfן!U/1c'X,zzxQ^^ޥW|X|#h Ap^zkuJd/<\0 (x4šB+ÄŤX49#^v3E zY:V3s-Y#Rpq{:@y5h27jM- ̭ 7|".w¾ޑi\s0mKO^}@##}2RAPNGn|Z:W0̵Nz-|i3vtd:qfy[8ZsFhhBe]-'9&ߕڤ=wKD5P ^b4cɩց4r*yTO|eW5N&=j勘+3;srjwF_t~cgf^N@g*Tq|-^I'},h=E0F[e!SsfwUh> wxXVnMh*lQ͵)(ZVߑ&Ԅg"OE SᒘJS,kRX'uܓ**ڸ/6jaeq* cf@.9URcneM~z9K3q^!у09S6 Qpgw;ti+j7U2gP. ]Qڸǭnhe+q7f4SM048ݕ!Ǫt9~3-S؝Z5NS8TlSuJ )C؅:#F2. \J#W(קb]Ԉ wTB"+BcS64V_NI+\؉w,N0Ԩ ղkRW4o>_Iǿk65dSW.*_Ljc JK]DT6yBW` #:7(OTXנ~, !\ LL@ +9d?b|FTTR.8 8b `ywޚPs2LájH~ֆ[\k_X%kʂ 4F4_%U±dDE3\xA$$Ra- `D]Zc 8!JDrrq4@ 1@PXM$̘` Tpyɞ4 yw%Â4s CJcG$=WX? 4%{yq]Ѯ۽⦆x=zAe9zOu28Ь?uj7;>lA T ںd6w\$=WP[`8aqLcsW(ͻA ZDDS9b˄|!? {rjLQ<ņ2YBh>,T_5̏+< iACzY]-LWfP;n>g{b/wsW,I(Ҡ% e[h!03mͭzbrm0p _'Pt:l'7b DMD]V|>^n<Yq ΁JC!QWN/'O ip$vN*D'3tG1v8cbBo"[$g%PB֩}|Kͧ^DԊV(]|zdQd+UR*큳bDd{ȑ`=ua$Xs`'}9 7=%d~Xln5dX.GdȑHdgvb ;czvպM]8UK= }&ïzo'/5r nqv\O qh`hNn`j\Oђ(wDʌpE!Xy*%`$ @Xԕy&I9PLxB"Fm;@K"l>x"׍Z9X(@+dsLQ+V(_,.Sg1h2?_9-y3-?F Yxr T;"Nȵ.r턋zpVxBWjJ50%s[Iw<ʧ{1[9&ARa}ϱK@ov~o ^:żExl+|e\y:O3N9c#T~G~살f^d'sPCtX!T2A5{qtfK Jr+UFY1A1mlD޿`vݳzA0J:efഋ*ULZXμPjv̖؞zS%Sn q:fu^"ɲPR3g;hٌS nCTQϭ@#>NT`&P턈Q7*C7֋Xz>|zT,v!X3ceWc߫ZRzAB/닐v vR?毯y+˜UIIT 0as%JK;j$TA Q D b&4PеrP)ez?띗~cr& grKF>qDSB&2N:_Fr>jZ8s !ZzuW;dJpBQb^V x17?\sN _ǻ_ѳ`x ~>o %ɷu>}g)P7e~}컫">a"z~?.g"\;{{s},~ߺKgf&Hf鮷T!*mBJڏU9 ^T hrT)ǩDEr Fݐ3hj 7o3r[*%~K$I[H<1 ʝN eP4Bd7 YIrNP4;H^Rb=i$6^gG)d,xBI4&J)3FeG4;jN8@P#jNQ8t&FJ(q{IyR)nq|G[A!q绳b ,{h,; .4CK,*Pb}i5MZ~P^kضZӹs53#O;|h42sbX.^Y#W4݊F JdGC Ixrzjy W_&r{ 3\]~{SfVh+|_ 1zg/ U$KbN*N7Ty3𸿺x9 Oy9 ) ѡ#,FBH! @(Gp|L Q\}#}j i-[jm3H1V+;THt}6wDR-j|"Q,GSQ9Jl15:,iFd?&VI5eBz՗0suZNt2X6tNB!E B;U. ڮ<$fÌcLhe"'BY$6yʈW1JI9 ]&Q>p(x9F_[}T`Dpp]:tP h<ΟKNpzZKPm=u{ޚӲ7^f f7?mDxVcVIZNJ%(Ŭ ҨCz!Eo "rXQߗW ,t)*z_!2ǔ! |ZE`VpHk=vp}_xW\uݮ ̉ Jtq7ƪÓ҄kw}uDT} At?D|]Y@l糿ܬO|W7_ٟg/*^?j=4{Yʭj::hx=P*>4 kZ=k|ʙJRn+κV[j59^&kBs¸%9Ofyw,/#f˚ہ "dcAaۛm;qw*rځz;|uFiW?hhnPptAA13&ㇻEhR2R`jX Lo1`PIxHd={@3CF;ρc`F-el3jٲٝ =#n1=[rh_۲1p鵔]t\z-Eۼ ;3 pѣ3L W[6~ڟ|Atw2$п90 82qopRtGYzB~1^(zUO6X3j6X.6B[v;A|4~.Zy3j Stn  d'se|2ssۂg /revٞ"g ^A*hKh?x,k ²3K 6 St6G\AJFI^FR]Ve)E,Ւ(ƋoGOGgf3GOA9Lѣ d)9[)bfD3ޕq$ٿB,>,v6?Yhq/l}#I6YYuQ|𨻋̌x= x{%joZ%"$+NXJN'tYYZxԫ=p`9 @l(rEK3QnELtm>:_^&C6ȳvZ[8Er)˥֒rUS|Fn]VJUkE3*JLŏ+u/T d\[; S;;ky4Ga-@!h|7X'NQeBa:M`گ3 ,+G'zgpb) TAKa# ,=@LzI!8avg,#D*OLu,e* W@!d L \'1X V}c`{:nS)LpSGxX"E-EzX0ю*D2a+e.qts^H YI+(`/  ` HÒS `P' B(o,**Rwgg+jdWC3\=)6] 1.εfOugW>nc5n]Ng#$'l2[ \zmlw${l;M$&~D%,xl$P5l#z4[f[ $c$7 I\WSAL-q0滧qJ\=8Zk4G3x}qLn'W$9ջ7^(+\Ȯc E8tE q$**к6Pr.:%?3.Ԟ9ؕ; k&>0i=EV! L7X^XxjǼ!;۽m]FDv,֕CPwR3a Оr %8%3B`&,pFWh#~ @+%a?}(SӇ*a[tQbg<}OKV-dp}0Z0x;5d+k Pz*2ǧvb+%!CF\dvw rNW& * ij-dj(2@ln?ۮƒTv%dfp'eWQF>^~s6-t0Hp^0Gw4'KWbH(]b}'9qqt6*mekacڷ%r l4cC}mվ@ִiv:+&Uk=WYPU)xi ر؁33]8گTǕX*Jgvz&8\*d&ZI>V.:i-,:)ΣcD~þq'W7=5vԌ J薻 Gރw= 4I_Tq2@$+%譱ں i-* e=;)w.HZB-Π|g|s2ԝAF1u v2v{(RP>:}MIyy=y /g gMB:@%m lm'Q= MvZtX?xp%複pQP]p-ftp1$nSG+ӫvm=Hc7dYBS⋑%`p(OQG΃s6J(* 1L`}{yUԸt藳Y_R4g:evsz6(;<IM `^($! e.q/bE-eY]chu`ZSS!I͛Ό|k$#Ý77~1.FkŚwss3p3 ōKZ5y}qgťxaf~ճ3:fDYÝ[؛D6`wl6ɷSDK PTQk]wQ>܍D4ޘ{quqr6omX6̎ jWkO6bB1U7wۨ"䄜 a`/~: >|լ$WbD0 ›;e4 |Rvq R|g'59y o&Zgfn?Dž}ϙ!;ՄlHֱ>c"'ysB^K3GP%QkN5$fҷΙ .㢆[!JD1%J.VTIHK$]"I}k_\YKt|x+Rn̽:U-͗EFV^@^j%>+..%/ϏĿM/Z[tm! imLT"@mbO |_䙖qDM4nٔɢ:|~~ab^tG QDu8aAT@rff>!IK7$3D~!evl\Ba*1Tϣ``7nOQv'oK)C'[V3#Cc`΍|β@t@ۗãghp{x?>/]FKtx&Wׯ-}$h{]>*Zw3< * W|=z?͡CR#xwC ©,䭂U?,ق^z;aq8Wb يqw}qLn'W15OFvsӇd"Rc^xF`RkZ'hvҠ̘v&.HFW?/).i.WF\qIhqȈBmњ'??"҂5gnҷIxyǐ!HTIc{.ar˧Wr)[a'=*4)9*r5vKV GYzsHt8S TER_h"i|SciDnxɎ:'o׀ΙX7P}w T7Trm?b1lxNwx68][ArJ0l;ٮ?l h=HPԡ8Pr_)A0vD啂RHaʸAUNZZQPCxA(UXLY=ݒ>fjk&'|>Oo"U vՃd*p@^[Z&'p:X+I/LA 3FܑV9 ˆh5E90*vJV\bj6O닗βQ o8=dAvvt<*E0&ZPwR A A zNyc%fXo%Ej~S|&v'{qR( UHbC$ Hܢ  gYG,IQX!XcZhRtCK?dl|e~(f~iuŻzz䞖[H/Vw `:Wd_ȅS.R'-wq=#2^7_o?__d:=g%B*|ztwWXK%QM y15&Ǘo$СwOKʑ@H)r1zXa=|Wlءqb%ъQn_ºI(gR|Q8E_P+/: Ɯp\p+k ,/ȀHRQǍӘod|khR`|Yw"01 $OUfԹ4+ Q3AR)(.VI C<V/# |s *G[@``ք|S \8~Ljb= iRx ZYk#Pdv@sAy#tX$bggwz99 |5-zh3ŵ#[6 S6Ōa猝 +`*1Gy\l - &1,*bxbD۾(h;9J )JL\)~%V6çQ}Hk"dKADd@%B| jbt1jIn{ P%~&۩bC#YD1S?vzw ݱjII%YSpWAU0;ZUHU{R𱐪wJ88ʭ3Tcq;⏝ӳxKܬ,=ӳێOtN}EuAܚ,>ξ/Sn`({yQjVd]r2;a3ODRʅ=}By.}ɐ>}3f0ZQ@WX~%qd h;)ָԺ~{I3Mn}ҕldT'\694.X)O#l0T FId V4fq!pTBn\Ta&PhPV YsMu%^I]ySG %Hn]]Pޏ^HXE1a/-"ݖWBEao̽}cV><9icC뛍5v3SbJFO'sNn#͗Uԕnn~6IE)f/q[I_vSg%~*d:{Τ2KR. *9"OY'9q͐z1jYgw2؅vb;hWkH˸#jAe{xks_W $uj8{~Ds}pJJImE- ф!݃!;Ble]/5+nlX.7q0}X;tf(Hy)5,ɮ..3˻rewe);wx_cސy)_ZfjEOθO#˿Wz3],i\[?÷<晵Rd1} B O17>=^K| /%0 @ʴ˧Gita+h!PTG^{G}EY][ps "ƈWw~Ng>7şwl͒#;K̾烀UH%.ۣOWA'T#w3ǿt SK1lѣH#ekK BAs '$s[+\f< HJ+=\' }R$U$Tj P;7b_= 3x6u$ռtR#;D]0RsavfZt]bHHԵX~TYӽ+Z}|AU%!ѷdpKGj,PDA"È')'pݳN"}tY5DXq*ƽ2~ 7AH׳ea3 /&p"kLA4s[aGb(vhlc)ӂbUgPh=Wg05As pG7p #haHn.}E)HkN6+ea)F ad;\ |Pqs#)r:=)ץc(j*(?""΍u9aʝ:wJCa"T rP剙^X{fU~!5BXì r0_V1 VW/Z̄Ŭʆmm1_,f_;0ېpO?ösfwX=ả-&>'%wοq>`Mn:}هF6V5㞲b.KeQ9l=lhgϻ蹟.xj:-4ƀT`-+k歾!M?<(};-R>7iX>QڹZxr =_Q3@,bɝcAH̠ sn2+1#VXl9j)"I@6e* yI0ʽLa$7yfb!Tnr56AjʄH_L&`+KZ4Xj FOQӆRUWX_FBaa ^z 4*̫B&;0Y<|'cRuTz.^KK7an("Yf-8"IШɸJ/ѪNQ70%Pm[jCoF#Q`4 T]hǼhs4o.o]r㍍x\/6ԸP[JB!#]QΘhx+P8g&>,)k$loCtmb DW;Zc?7`r%%2ې "Ҡ}|*xgUxnWndMnd7bi87onc{E1O[ͯ0 ~|>2}=~_D>I:XC'EPAX*XP.F%Aeh4B e?o-µVhqVlzsYc V@0E|ѻղ4;vru:NOMJ9!';#j(f_IT~T u,ryܴXZ.241X3GhؾwT G< ײς7^:XU)}(}} cߗ%8\%4aa[CE\Ke6s4xď_cN ;"$ֺEåqQ~M[K~a +4X뉊<_P?WJմ+q;XPv{cz5$FH}vcΏK%*!t|6/m5cBrmfWz*c)Y.$eJ%鹰! YL`e+s`JӔV $2_I'x)`Mٞ }X3(e'`=qH[0lALJP%ꐯi)c/}JB,.Ft>J0lL .j&LiΌz wru GNxnBr^/Âi0.]8z>~,,g`2{3"\]$,̚wzb) C /#_H:DH/8spڈʰYp2rIDAX{0`c6SJNmGRr] Ŭ"Dq@G'hXMYK,-5'D,J!m0sy#a]DSMa*2gFqDbK= 1Pzk,曥M72b+{';E!C8,D- + XSx~7E4' we@LEu6@L!tC+9>Y$lt%>dR K͏%QP{)%SW=R620캢Wy]dt[-F+NiwOV'-1ԟ+1TKCczPj ןe-$Z`DSw0"S!J#6l4VuSpn`N{gx4 k0ѡ֐w߯xw_!C~_6cldtf֊(ѐzswQo=ػIoIгAk8J޽̘F Ӻ062:0eD-eΑLpS(|A1}."pВТRz慐#KlN֒F91s2PeHY {%rKt{/R PmdlU3p#$rAUXFva)$I:r(' 4Mk$;Ej-Hv ]K'$Ute$M4QTYMp JR&Q_H+;=s3 WKD[.K%y>/ܫ]\ 8DW>̿ CB虅ӝ0vyw5o6//RXH:&=W%:-Cɮi1._tqzO=OyE\Qjrr}&T9[_;M4ɦ$ıwK tRۨݎ 7ͻgFz6,䕛MHbLV'qFVӇ,bT}uмXcn?XL]Kq 5mE$Z؛RY7/4 2Kydakm6[).Mc&X Bh)N٥nȮ%=QRBړ\##\NEtڦJ8S- ](Z+y kp!0G#]X$)\OHte-ʴ>H X%<^r5k-@KG?`4RӺE!Y̐ͬ(|\1 a",íae:&e->ͷn U8m`F7Sc4s{Jx5wŊ7Ɍ }W늢 's f5&&t *g5שDU[H[uef;ZXdЫ)#;bruk#pZ(쨱`K4b PcLR!X,[#jtLJɪI1. A;n|gjhu_HWJ0L$%8…bPPK*Q#(Эz)7eƩہŌG0##H!=F^A .W۶"D1)8y^';aNPn:Na&(8u uB3A=uC:Nq&(J(T Bꁩa"%ywl8ӫ~5GZbp`X?]b3(KaC޽kJnQʯ579?hV3vӻ~nyUvˬLoһE rI퇏7ƫl.o 4^ɧjp:Itr(P]:mZA39li\!aJIn+JzЅ^@ιKAJISi@ݫUl%(օboLmtrg@ k% u+nFțm<&$b/ j~P:[ Ma ;t-C" gxsjf4^f <Ѷ]dq YE $,t BzT*Be:^R 9Onwߵb >LW])',`kvkfw*wQ"hdeU9miޏﷸ"Rx%{Kpx /.Ҿ׹q:N[|,(4od ɢ QBDU0[|ܑ†vQL'c4rC-G]lv>ͻ5䄕RBl^t pnȷhG3H{~Ot Ң)E:dj!}DZ]a?{"F|| mO^С>Lą\gDzjR2#F`!_}xhЕ79[AS^~VGy]JB@9~Րu y؁sx?>w 9 *fT!bnʴT_s1uy^m@,(5C|^LlA~Vm_Y h3B;=>RBTΔs*dQ2vjPn[ynގSmzk+žʞ/zW/zz1EG_S^UMg?xTϬ0S+΀f\Zš1h|- Ķ)D1j׫ABicv g!B7g-cBT8g|۴y}Ųeׁէ^kټ~ͫ{l.-ik6ֵ~t2,R?qx~~%E,$mC{u{e}ql#ګ[WeԹBvԌ\흇߷?nIqo4ȁΥvdhY[[.Jk=ƨ[H7 @kE gyŤV򱍂DuzQDxuD:qxC#k_T v5[_B,¹-J*"d%dbD:ADBw$UtRRBJđY(! !3!plpB{#xL0L!:.sJ!p^h[iY^pW38c'SJpew_~>I<#xE(DgL)@xЌI^[u 8#V9x>j )Ji3v-#)%H$HĶK('g8)"A*Sh!  -2xh+%V+"Pbb"̯Yv;-#JdfͲ<.#UtO\D +IX=/r1dy݃m|`F3z~7t6_LJDE%ȼן] cCNI~ogGSC s?iy\&KLz7+UPy g /b7}LJo'/!`#3&dmdu *b :zΖZϙܢ$-EـF "с`+3qN#$<(fq"QX21H@8"c:$uJr+{=FhYh` ~?nןδ2ԳtܣRT.Vc~i]0qE(<48>|$2#/xw!b@ߕHG!C[) R|rz3,^1KK:}Ob|RJR8u_TV'I ~oSCGr{hb$hXڪM5SSB)TwVa‰lSY; Ez01quI:*y?/Uѽ_f\l+YytU'saAK s@tَ+[lˌ?w/\_s\'^.(r[]) Joһ&=oRc!:H&oZ+G Сh(ܞSAhʤM0J= jLdbr`54"|p )NPjIi@W"\w^s>O(V@%3(k4Oaډ$A6&T3JӁn $ 1D`@SJJlf"¶HnX )?5HLrh%FD/jS՜7D)j-"m}>W NUfqc+ ŽO>+_?P}j%ea: m^mh][Gw-y V^% ˁqeZY5F?צFZoխb1/*QLC0Ju7Qk*ih{q_W;W֟WQklѨ Rqb-*f 1m P:G; , j\j51(8:EY)FQCҶ3RMbŸKOVr/HS{O#NHmmJBn7 ~bEX&]SZ{\ӱȺ;JF)yX)9bR+/5-doR5+L1je@|9@U/Pz+U쭅*K c)Pg!)2tJ*Pf .'(7eEmbr> OM&iߓy#$x8N,:]w d}mXw67O&f_Gy0w.EОw^감Wnlw"vu:nU11(ݺ]:5V}sӻa!D+ٔӻލ:3CVө2ޭ JݪӒxz:,䕛61vL_>)xYH@8%(j 8ySgu }n]H_Ov( ٞ"}%\m]'GI|?EKBn1{!EN#~D=Z? :pWӵ'y  L {k}٠KaYm&RfTnSPK=vY5P)% +Mzm@S!ބXƝ7"~y8EEz}. &SRW&ZFllHz*HTmDKn7ˏp{:80Co2}m!]^س|%+U⋞^Gv^ Bg"oEDi+P)" Q%@iTˈfg )é€ N!#Qh!QJ|Rh(L8’Xu+(M2,渊e$,C<>F ?R$$a'B 2̐9TKsS)@a)q @JDUBKV3me b1:A60f*&Ʊq)r-D !@fqyDYwX4@'xL[8$IAX*hqSDed!dV}P2O6bR}ų}w?c%yNinօ~&pC{~1(ČھFC&2_;W1a$vgY%K^Xޢ*g$62wę`<&4wQ?voIWgGLHۣhf,g<"AG3cǓU2y_}Wq:e0nD3'BC &T1DkAʒCb(L ]rF^:f}QD?L\w}xBQk7mm&),P HTXJjI 8*3+++ӳp ,:|P`,i>~_mίݵ`AP!(G#B=J+\1xKy)iޞ|3Y%L=UmG!hYE,a7 *.z,o* T QICgE۳V ^Ri- iY2Vt6+jQA#Cb0TP2ڊlEf庂љnVԡh<ڊaP#àW*׏e[36OfS1ΊfE-*Rko*#CVtf+x^BN5߫aSfٶ!s=W4G[JBRZQz ?54HU[!h[ޢ+V~qnrF}rQ9wF3:|6Y`:`3H%2!pnxت-%u٥T- nftxS;/N'\]˱MɇOӫk6rn_O[rF5Rs&l%zMUKFrbl==o \@ŝaY9 \s_>}qB}0bX a;EaE~( G UN~/ ^-g)T<}!FtH'1jn1ZJpFC{#-[l:c>;h]Er$8;i)iN#]TFV'vØT̘*^]UFar jjxIsXC9_C JȩpRly}tg3xNC6Mi{z9#(EˮhQqrL26=Bfi!.O~{p B-&C+B}6 C'ՈɡJy uOj}jʦ!# B3-/xJ3FzN(S9u.}Y_ju8Xٶx!%S+Teޫ^/*ϊ>#Se1?y.i X3kXE/3>T㽳HFI;yq*hRY77MѳJ%*!F+)B5B-%y7ռ:0U)mxd% _ZBWV4N q^I.f"R0E>9󻌆FWn( EA/Lbv^FR }~ʏЀMrY0#eavwѸ`SN[q݄`%󈍙K FS7VtHsTj>H{ѱ ?YJ<$$rƕݦ2RR 1Mb=-G-J賖#L`F(:U|QԦWSb4; 91k^]dg3!@8_d$p;lKiy.ږܨ@Y\b{~':BRy縞 @vE*Um6*n6Y4 ڏ۪X h7`k^0l5$uu(ż>PA'r`V19OxXh3876IMb FA.;YD>Cd{[s1BNç",i1yk^qRrvf1ɐlsg?zBdhƭ íxݍ5cD}5s6ls\@G/?=}.o~:|y['\K2)EYW|Bs\u X:֐Lp/xc,qjHǪ_|뢊R g?7Qccu{KDMI1҉j)bL%Z=JJE/;$(´P"j=GV}GNhkI0W6[Ǟ?  WKJ6E&ߕ}*wxlq9_gI">^aHr :rFP6^3X{n"#XPHӣ<B}lIҦ13(=>gBJGDB ^7{==;LG'mwAB{pAd. ֺN(&B)8>C+zA9W]Y^ X/sl?+eBo~h2]W?PHg4.9"C-OO8YlwS/e7ĞfW [1 ~q]t#BX_uiX+f=4aa`_'ZEPqWZIͅxhbU={uՊnb.æ̻z Ӌklc bQ 2?tNB*,?+ė̸0L7P[F i!_X+t(rRLOZ\WnH0V:fc XsgEHEjUP"lxRB JQ>&0sq<0/oAR'' 1?,_%yz@u:#DV`~4? -jII,9ф (AADcpr#n(D)anJ#R W׋i <j.bcx, -RpΔŁcKs$eB5GuUx@\I@3S" Y"NI_m܎lN8NȑUShǬM3zbgXȘ Zۋ<9Q]q c+pFy ct5zƭoGJG˵vp=W"?zN{΂1 }!]̂ѿڈӉ$%$  +h~BZzɂ1p"RsDuזABdqi{>7?w |$AY}T=?F ]yfr;(]/&2N"Bi&H'evђK;970) m$*V oZ:1>]%WX^;,fC>jV%9vus=94myƋ"d?y.DZr1{e뢭=4W2c3*_ݍ%CͶw'gY]Oq\=h ~.g' G#b`ix8b4 ',xQÄ :q'F0ڃ4kEHfS7O9+рMLAa,0»h\WTNFtB05I΄ŕSH)&o׊; ,QG[O:`㌬ZɆ/N&׷'``ڂ:IV fۓ< ȪheGM] Dz~6Ltx)GG M_^5UtD4?iooYY=x*Tc^ooSp~wJV/绫'vȴ'pQi>O%)$`_RZvY;| 8%8xAD'GI_6wvmytlgȩ`1}1xX(́3q20WwfRg @LO'kln^6O-Mv(rB5EWzfM7\).i` +%*dWg@ƐX&ckOڹ1ufQkeB8|lgI@1Zh(J ڦc=MZi(Z穸yVuSDR! 1P1iwVNZ0/, o)LMj)k\&K}vfD9;_BLSaZDKI$eXQ'_(A V8d9Vpj !8Q [e^#EL+JNZ\†:I}I^z(gmĈRl૴KX":"ae߱T  KS,Ğ^ha{Xu;% *Q%ɠ& p4v2p8@MQ)m})v-aO3ij {ܩ-àd8b>1%D88)޵q$t@>pX$//u(EM߯!9#p [տzt W:2YT%#]t.fY~6u; b KKPv+Q ~AzXͪ[oĴyNLĴyNLWJ[T.M3ϖѽnւf}.Ң;o5Ev8uUb NDPU0"&n*2(^F?)B́C r:qjh;/CVpybJO棗Q|gVdiCQ}|i@ \QmMr\G^ ([L."YnW ݮ6" rYB"` [Pf}SV5ԅ%\a;Ra6lSxvGO'X",WG}7ʬ٫¼54yCrJ+)e1놳἟@y>P cL"!_n.*[s当Fd2]l;EL0~?V;ehv}m"dk'Pb{bq!͙DΣW_=-D{4F7c~Þ'ųEЅZ/cW^̾aƯcy_/-Ύ7![E#[OC[Z 2409W\'9>3G[jh`62Y0sLL_&šf^<9>HL)Мkd`ͣsY(9a|prM~Ouݷ=҃םEq3N+r=; oo'j'ZBkEDJ֥\J Ի}9'E&65-'bCy`r[H0&qtY"*q" -M(i,7$JȮP7NrZc|]5WUJY"7/*pT LFTN%j\ڱ{AӹNM '+/+<ք9oX -D&&%"OS{ʌB<1EN 2zu4*D܊#F`ԁ`Xks3ϛ3-ge{bM -Բ_)EÏ^p:<wJ r<1ms &ڋjaFO-$*$]y"&F[.IiMS{~y09&y@RBk` E[PsY&ɣ$߮,trdKE[/xk_ MHGө0RM]WGaR1ӸJb& W@ -E"F!Up%n(>qfZE'LZ.t^j8kmW Sf+T =&P5(ve+5LqJkFr_ {HSd0(ϡ2%޺RF A &n[ 9qwڸt9e;o߾q40CcjtJt O=JSfAɣ8m͎t,8ݕD9GN`W71W$3|ojf./gK *s1icm2;‍ȗȯaEF "]QǢ1̋ӄjg6c49 c 'D_V OLXOv"TU8L[P̭YJs7[FߊO JP-Nos$2?;vw8}v7yw.ޮf ]Ҝ<kv5yĿG߉^UJ#uH]4D Le5ktJ""'qE/[[*s6ߓr39/탳\R;Շ/_ho__^ۡ_%ގ}:OV+x{~Gח\me6ґF́k9^;%UzTeR^+ш} >uڦ Ezuy7]׾<׼ȃ*3]%Qɷ)S&dDrBX5 8Qժ1a4eN<.h{$M@j5hׅLɔ9#tP2k\N'Wlk?y)` t?][ aR 'oetg+u@6yd]A>Ԝ&ןdW|y!Bc[: m\pSߴ峡R0gT?[rFq{FJgsd)zVfJ.͟f } 3&ЖĘ= B1`)M,X&x2. 읕ҽ#}T. N FxϠ=^S5TOWK6[M'O 5:A0][ń ?лe%.+wU%}n"sK~$gNd$xH\( 6T*`>2AIj$М+d)iD CJ[GB),ќP)¼#:5#I]PdWS+2*YhBh(D8J7ZAs ^C]GeщI$ET)+D-ڐ=1UhqPс)7BєS5 ԠW;Fp$)%zz-;b0Uw&%% *]l7APsI -s"Wh^$n9%M. ÒW:D"cph,$μ( $(|2n5@:{~>?r~̎q]QtxRnնɓ/?D{[^mt/nJ+h5. ˻+{;LҨoCqVz tBb% hڱTٮ6Ǐ#s{峤?-"5H\X斃qYK*P+nVwjlPwRJw&E9z!+g7hpp&q4q[%߲D!bAs-_C {`6-ئ?@}6"$ (> 7չcJ;g6+q24;ƻa|&>jsbMgQǜ6im%rI+jC{L2= =sȠGAkL)=-Qh7æ.a ʐH[)X[RMp1DӅ}G{jjwnOx63߹c9<+{8!˕ǿuY*3YGzV;}9!];”mI4L/eZ~mzsncҐ7ڳ| Oo.;+\p]d>~~MAs)*} =alu随gi]Tx뗯>$=:b{^:/}\}*s]߯n+\pB,p ?W'lQfo_]^jJ§~_*}oۻ/:`-(UdX_^`f#-?)^7pY!f]1r'(SDZBv{@`j9pMBu1Al =b. f#Ջw6Oom _pi|gCQtGg1i3~9$89nLDo54BC§K˞n9Ӟ[Ѧ.?9z:N.zO{սǡpKLd!&r~A5K Ry2HeD-~"JVxn󤦇@+L%fγ29 zzGfPqRS>F%l[=߂Li(hJE&\6t8 'v(LQhϸ5!pA, .%bRA{s_,?&6Zj \*$4ʩDEZDXV 2u`BL7/&E90\L{4T ן|r^Ix+lӋB#$sG/_="Ƙ2ܘ%>KNSc~=mYiQ m8 xUGqlo '^9G ջ'N ;Ǻmiq8ߘ&ׂ#Gm MnH9:ů ?6UZ䂙tj,Ac)@϶G:GPһy=HAW;8x ZHgEO%ΐ;g3jZurγ493ՓVt?x1Q7O^W Y+CO{! ZKHQpHɺqaqezO"u69>Q.kknV:иCUy88[yس){sI0X");ޔ6"bJ*l"F: "jټXKP7vt3XgLog;=dg˦FRٲ Ď9э?~!f}Bl^v/.}27jbʼnԿqۨ\np4^yr} eV@E;k2ܼ}G\l?ha=įc8 l79(`hʔN<>"c(28I6uCF鷨\WxLQɎ6R9qR_SuBLPu<:lުj?Z%(.c7 A JlZ YG:v{ǫVL !54 ږwMJZ?7"8;_dHS) @ѢpxeP4sN 3R+4*w8f0N<,֧.o>1d E@m_!x['`OߟLofc}?ut~!{nd6s-~uu<`)h|TW!exHn+^:۳ 퇩EAyycWQbWU Clg:L%x>^ڧ[Glo;5L QK(7*OM!Z S塔h"VkÆaCI*9/qwu*B "7Qj K;î"ft+ѣqu L \Fqe*ji3J5YYHZɌ1 L a2"%ryցzxjwzԖb# ^R+ŻؗCק74o-&4QO?,TѴUyx-4Fvtg4|ܧmF47 b,RX?yʅyCFGm8 7 k 7 A>A R i#̐x 4Dl=hiEUMzGÅ\ G &,>}O萴'tHp*BxTϰܛ8\[ʱJP` m+&|ZˤN&]Q{2-ַFhyGT˦7M2E1VꨈjǤW3p&ث߅Ob|`[]& 2 v"#=D N舸O}q%D]E4l0@4jvC..g@sB& 3fzoQ6QF(HFTTVYe %dJ_DQsVCP0f(]i8DL 53tvD4Rq%Q6dMPAZ v]iw0;hB<;úAv1dX9pDɂJ"rt1R!qEa}FYCXM>yC)x+ܘԁO~hݐZDkAJ0R6pNu\u+1^‡h1%MްnAD7 e4mu;:kϺ5/h!/|S:BD@ʦi;gDwi߼y̾6hVXWc7!x@ўSCD5PD ۘkP3gLeKA2%û_V |ybBKE!s* c\E@G Ts'hܧr0wmR9V"J_]MRfBf2u2 HsB% {UgW(E ?xUp"" « MPTQrf"~$ʴX(`Z\"%u/6JҖG5#YG$ ^>Shr$wh bQKkɒ_H.;'rA$?@>].3Ɵ/1E8f1 tr3.&'|-[/ yk~=l76>hХ26N/sj@};~,_I`fH?i=;iyΨligOu> t>[VZJJ SĂY 4H^e"+)deE%+\2RH.7(T!!Զ4ALrV b7jgB)n/}zL íYBOT*͔ dʙK}(n{ շHLs>(.+IS)vtWRe!LRZC^o`4/T,$s8 ueYd$rӽ V#ŠKUH_)*υ ,ce3\w24ZC + tfRqAXi.\V&uBwaR)|xjJ^0m4Z22Do{ )U|d T0ŗ%j^R <1*^_CdP! nƟJS4%9KzY)+ӘuD^NwZ*rlE H%mG]p8AIIm@t?P)AO#*??Pz '1}z9+%'-nq=N>gg@N/s.=ɒs>BZGi\Ni9 >p/C/NJ.&2OgkEb73\uӧ'unfy e/G&]yJ{uq/䷻};i|R΂6\pB{G. Ԟ΋‘8{@5:4q$B]O׼$S![ 5/LlY85gQ@'$ È%k^҇CoMT\qf[]H[]|]V+i>՟F/]|zѴSF>хШԲ7ժ;*m*Z;SwWQ!š*o- fVݩT}*`6Srbj>ף\tɁ8lCNQ8pN1ĤAn0Һ)[pAV]," &s9()p^Mf!2"K F!jI|ᙅ"T*S.LQ8jBɌ$$nFdUA莿RƁdMt*e-SFڒI0Zۑ1-lRH8GS\b X\w u^DM;(>VCv W/O.m/*+aQ ܍{3}6|B>>mQm*blQxz:LO(H)(5ҍMN("! |fV)=hڸ?3:0rZ{@Ugd"f&, as?K<^g&"s#GݮjeV|~EmGJ)Zͧ@'s(t!gOηm6n^R.'y3kҶMhui '#ٕe3Si ?BKJ?:-IS=`-콻$pg j1L=61J=W %FBbȐ$N]Ý  x$%hXa!>)ZPnƑZ)zu'L(fӏU+RzH[wǣ&oDw2IhXuԧU-)iyRshp0 Ǥ,|uWױg'_}ӽ#]:*)2NˮlЫYO!͂O\ۛ332&@. I2p~^N ?7#m#꟧ACMPh0}cY/Lhl+dk* s2*|jF"`/Ѿv\Pі" (I8^s`z ϥ!r)|r|;NoʲSQiCeEQ8Dn&/[C|QmG~?&qk#v)S.Q*L)iJ\KLRe%*С7\d8-e(PG3ԯT,F`Bdy..LӢβLK€΂1fs+ atC=6<5(hBoP^ L"+XeK/𹦍 w}]8)QXάȸQ' K !ʁphd8 sk 2T?p/p &!#-+S_| ޵>mcEOۥ#3{IԎV\=~AR)A([Op;y<,6OZ<~ (Zڰ.co S??|êZew+>-'~F`m  ="Q,0هFaj'lj$ ~56T% E9vo^ A+ rO.TmiJ ‹QgP9JdIz pmF1 /7SPz/ H`C ^d~'D{(5ܛ9A0J99=DzR@ܑS; D 0tun 52K&sw}ýzVABL0r4!DPgŬfce N8]"[V3ac2Fn}ήnpNe΢(6a<6z]])yJl~J Kzk~5gOg&IkWzάJUezVte.ĝsوsVp-F"ҼH1|jpV|/mpӺ"cJެ΋>p /2oȲ}SN'tϨtfwm&~e+ܧ "#Wa*KB2$p `% QJSL w(>- RLCkΧ |~,>X;>V:(gdKMM|x,+rZ5'$A#SZ)H}0KR%5'X[&NsB#nk1:GGwaB:s2_7= SK~!0`zf3u_L Q$Lv|>@8~i*N4Sh(Fv 5grf.%7!7ddVcs3R \68A=9:EA%@=⠒gUT "ok̵S@@MtBR ?}DJ{}Uӻ=F'aE':+֔Y( cK/!RZҔq ԍXdzg9"%( Nz%)Ҝ;bY 5 Ai}'lH&RnVo cJ r8+ՔW )T2:67@{pjUGz %;dݴ P O0U5ʀԜE]y oЮ9ݲ/IkC*+eAI #W :zZEay4KTV7<q4!qn"[%׿Da Y{zӠ/;l^C\h맱 B2mE=l~SV{5x͋ zB>"Q{:ͭueFJӻVVR\ȗOu?ntm8clg֫~T;'7dy$>,6(`u'plNk] 1l]jqMxA#o5^7%SHf;Օ[ 8i= &~C,":Z|"Ld:u>ѹ>u} 嘷`0rιXM!Qi`Nu'Ni4h@E>sd!9b0m$GvD@:́%poTFBU k?"l&sd_bfY0E@W#z.@;_r;>mta@P/mojCB# Mm7WUDG˶VGOQOz8{񫎢*ʗ;Tt9TgFOp:W]<)?r=R:^=sQ{cPR1Z2~< Sfsx\(7ȺqigTнF{HH AegY,@9!f~9 {.3 gH${o3]yk.3`dh ~vǻt0?<Лq3#h*ٿd'酅~̭@ccvNtg%V& I+"~ݼAb1>hm3Rҙv nMH+(; KPb":}n;<-}<[E4F ԮvpjX BD':턧R[|@քr=ZDt4, rY?,$c >-M/1i}Roz~~dlh(EbJ;vȟ* ~$j>MT˂?XPN)pSpH|9ԀvAU!)ى9X0zbtgO >=)H8(Za)9ͰZy ? jVqco\X I8HVv_o3_a:+^} iy&O >F9Qۡr({ƌ&:^X˨lUK shAMBӉ&M h"PyBCЄ3G ~(wKoGu16AW/v)u%"V;ď @ޢX;~ W bž>z„;Ge.$1sL҃2,y¬/mՂ|,gz3/tS\]Wmݭn)JӢ2-G)4lOl=l Wk:{YUB]P~LQm,+0G%Ќ5.vȢk WoqAJbmCL`5 y1x$6ۖO.0ޢ!eG!JI8^-Q Pb9? wAϐs{T,J ,W+F{tz`e{B:kiXU>'7V{o$Gu[ bf5k)YbyyXOow>ͪhzL r )o< vvSk3s=r="qwRVpBeQ}dZ}kpS rhMx0a efͤͬU cJ 뀱 b N*<ǔcd LNSd t13.P+,M7^B Ae}ЊjcyI?K }zzRRV ~T+_~%pqzoO͢jSTfgos|\=;LA(IC+ڂbKAftb8dLJxJٖe)hMU\ f'z}A2tF(,&<ʋ2z%%KTR.(W [Σrw!MAS!y%h#, 7Üpu{3\NX5^|{Pz0Hj;̩mZp1Ts_PJO3œo 5%YvlU潰|Lg_go KP]A- x^]W}G 0JYÃwh!-J@'uHvFUD}^˧9 96%iRe!Á%;R JˆQ W +)EhU A+@Qp0Npz1G p)X=$:4'7-1O9k_9=Foe'.!wH)+$ta%>vԬOAprh=b%5 ӂKptGޓasX N!M8!ր$'XZR|4>6},mCڿl~fh3 ?;O.OlA{μ 3 @xۭU^Ԩ<%oy0":yV-֪R^s &XuH_n_ .ɜ+30AR|}Y(OgOI>fٹ%3*,D mBJLe"R@[{ϊOv/a?.=ţ0P(igY>B`G`tN;D@pFWP6~PadP@%ckCi "K)VYbVczFX#91ETX+ R*,%S{젇 !L50$+A ;]a@İHgS)1 j.8JGEEY,J¡3GaR0J e$xI&F)dbžJdtNʢ /K!rӬ<_Vo̲L(}x|4 A\?޿{|>,0ϛx;l5}?wb @I٣{ ~pFb/nw7C&7O;c5_J Owmq d{E9g1dp`d͉nь.Og鹊JH8_fW0>G;ьLӓ7b& cOȯ~3.J WH˛deC#r-xmI ~̂!S.,a^ _”@rsRdY,X/o,&a[qa[qYV;ɻ [1"l%D1l +e0Ej A|&R;_|(@]L:lguzɊ!Ek"oV?+M3s7߲iHpˡA;\1bvQSK#, 0ya|a.oúK-wW:i×j{).ÀXA,0%G\WhRшbx"NeD,xa)sǴJ\AiE!gRHS`IGf0,w<L7?$% >P<ډjI&+N5߲SzgVg&,l;&!ωY8?`'l2y3T]bI)xq~8 X%4l`\Z؟2̀Pϐ08!z9a9[>ڻ`hIn,cvMh, $i1`#9'(Ea-aqc\h"wbA /Jk M6CA>ehGGiF6dE^'lZ59q.=3Н%"WyAU9U RKx_:D]<>|rx7|7vMP5 ԻOs [21 uh;UWgDa}cka9+>jzPƽcHՋwF#w#|:Oyv:^Xپlj[~)FKkhTO/LjQ(Bu1%=?hJw?yZz}jzz S92=mյ5wm $L'6^+p( ǠN*.%_awȣ1o,bQ |#o? /3VgVr;xR1/Oi>4dyҊP4pxpk02xYXHE'D %ޣV':I&vlBrM)Gx7]Mлbc:݆ xNͻ%?hwkBrmnSBaSIZ}eM8X$l l\CJքtρ$2Lp{=:Jʹ֫,U$vjBϼ_;aJи,3 cpz1MKiٞiuRbQE& BSӪObCl#Қ %Rvyĉv&˷mnѫ߸ru586v)F[u4sS5`T{WӍ)ij F8=pD/##vtDi;Cݎ }Pi G5Oja m=8SpX\eO/hK]o;CA* ;{sî_`ZBydv>|&n'aO XQ~u"HnYZ7Q{ÅJf,T|,T+0=5`KK l2wkRX!Dh\chzw7:yiTƉKg?™k'!{_=w GT"Xbb1s(/a`dRNcgma@p+c s)-[AVB7FeׇU1crga*Rށob\.ŘתMJxd/΅Q9 (T Iad[A\D #",?|8qSƤ =\˥\"u;fa #[u!碻SeE/sfq//9f@#:u1(14ns8>DxnjT4?Y λ3 Di'CZ &y VGP/WNqVNQ:w=v/4SH)re_V)i4:=6W Sd'ǿ5U1Ra|>fE=miIn=rCɺN;vT p*d) -p Qwm=C{{dKh%b*}Tշ#Aʣ"V*k]+kCPTg PNA!MWvjy]0VT}kEN_dlEQQf"/ 0?%+|g+čc(=ObB=V{G]/ءB\_狴M|u@k:Z2@V3k-uXH# ~seLɶq|Orxྮb&jU;V,xH_C `ܶ; 5dG)ǰPuN! 'O5*r,|s<܆\lhx]P:w6ʎ5bR~ ߅3]?am:ߟ,{? _6[N윸QeHbb$ X82K .od7gqp|7 ^rPa%u c?N~|%aYW+ _f.Ф OM9l`ҋÿX2!ָ e5X]!e"w#xH &kRpB  W0o #8t{`\9YAH9밎X]zEV )`!U !QО FH- sIB|g )X)q%:Tːhx\<H'լZ^OO,eNOKʀeX/3IJp{ɪY"gߟ/V{,93AS̵/~\~n['H2Iӓ7b kE8!dg`؋qsQB-vA4TC$E|2J֙ӿF)-;p7J1רk۲:|4*B-ƌ9eb$0<%/ . Խ'cVuϊDgEYq+FtϪYmyFt_gШT𧧾u:S]뭗%a3'踴B>uM:*nT*%D5˘ka¨jG\o+:̰:E;z 9@&{PPTzQs`E 2Sdr$&^3n rJ&8[U=>ɸeDX`A$RvKqhC.-l S^Bp+:C nlV(P5:%Nsd\[bKڜ̴#c0dm&-8HI#ļ6h$!) ~cwȚ+N(BJi2&,ʬ%>`ΰI@j$\poA LOtq'zTG9w6^̟>8cֈl4֑K!֢^qg)qPEBj46/|93[ba'ׂ\SWt#):댰Ϩc!} cNF!Tv<^]!:jys,5bCKK$Pc4Q1W8D\2>\ÑE.s"W(#>b?jr 3c~or^*jV,N>opu&=>qj8?DN/ʛU8r3s=Y-o5y8Z.m0+2{>I9%zDOr' cY&JɌb35͌u63[o &(mj+aby 9C5u!%)Qou1DW0f ePa[8yIsic=9-,.1'<0>唫JU2BkBtS_+#j="kT•Ử& ?& c;1"͉ R${6#V##i _Wq+₨/K1"ר jECgӀN?EvE'RGBw?)::=}*e<МF, 4Y;E Sz;G!&ېZ ".S]Vi:e@ >StN^9x !_)J灻YշƑ9´ DS$Ϫ']_k l%E p^;k̻ P 6k")@P!ygTwVpj=.t7 Ԁ1*)PIFTA3QeUQF>^]w{Jg-XW+ߕōg C쎛r)hy<f1'SaJ ORwÑ }k4&SE]Z&TnV}C9\xu {#sO|L@.W~CW>wh:?<#HU qO=y;QD?޻g *?_, cZu'Q~^;v{]>+s@ Y"[d,Ak-fքS&'+S0ñQs$CJ^;{@Ά>j F@Bn e_E912 H_(V|=s͎<Ĭ+9ZZ$s0F@ )0ZMʕ$ skAbPD [2),%EٚIxAL p&JvYl!FRM7e~*=Fsh,nm ITn/֛w"8 "¹uL䧻M:[{qr" Ge|Q'h wG8۱{;"HUf0Nᆚbu2+>ss'Wsr* N6==,3faӹSf38}U򅕗R/W AtvoF۳9.Cp eѮP )BRt%($VUp|S_Nil{mSҸ? J|ϓ0nu;Y D$VRLLU@gb=z(H0 @Tb +Pi"g/P[Ah[ƍS(9; GAU4L|m7=DD@z̑de?bj2vP6"1: AS9!ќW#TyXU^Ɇ87\/6kv =9ֹѻc7E=7p {0KJ.Յj;.Dߋ*0UVg1KؕP4S4Z1S۫,p-99U @(- io\\ILi#^B"J/6o}2o \JH =5rd4d(@&JH3 hR+h%/beaJT$%/d%6T(n. j@L<5 q= DVwڰhwB^80?%] "qyUR6xDf_RuLp_pQh\& Qkzy~v/ ?Gdmy%I-MF;~,? -vy 3$PV\Œ' рo-1/w_Բt=?2W4!^߬b),0vvzwI#ꆍrSXx ~gX4F}n7?SȬꕚ4٦ ͳlKT4ڔ6.2!"pcEfzr ɈC#:`=w.6tn{]֤~&Ctcf<6ێ틭㳦Ce`&qX;vƋGf#!Btcd7#X ~y!=zX4Fmc )8'TY3rعKLKuߩڮM/_>>oa[,qHhH6s,b# x 8|% Fw4(DĜ*t)Yft=* ZKg=ڎ3߁xF>nԞ9Ȫ0@]u "<bk>bK @Nپh9;v >gwSWh!âр l\6wJ,Pr>>_wx|HƆ`bƗ·3RD޴%#/ӏWs@nh ߉ewS;ILHb^7p1`u8~I[,&!^%40z/qd}IsKos4Z]7_-HrznKhJx!.:UrojlA^'{d˘{;ȽssYGgf3g-,Щ8ecS iY|ޞ^a00H" \S #ڦV3pi-6a,Z;;gvJ3l+:PBnQ)7F3n862jEAhI#n6͈D}na|Oٲ->]ߪ;kUW'΄ÁS Wz!{fPSZ[} ٯ7Ҭ5P6zCsKyIvZLIw 0xt PwzqYn^j67{[_>eUx𢡊g]%qt^_rtw Tr~-)]/\X>s?04SOV e`+98?[2E>7 :7m3\ϱÈ Z6980g#~ ;ǻ%螓;C,@N_C0WvXcZkn;ďOBhg K+Z!ieAZL6f F^᥃jUsTuTZ="䡄 tR{_R xb e*)C9ifCRKs:ނv Є_ÙR0HCy.* 'm6MHcX Ȩ:1(rB٣w!Qd0 h:YyU_3wVR*yf #XFVH@94ָg|vRI)vE:PI D1'8WnbhH<JLB k=Q`n]yܻQuXXA_f1l5`?Y4jϫ2.<0wo~cSnBs '" ExSgĘzfSN^!(Є,cӼZqlBMp](ĔJ:[p?>.kgs+?Drmb= ^0jnBlb悌C<,x Jhųo@[qGSCOb(|E`1Ճ$|[ُz2lF;co gbYl_&"aD3)TY s) YqQ`7rKC^Z$@yIHFd4Ӑ]%8>wJa:ZܛrP=r.s8@ۏqX[ug͖JAO){̮).>g}?+0TV%fQR\cK\v.q񺞸A\g"+/"HrGTYgwe(רT),5W ~to y5{\Z2rq|kJ~6AWk^ZEYQ˺zlQe8[SrU8 ͒Kq]r7W[.Btq_.Z];{G>l;wj'Gfq,(nX29RrXh @sY9X˗]98T, \d >  & (QW_Z*%Q_|\Ri̭yv̳kg];n5&U@lmc Q1%2GX XBa9mq*Z\{k̡1>|O>9n=aнP]ÖSn)RBb!%I6b0wc D ]*HBDIOI n7e))A  `J{PJٽ/%u ¸,84~md ((&o gcbQoPNP>++)3LPdea79YcD,ncˏ~@P5K_W[ԅg0{Cr16  p.b EȔ9W~(v>Us޾Q4 ߭W;/'auLق\B(xGO޿{EE|W?]߀in./""CL+hAgO\|:?f}Ɋ'-.ff˿-dgnP0Q 0gE]ЧOOn..*P ^ .KqRN;V\L)tG_| ${VbIwY<UfR5?̗@7ʬP !Pj`&a uJC˱3Jr09όN %' ZveUPރUpVp) "L 9'șQ:r+8nо/Fv@5:TN0g-AhJe"ޑd%∱~v?:D$\0j ʟ1}D)9K**#="N"IJB$ņ9tjQDJ?zVp\nu'񓜜HcˡlET#.H>G% ?+2$wx]?uE_bcVn#6[/iwuMnƒFcRw1}dV7>_ uq[/2J`1z+0!t#iLsCQK8[l$rDZ(ţT {EjWf,,?X ~ax$|4dMqN5JXm.j}!H aѧAX Pr,GA85J7B1MyRt/ =/oӟwywpzwuxbf ﴇʱOT.:Y}^_lJ=IgŇxzk_Y^~m P9k`+8g^кI&dHIkz.uGM6CcUlSB7iiEeO\vJlrIbX :G r}9\)KȜrr(UZQ,eo&\StrA|KJ1ps9\m _骷L$!y,moZ 1 ϴ Aqu9l u(6Mo@is,s;`gZ*i0)Q{!d䖑\# ^(s@`5Q %^&J > A`}SF OFMʆkR6Ni@F@L/>W0]'Am4mDԳE^c g O7yПT55Gw'Bv{X#DUH#R>\Wx^}S%yn(AxKkTIyǘ!W$R;%F £+Gnvm@nSJ6 CdGDHXRW,i|psp{N_ޒ+^_OחNܳ=բ Ċ0qeE$)52jEj:!+ȸlX*k8:-Ռ'L%Ok _`r4cm2 h5{(Ԍ.(U-%kψTϭ@܀7᭰\`I/ \qֆa 3\`UzFQ k;,Js[Ո:F'r8A!K9A&E%BR({hRc+0TȂaJ)`F1e?Tf&M"a :<K5"3Q* :S4$1y|ќYKJZ1|Hhcy±K9ėDPz{tQ*2EW*xjcSŮ1 ,ZX ɧ8n}ϊ2v4):#I tQaAAjq&MݺdŐVu1)8A5HnMp{Xb['9DjO onwv a6cSɦe /5E4 .?́/8kW)\; 7w/ \B1'XfSN !"ɳo*Dr7TwOWfS⼻^^E䔾+EazY2_|jPC%=/jNCjR9 Gvt:jrU R"{hؿyssJҫ]j pDAzf>g u:ntu8v uSXi)VbPuڣ q q!%n֋QG&*U>Y{`K N!a%rm^=0OL`7OӬkTCT&Z%}"Q.zE {YLImv]ri[>ӎvڜ\/mGjpl=&&P=EH:k2IlYWs~Gn;mg՞Yh*_=)GI݆F>X%|]xpbqA r<q4i&x$=Md6#o.6"RApt$s:dX CH}fR^`ٻHn %ȇ"r$ \`4aZKѪ5-i=6vGj5bU ϩI3A_f1ֿGsR1^B9)vSuxGTJζ̛Vc'o~rhi&_Ml)-N^my;!6N?Xk"quRI7mwiKQ>x2muƖoDu8[\ݮ?߮Kl֗?[g$SE҅مdL/`.Tyhet:yΕşNW wxJS-6^< Z|҆uhFu67˫wh]~s;{#gi=0BzI /r4m<\%FMՒtbݦc c/fhތXAӴ\\fv|@2 Q/479KQV(7"" Q5e'(+RQ۰BD~qJ82\rC%'d4)ņ#`t~1r4b !&hPg#&G] j:>C/e(19H 3jl{Ia |M=:d ;Z_j{Q*. !~B0N[ڦ5@-^B:Аz.I# +f.f]l<շBbszDKFi=zZPGP2;\ K >X&!ܰ|HHɽ* m,\=K5 ÙRSXI&@j}m5 >ɏjnUa &)A& )Xxs_jn+YZ<syhq8}&aGN %XH/h'1(2@ m$S|="vၻ@ r NP^% K+[6B6ipq{|z{ &ŴO?',g]8]-Us@rwyにנ;*=[x7ZLaB7g:9%ToIE)3czuvj촗 [SD-ԉ^־'ItBpcCъuOኙAE .v%4 JV*0S_8zml"aw٘wOOd 9PQRpTM]Z0 y%ڸL\K'0GL  $US:95t+ct"P#f>t%WK*x2T*y(B5V Jhݒ 2U5=X'CEjWQ bG6ioeh CX+<ڐEJ1Z:K Z߳pXAumSЋLR$|i)D6$+# >v?uGwhY+7*&\\nл tj(n'N;*c{jFsMƦsл tj(n'DbjFsMM)An?}߆-VH(obSo#a&˅e ǕLݴ1Y* @i4fWɟchatv}p.O s?7\ۡTv]3IL`YJqN+sQ@]G$jh'@.cL`ܙψ$_Bb]߼-N+l9%$;G .Mz'ħA+мjwt9EB8+19DBM|N{zrS>LPEf*aX#ӥ90GsQQ I,9#󸽇S'v-Оx4E"THН E(jj(~n5 뙀[? zLȍ3LAiB'$BXJYh7Br.g-!.ׁq^bhӹd#yC50H!AlE)'[N;xT?)X+7*Rzw#9xZN%tWҩ.\6rnǙF{ލ3jec:cx"sɼ[5m y&cS@REҚoyOe`D Oϭ$_}ҫ~tSB߮Շ':mX:M`~M0N BB^yr&;ZI)R JLh=r\N9OYWI O>KT yY2JɐIV!u:z SAd&LHo]̀yP̆:ܜ:\!ӹOAhߟ!8AZ١W t/ 󳼘\0K6hp/jxiGM|L\qv;cMa^Q]Oؐwbs?ouFWQ,~j!)ZMyR t}2u5LҷG6dOFVDT+=M:e@+ Ε9&=^n62Y(@ X.*R6mZJIN+NJJEcW,}ZPj,e}k9ˢ,7DFGXAY p3hjP ^,0LBI'RakHm@AJp 9M" Q~ lAmYZB'EI0B5ݐ2x5Qh&E)dѬ\~N슫?D ~xh3땣7ݰ^l6Fk$Xckec6E䭕0q5[?itc{WGj]"C !R郭F)Zh}]C6Jqexfkkd*{^I͔E^k}2Ó@ƙF_8;K ɴ 6sBe<SmANT5 3$:R`dJyzx"2m!jxϒxZ[}*ZwB^Vtqy7f>Y),h1'OЀ0]V_>9rݱ)iSЂ>wdOF0y\%a甦&So3RQu7=Fj܇t]^M&pŧ&R9tK=yPrCJLs:,nןo-Ψ}X?_sA+ MF&}^<~bc[e:7owy.]yAwSlػ_n6?MٽS`%ִKvpiCӢsn/k~nn?\y $] f_$di ~4ϫKϭ({1\'g-W Qk1xkZi/bZ+Ɲ TK/~T{o kB|q󤴐Ρk|hMusa$zƧIT+ {VXo"p-H? =r#zm$nG̋{OvcݧqLl[׈1"%)B~nET^ L•t:ZA&׹ު8j?~7w{3ˎ^|m)B B3?z"K "P,u][C3Wڌu~~K{sBu6_fE?7\YT-Fc,ڏ6SS}^B-@C^{ UU"yI0thZa<_ݓ=S7\+Iw^`jk nOr(N$-DNypG 7-ʥ3**QBgFZHHƙ`Qr5(<h(Mnnmp>.$maav3Dm(EFjB0Ha@CcbU|f$D<-&Jt檧[ӆN v>Th=^G]Ġ@_ԼWPG5T^A )g:!;mײmٶMl=Ygܰmo͞o'G '\ 4wգ SoJPJ)q@ o6ғl³ү,뇞n#^R]ncUM3jJ-! k+`P2eF},tCNʌmB\(1څ~ET~kޱ$:ͥv{gy쮽Ke*lnf x%gO`T܏Js)inlUat~Z^>>1:_=2Wc߽x_ p^h/>NT2Qם:{,I%SH%3s\4 4xB4*<%vI2plL$)׏R\?"c0׶F Z.mOxnܷ}j6~9gUA lhqXQu^@ϣB0uműFq(7tfFgY]ajZV}x;?l^{mHd?^50xz7 C鹻mps{ iy~aaρ4ޱkU-$c\}k~(kE\ D$6kq-;S%`,7|!EF?8Q{ ;N+2D"%b,v36҈v5y`_ njO8)(<ǝJKfq\h%IJТdBAk%/.hyxbӳ&5UY@Wd,gۃ)d+kI~,#@e "I'{) kc)\WIe d)-,Pɕ-eD,Ν"HR#ɕ$hIi6 |e5dH>W/yݖϢ޺vshqf*jIϿ_:AH0Z<|~ ?~ %%r;ml?egȆ@ӵ"MQ:͙miR.!S}tß<FGvz? %1' 1JZ]ǫ+\+#j=d2†)0ϹEɻrQ΍ƅUm8#uĴ jәz#~u ˤ/ u ҏ;}ީ>5@xuWl>ބM^-QFmUh اw6hT YHB`FQCzC@Bu .\ZooOdA-(@" (;,S=_fA쯾 /:N׷7g75(wR6&1d n'Ź;$c}R+Eȕ2,bhFj][}2j ϕOGDfqk@"QlFduȝS s5ջ˳[LI`|wa|Ww}|7ڑSoBNAA_{9b ]rlM’hUP1PK/R Z'7֩>+Qye!)P OHÍ?ϓ/D`ֱg/YZ=p2X%V6\g.L^kMYWX<+L2c|Q&/`,3B́+L2@5v +R1 * dhs* :3Llea͊d5vmkG-7C} 5쐒}݊&\O4J)۲?4jvTspyRt UЃTbX ArӨhQTQ PN`7Ԣgg=wENH7>偝l9 "yRe7 Уt ʢ)za6.nt khQxB dvv 'Pku7h(Aw֝{{{K5陹YT+`F x=7̀>yr k*K`Y3"hmF$Cc-s `CI%I3M k2A)7C} 5'+WrP06FEUKC Ky/6¯[*LLJuFU[Bt湾muITe])7z#A=Z^rx3z\Kg2Co3GO8&B/1^tT,[·jw[VP Z%_k' Ov] 텀VAƖG_>#JH WWP=lĦ wX1WK9 R X_liT:ldP0EB[R!k:Ra4f+15V7>&QmJPAT lm*ӱ #fv_ 9iZT#01E9ZևvԽ.9A6k}j)2ZH 7ղP|-!#!Jn4 gZev\. N'P&]xRncAwךh^(YcKŹ[Se%SN⊨= ilkסajj ^ʩ$?@H&u[ R9ѩȥ j0cʦ4Ć 2&¸+5/nm?+Q0к.e|L-#@5o pI VzYqS~h v a7^p7nHSԐ$NBO7y/RlۼLn/ueJj['qŝF3Y+}~{zxrEf(OazMD7(S3jE$Yѻ?2$^Nt##HtvPy:8;^^z }-ޙ# ZwZ Z?icЄ켖Pt|T^:¼urɱsa.&a/O[nZلEs^֥6b2NA[ 4u)P8ּq=27x]oHp.%Rם`yd^F$]Bp]@y'59:^fg,Z%=ɼ3hr>-h2OB[ϭ^V" =/% v q.22.\2y< eUjVBvZWeM=YF֟1 oϝ>?v^ݮvxSgn>lSfa .u(J[(MzA69aO[=FlӈaR[T?~d`3kOCEC]<=}ԡ>π^~ێ6xFn6MPh6Q24'ibkqҗIb ޿=l tȱ=;95y ;9OD`Nސ?ZB#o Vãi#op]Kn-ǻ ;t+r<)b6 !gw %DCOPqHJdKuHޥgARh-EWˊȄ'5b5{'2h[ct>g磅lH d1Z&a)Jt%iOL 5疎Z G)%XBbM(ThKДR5NX&`sBH)䎪lZ0+d7R$ʘJs! %( ~;U‘;'r){+bնΖER"TIid{ck1z?9w\FHjO)QO  մB>q,zaԜid3G ])s?"Tu. gb 9)[ IEf r:/}U*[q7?ϱ~Ԅ5YO KӳBW yuSͤg},p~{4PvFX=$-t}':/q;c_tDi?Ϙy4sm:ebY?7HISP'݊qf C먈|G4s&!QٯjolgO48Ia͖a?rŅtq3n,ZZKz_[.0Dg& NLM& BJ82*Yʋ~a幓ЦILL&?1,bQ^;兕ϐRa&&xA'䗕aAsʷfjd=huiv}-j##R~"rՒRx(ރyfԀhx*%js}@K>IjYִ59dJM0փT~TP“P'fXQ.^##ԳI| FXaLyD0@s2v^ݞa+vmgNrGNaNAsxmg r^pcc ;<`.V$ۿ0opMLF^<o-ǵs|ps<+XYcǣ\;b,ĵ3ǣ~ i.˼y0εs>-˵3OuRw~Zk|\kgwUZں,vd*KaTfZiq>Lk'{vL -/tXjoҙ*ڝPsۅvODǟ݇#9e-B8vfn=Ұ㶓FۯwEszw1\! M9jRh/Zg؂VrXS?@',5o 4Kndo.hAs@L8A\Na&2db1BDj[{#}x޼sdߎ <|Q;덿&!}7ޖUe_q9-E2>\,C-uE̷!ù],\mt`؂VCzs[V^p_n"B}Jfr{<.Q?=zڄ$\ ͖5H 8c!{S<"`f(A;6*ɾ:ފ?T1UH60RCq| )EvA*XȪd9 { BY Q' ?NCv6 EQD8=Fa4]G)Pxat!ƽ/b-PtǂsZg_"8B2saғ?`}# ,]n𣫆ΎE돻kRgǢt#ת~0J;8]<ÙZg2?O+yj^=*, ˆnipSϓhL! W##<Sw K[a%J-םfDiM"i@5"#ve{e5`orEYyf``Tr;cO=Y^iWvgC[řp>TFF?+$q I(*QEPz]:\.țϜ}?ƅ`/x˸ǸWե÷'Ʌouw8-JƷxfܫnrN-qkzX]q1fa'֏THL Z*-@lc95 aiAvA՛JL {4hW `j0URBT[q/%ZRc_95qE#U))Ř1bW"ٲ Hc-Pހ r+ +8l ?檠z,[+Pa70bB3hA ?9cQ/3^BVnTfky1W(_O1׿pOGKs$+jp졕jRY ^Z1W4^6Gɔp≽`F ոi:?+mU|dR]=QΞϻ_ׇלSsƽ~;C48A-,얽ѯ9BѝF-a X_p%߲,O1n=›y_ ՙ̙j̃;jk@gоm=.,[ 㨰OTGbpdOj3뭫1aMٱRoZcq*sΔ Ԑ|G ycgx}!k<TZ0b,.콱l=0 y'Ym>Kk/20z:ٔ;1tt`ؑNd-R@2bXO ;U]  D @{637[Pg'+~p3urp q`g0P$` ;h!@1gj"lY<"`s@@3]LRD'PaW:cs~nB.İ3}tOxh&>jvuaIty?+7PS3:oхYƱ#׿"e3&Ydx@lb/6fmaǒ<3RzbXԾ}mu"YjkzT),엌[9C}=gaCSa]PFndjy"ZVR,2˛ ~.I9[}x~o&V[(W-_"7*ι`pV |a~aOwMW4*7rZ/{r+ԟnL'~[x&_M}ޭ/'xwS|&/*3bqX)6[IrY3$p,׎VT;9V9,93*8|UC)i;5 $IB2\{aso)PBA8@ /Ɩq',Ρzͥ@hh&u^UyKWs mdqQ:d6GPBjPj^P YQJ !ֲ"/sEacRvP6˔3_6vh5R<'T&Rs `hՔ0y[v@Bjih.5Q55򌴫%J㑦3B/t3"hiBjﰆRcIh2 3 #eNv#L\xUrA{!A K\fUv)ev{-~M_+ko= MFExxnB鮮waQ,)@֎?BrAjxm=W[ilFK5i&F|ZOR|y$+Yg4 ; ﯪqT7՞)gj X|DZm%,ĶPr\0-Bɏ~,ÈRuG盛|B(I[gq<_x9VqȥQ{1<^Lqr?7<`xp=xh# y&ʦ`kw![[ĘNox])Ԓ,n)%ޭ y&ʦGqx7` t~#Żsa[{jwB޸ٔ)1 a]ϛMsYۋApw~x>[_\﷟~No7xq0ރȹ2RjY&H$An ۘzqNhSHJޚR[R"FdZmip] j%;ntH9 Y&Uko~&%W3"$6u H9q{'X܅q32U:߾4{PjŠ0T@hZa*6f<01ǎy)2xmϪ[^rr)Ld.p5S$ (9S77sETm}i*PҼ 7ݖo'eHڹ Ƭ鳟u 0{KSw#-Js5XČR?Ai|z(/]2qsN&%So;wlL>xX>>18r'#OG%m9L*"m4 "s#$SrOidJ]aKA! {9:ȶU _9k f_oVCެrEd?w{[v}א=B5yez5ڹަPF*FбX {..>_=ju[śM6~ P5p+j7P|G`u?7,?Iv-@ь!|:dzC_'$. jHx^먀GPy Z-{n G蹙w\L| ^\x)@6o+.l6flƛd?(7ع+vŵ[vGaMr,nc4yZ뙿J^ٯI9bB.)'wN[b$gQV̗9P_625ʆkWP1_oz5;% aN'^3%f;}Rqs(N|A'7Ϟ7m3:~/byZmL' P(X@MZ .zn('P9ݶLQ챮|Zn(wȯ讒L*3sԤ‡uއ@> )hJnD JXGާ`FHU<*} [)H*0sA%asE$+Iv:wb"P>@ 5H.W`&~$er0BP8=rnrdΛ6QňݝO;h-5$nRCkm舯V_FOۙц5}е pL֘[?>0Z $n:ޫ~5iʁ#h`s` ]Ρ部zth7)#֠舗}],o, 䦠CvJI[Us6 UJvۘݶouU}D @$vH6kChσn]sc(sg ŴT.PR\fVƒ2{ ˅dyZ KtˆmClFD+1h@ }vfs.*cRb_fQeOR8T *(ZwIKI^o/1ds %&jvy')1i*N/fudd)(SF@2jNh4_mwg-#fltw4F݃Y&ƧrrƂ Gihm.~\ŏz8$#L ZF4Ψ If8h&ɤkA MCۦF2G9Zi=xG0,אlԘ.^$DԘACޔ#M~N)%﷝QѵElр]f|~s~ۿ{JOnӶFy۾ XAJӶxG'c;&͘jkiF=]7-Fd8w4yjd70~JđЁF]&8KO1$t}-=fڸfC 1T3>g'񐒕zw ge oQ0aDx, `K"GEEH$PljYCQܦ J~3M4W$i{ VO*!{H9adR i)J#U ߽:KRd̳BcQ{ϸ*̈́cBQ:~Y 5TwFjEd:72 R+/ _z WP:Yab*RQ*r/x/hV!<Ջ$Frmax/O"mmWkgEZJvjz[[x>עKxf뻾g3z=Xc$8gBz* 4݆0K0%LW_cr=PJ3~ӥVSP7O7Xg[1>Z<]Ww`=>M R:]bd`$ůmc`۬7;'V{u<:#@%맛ޱ',k3 Zt 0a>]8ugqX?] t1'BڻttuF8uy{M1@k7%4Uסt{>85(R.G.vL +gR eύ2{1D CNygř[[/j4ꢲ<4_ 0 Uo:݊1v-,eEn2zBqxD¡JK2 EMb\[2Z+ K[h\kSg2s>xř^BbZ>6VfbnQ>\o Ո,Y* |hZիC' Wh_OK=/:ig!#>+ޠyt?ݣݛoiOe`㰌 Pߺs8 yc(~N؁kGKчEa!G<X9H d.-\{M廝[P/6lP?xȱt8]c4i#H׈r 1 C{r{a=α=(MC.ag+d("}!@Ӡߒ܏}}a! {!=a= KNaL: /uAiXRQLdEsӢs6'8QiH2ދܡt!aF!HX"1e *NzifE62|ϣTVѯ )IcEƒROXbs*ƺ3!pW $]WWj fv!5{gT9r⹈TSb2TGb^ L @xpjG6N 3@@zK$rjykU-ozüPL˒â7/bC{ء@iܻ&C. `pVm-}xx.\d.]"CBBۣX@kd_Y5ڴ-oOFYUY(yglѷGXY.7cd>WmLlO I\23AYLI ` \R0Ka ؄x6XfO ( _ϧibFamXM-`r|6[Cm쳱hƘWۦ# m&|t5;&B_h$*$j%c]Cgb?ju@v8 PZ乁I(c 6ȍ9#G z^ or5}ɐ P?kfH .{r}y{ò;vw8sᾅߞw(9FN@xv0bQ͜};SB tl|.!UNmv9W<([bSc׊0c#bwEßȌ14kΘ1wv@k]' 4Opje::4ꚟcqŴ V~/ebw؆LC|Hl ,}| m ʡ0 66Bj3s6܄=[vA0 PohT_]J#quR-C ي&-fWT.sogPl&8dqī Rzz9L]CƔ0R8el1}|*rc glZ}mAv2CR-'tW]ĩ|?'lOhW` u^!?:5= ^GLouM~λD.|MsLHv?~4 =x Kh?2Zʑk$ZC7]ؠ(0;VrP{&15bss5TQQ]%a}\꺢i)mdf&$ms]]qAL! .Mx`k{2I9Toz a\D 6uT'7v澋a3[ 0IPp r52&kCQ c#eV8ضPX%y3r(D* LaArEE0P)&0n@ b YH&ڮcC8TO;m0W޹bJ+}' .70!;W" W2!VV3aa|iYFMm&O\Osݛ>-gO˻Dv8obxm/qoz- N^~0keoiOُ&J&UYZ{_!zAA3!x)uJ M Ո'Oc .b!nu"3xݰRY1I_[0M/gcks={Jie{vRX+ 5..c]Λt飬 qFdd} U/}d,9W޵اʞˣw$:! A|٣} =x`MEj,{;1%qY"qF o4 m'\Q8*:ǻ!@ (]:!YI1y ]:<W.qA.^̂\No]6[t9iC9fu -^˓d$b9 ЄcTPI1JOڌv@mv*8icL$C3d.)R&ՄiBTBPT(hbC2ÈWWK4m"ܵ"mlH*%EцB H J̘K8%45*YB4"ь'"7c+ -!hE!*XtE/H NRm7(:*$|A߶T{(B6 L:l)ϫ{9-?GccoZcT9&8e^6R+t+0F=n⣻0"*VвdwWsǫ9Nqy O9BO֝[/F+ŀ8>\wYo'xT6Pi []&$q6uXnU]qEέA`{1 @`1-V|jy(WG 5dkZ{3l0f{Aae wjP`ZVיA f* %kMe(Mb\%eyT! ip*%~SRrx+B+f5Tw=>HʙbΦ(f1!'`cLQ\$X+Jy2i@8iFQp(zJ(n%F3ryքĈ<=^PoC)I(O7>#U{]ǣy4_: s* t] Nq:` ]=3KSF1]/K^Dác:Zuu'jzK(Y4]}e֨(z Fؖ= Ǡa-a쾯$J7k3Pf=N (^kޱΎVo%UX]Og]]'(_?HN:el)FV%uB&=v&g=֑ňq#av~A uk(eWD;E8 20.&چLhk,_W@b\]۷Cts­WF)'18sJ`8ye< Ăؿu’sn.%b/vVdB5*S@9J0is(T9YdY$2 D4E 8K{)U*7S)S #A`QbJrS(,LP,D5cQ0PdiR7X;AuɗI*1MK~Bۀޭf31tɗb@T8)XR{iAHn/Rn[Pj">$ȎG<2 )A[c)Umެ?ͫa8#Z,;i;Ӌ;G^|%pq_{m9t?"94}ipi HctUy `sKhW/S_~> M*#^7.TrԈcμ!.9[_>_]\3>&);]jaFޅ}z%u@?E F_Ëj+gwM$˕'|s L/R{:}g%!|j~m_onwOAAwlnó^hhfZ2r̎TSTIe!哉>z)/QLAzJSȽ^V,Zwnl2:OnfcnMub:M{Lewޭy5{(~={l5Sˇ᳚tC'?\{ر3&[.oE^1lhs(p4 U`P^gI|ߘ@}Digvc]ղ<޴VY2ap$h\O'3~yX.g!D|?x+2f6R2[F;凑ruy?f?(,W*^E%]{uFw7k֪4$LG %}%$K D0 |'Oe 27]@ JNWP.1xmJh?bM8w7i)Lq؛*gDKs)in l>D !KL+kH/;nUfU֡=xēh<(WdS&} &>YE3/+/M(뱿ob}WŢ pkÙv@h؈lf~5Օ1暨!gs= o !ɧRFQRRrv)LxR$>Iibe$ݲgLWUzr$Thgے#Aղ=9d \>DTs˧,\4X+]GӸ˯:BV>! = ϙRlѪiQфF7 8k}0~\#@Xm`NyqmIٯ/c)$.bhlaV[y!cBHÐ1gLR%YnK g`{"uF>3~9}Ўݽ/R"QcggUTъHLRwd|C?Sx-=|$۸J]?;Zv=r`ƸŒ qLBxưj'.z6dt ܳqt,ZlşB іN HxNAnXZDsEZ裡 5RCrMɫb >Fj])Il4&NSwE/j y<:8BjFm *R÷%Rk8fFf f͢@АW1:urkW]&`mic1QoԱn N+iڶn[U4JwȻ/CnN7XN=|cg-TukBC^n)1C.SÍî5n yƍ~P2pE^4":(/PUja8w#0w 4Cuעo,&P^33^e0hRꄂF g`3.M~@}1yhNVBA{$䚆 W|b0k](Z= Rς,[rz(%8tt,Pe_d(hw, Z?Øƣ ; / ԵU@#pV5 #dEP(h=4PDZ*40lzL] yEk=w31֯炍s :jro6y>5(ז(5̼&J-^m Rs b1~N8Ah"v,k; ~NЄr)RױnEAKEbZRq .J,Հr)ܵn$[,!:֭SJ[|@քrҩ{3Í  z> }!qs3RBWTE_HrKwnKdo.lٳ\xDifU`SK륽.nF%O_}hb.ݸno~Fry~ 1_xoVnfs,^tw/wzOۻ+#SW/7o c={!;FME5K-&SOm%S ~as6fl'jon˕̜uAM  .+Wl 5A&V4*| XYĨ _ dbzQx<)3dFHzK}bd ӽa]ʏtm>~[;t|8XC&eh)Ves<̥$)eI. K2 w=w'~3>-6 wX~j"8+Z1+7#Z96r8$c\or+k߯n{^'8&_m9+[%oMEDVן|B^ {@~E.4ݿ>,Yʷ7=!R7CVou ~o Ym:d loFC T4Th\[439 ]m.3:WP8sRJhH& ;\m5hh2LK!*Vߧ|K~>G- j%X>絔kcRoK->c-ŷ,h/O1_Ǎ)WSALNTYڢfGk9 4#JkQq23R7Yb3N̘R$ġo]5*Z8oS]_L3wv៟><^L̿oA8>YW;?xʢVFDK!'B5*8*5cJkfZ!<Ĭ[޸}byț*ao&EeI45 yQ+ =d+ r& AhƠ!7:_+ͷ&5%*]Uӕ3u=MAáU%Vb4QU.jfV҃Ή="1vW5J-B%u>L>>9q'BaӑsAM\kv:?iV͘>H5y*,8(R%rV1Ҭ1:Rƫauzm`15HC%J)L8攩- $2 !\w)&y2sk_ O:JӬe5}ny{G5-y!. i r\F-H!R˫6HBFE'>ߑZVZVwհd;ai,A)~ZVlcyce>l )ݯ1\Ms{;tyQm(FMrRfiQit>=T*!5@KM7jD5ZaBE6 sEW%NU{(1wJ\PACإD}{(%QJ̎ å_bUlޟp{Xy$PV1mp8~P1#׶-αձu̓i`,<7._t}}qvŻ[|هrlU*).9^/쥛yp=MyvL.+<Dbi咺w DIZuv5(>Ȯ/&?Mf){#B+Xel?u9幽[x;t5F^,VֽuǍN⼲$no¥nu}t oxPW ?Gkl|U-L'ߊꛫ+ȱO1A M"5M%y-˥g#\ m>VVᕰ ݱZuۗ͘1aPˋf%hܺNMo>7vՒV|tj\ F$l6$fٗ :|U)Pw<5O.uAf|-Uh#~O{Si#yx䍷P(s ǫF)LDx=-#9^vEG7) js~KPKx7uZDx*sĐ9[vXWI %V(#yٿR[~_ fpeI 5iI4^Iv@eNnO6?]gP&ԯrZnKL(IJxw~gSúc DY#E"g AB*3eP! H1 ̉ IwJx³Z`e[p58!'IurÊZ5h ,1CYIbf4]Z_~-|YEJVӓnnE#;!osxϰ1am7?Eŋk?LN_?ߟMrMOwe}SZ&_O.Q(T*dg]ϺSVLyF0g5?]^w>s.y0Lrl] ~=^5H3&`k FlDK}#gl$bĢq,JFcZ&LKK?ЏZouT 0[RbUYk)vХRaZZJM G[{ bWjZz\m?Z1LK\҃R0-rW{QKr RNU0 -=&Bgg„i(Oˬߧ ZV"F-}~Z*޲b_jRsSI19j)PBk@aZ h?Jfw3Oː^;CIv*B[ #c}Vޱ۫e S͘6Y'8#j D%ٛp}UG`O%b{Q=@ !صwg5DԅfUq%-/s2a>!b/%E>kok."Jb}׬wm+/af-3i%g۰t:_Rrbّ-ʢd9UԉL>!yHU @=IYyZxłb_yCVrp:k6ۗ ѧd?U9IQ"h}AJ˟r1J`Fnh{K |[nU̹9Ӽ[.Ls/ޱ[rL{ڲ/2I:'Iq ºސ 'UBF.c'ðOp?='.]@=r{cct.>|5u3}ue5@av/O$VVŃK?<&~4$vL4Kh-I# mtnvV&3"ېɵVx3l=LmwDGotwC x5zSGړc Ŧloj4@lfr+osle{b$L@bGF76*$9yt,-d>g9;Im ]v99cbW腖b駥xҴ"?-E<)LjI-h-oaࡥ1$!Y?:M'F'ߒ6׉Lb97"ѝIjB;=;cBVt~w(WF"lt F!&a #5C:1 *G< ;|`HTh3 >o?_<ۜ1IB+m|^/.>is/3I0fda /e+$G.G0 0P$(hcXx19\J,zy=;!euKq0xalL 0qjPۏ1"!F4 "af-b)oJzy=|ԛMwꜶ7V3b9 $i,ϨeTZ@6ҩuU H]0+1uj iNwKM$'NIjwOO[sFƽk9(A?7ih葮޿8 N͇t 6F0%sdfwSfVrqQ:)őh`D2":e?0D b}Noz:7v\pyrλRg?~\R3 uϹ-uN-=&;Ԥ8;ZzZ"5^h)~ZI)#^l3q?7R[+ZzZJ=±?2~brNDǒcM45jP$8@ +ĘB5%BM}=C%X?cRKMغ O]H%x>!ъKAHnDD'11@A*eHbj_n_Q %Ta$GɄwxҷEY"7@Gv؆asP'٤W02Zn"G5';,d-e3'YNpOY'ߢ== C"&w#.9ϕiq0CKIERrl$[6>>cc >{\&[T tNْ?9bIRQb-hBg)tj6898,@\Z{h{u>~Φ[yRcYz*UDM8_:okHi֍y?]jERcgZ7;]jg=F!d8S @>1Ra-e類}%ۉ=偊I>gWyȝd[gㄲqY.Y)Vz;YT9?Luf3",u65XBhw{:drb 4XdJ&7u1g Z5M,"a-A$bM"C @.cIu,AjIdlUC b06\7_S"d##V2xxCDxxx5W#SŀNU F&IeiT&H Tf1J8P~fEHi=D1$՞u.IeJԸw:W+oESW\~d!Ӵ3x2$/-Gwd@h! Sf{ǀNPww֋;;.4rJd/0(A 41T(5JIQT'RX >o/?C֢K p~\GKU}a.2 DwMNKkGIʿp]Kt8NqyPXwln]g_!3ɫVk(J P]gi"}pK\Cs3q36sEjkN*v-s889Q"h!C^ 6; ra0x4cbdILﯚA]vS<$;IDfXky 1COyN͒ү/$[}2IG0־tq?_|;/f|ҳ.?Xr*?z5 b)bpa{8; OMc>Fp՗ i,۷[\з]4gb-z/t@ P`P }˩ut%ay#gQ7B&~^ {P}` HM:{ Ÿ? x*DӞ|%^dt̓;KiҴˆJkFB4M4HB s|M^lsChB|; 7G@<0]H~e.W6o'<SwA|d>fl+fvkUn[H5T%Y< 24i槩?>4 񣉊_LF/&l,} ~f6uqx ~.eմw)g囮RX?s2ox6Y0 s+E2Qe z=tу7-X✿֛q-~Tdlz;}3c! w;F/`67R_|y&iWWOxĀuu> q+t.zKBq}LZK&ۮT +qs%R8WaqrRIy-Z(=k9.'eW+cvz9 8X]F/@Jr)5V=JCX%Т:jXf:BKKgQ?q-f[۶:` ʦ9O0i$z#qH‚(02 1!T@SdžQM&!8D1xKV 65i1\[»kk؜@<#LQxt/9W^2OIDm\X|dN04o߽w;<#>"N)wA7$4u1w&ߤF6 £+ Ţ#KbJ$ =,`~@e;=`Rh,f{.-%L+I(8? o^97wW`" 6F)u7J<* Ѣ~yNE6D7ڭh*R^VAt>4B80@#-tt)s-AUBXD"A&R o>JE70ENOf>XT f' nO'6yg`̍֎VP^U]v?4U%y[?4򬴕TT kZ*pD5>9uu>ܝ\[Ct*ﱺȢOAθ11&O|h]il(\1 {9<-(-rRkqc*.+@ڽ'4kF# 8BXs)B$ a$x,i1aXS?$qd.Qz^6R$!XIQl$J@0G:1,BFbW :itU@a6R!TBD3$4:Ò`CUqYeF1DK)5: RmB!C820@0Fu-Y\w9Ѩǩu`N'v}Cʹ>Hh6}h0zLKwĜ}hTzL[qO+e!>kԖ 6%Hj>tjԦBlИr3Z36'6Fog˳ԝ5fltcavsׄ\  nVtD״ 9nU#p4FvLAu1/^^bܑ\!(}fibU+Im2%&Ӄ$OËr@a8MEI뼛`9R5̓ݽ#1H}޹A~tvoNՁ\Gډ=Iu{0wj8Ah@n;O_޿n8@,3o'&,bYljyl˯zɨg罢_8'kp rᖸՄ˹Owd%*rp&̰uM I v|L7uBx#`piQCjyNfPHVN 83!XJu잨瞨'X!O¦P& J!8_b|UY}p0`ky յ5V GR@.\^;F꟯8:TFS&Պ 3Wٕ|+5$WgL̚1QɪPgM/o/ l((ƣI[` CS՚v%eבBxaW ^69*ާi ㉰4uvhɱѫܔOLU.o*FrS~ NMbOwHPiEuE:qt5Du\]AsA =`R%q 9h h M>5XOAAXNƶҶ͓B'% 7k;4 vԫ:;S(R7")Zt}_db )R"اa 9wWxY8<a-s!xLeXE,BCاX0H ;HI1lBU,&{_%x 6j`7ˀZ~b46_~ԯQV9Џ>'xjYj3>|NwMJ`3w"BzKm <9޿?-aY\ /pεεεu;]۴u+xfžÝy*qKa": =ӝsq7 ypƛup;bK{g-UԸ[WMeY j74XZKY1}! Fn*bSX۞4|"J$éR3tյOL@| b tj^ KrrT>W{8BSW[TkNHy] \*@}G"ӪBHygE@Eʛ\Dk'?Soi75>Vѩ*֗&dgڭAuGք|"ZK(?ys[R#jNwTn\EHMEgڭ~"Gք|"!SyYӡN[:'BT{b\z8Qf&] p:c?MA/V"j'vg_bd{|66P\D .2gd] #MMH_vE%3;o^_0tSz4>Ttaʄdry:_H[E8$oazMQʘ Gɢ(ɱ/6Crk/dL=C\mf8,{/OO~qoc?]ĭYnc?MѼggy[G_^';c_Vs,r";F YӇؿ x3w 3%RQ/UT.jlFUGFjH-Gj+_3y!3fz` !Vtâ@t,jzJsx˜yIzXWR ~,2 #P`)?RS YʨKuV vӎWRxKung ,/(5\R,XeRZ}RRK^_J9kK&vkf)K\$Jdci*5t,j2R$׀,'(5äRf)Rwc)mK=|&nt,jǕ^vP?RVZg+d_J5KKVKSqҫf)(?B#.1iKA4luy RډK,q b!12ȀUXj!]CI+FJm IRCj}PlU( C( #%\ &BmeiB53p(M %tl9Y0*)˜k$cnZhN`Q(b,hؒ{&oR392_>|#~>KuxRk8hAR}jz+}OUârmܣEx`~,FyaibuvfM|yyZ ݃u،7Gʋq͇wfӳçq+b(%q,^ݟO Klؑ{%(h ,a5@a(5sPa|\bZZWPIdåFQyKr].Io!n['yTdSt- %\sEg(k|c~ c" o^UrU RQ @l[f|wddj/@pEN9/겾Lǀ4;5GV^ "<@$/p}xk'lD"^ԍ" |7~n])C7ҳOK+vp> {CVElĺd﷋G{/72JY~v'LSJ7 8xǩ9ϴB# sIB3uwƯBxF(8i[ܛIj[iBd*P~D,rZvlKDmiucF|C&jbn3=(gc2]AvSxvуޛc#?zWA ~(?[ߒJ/`];Ik>θumݓ|m¼ogrh֊'bզEъtIe-Ѕ n'ыםEY'para9) ĥ8/:wmqVSڷ8ޞOk:x+Qmئhu{qyEfQ++sCح?o"ej&/5E,CwoX\2Ÿ;Kqo|_s} Sg2u^^w2əG"{4YbH͝=>T]|(~[ץLr7?Yh6o=u&kcg3QS|a8r{L- ;qs}}sPf;e 1ON:ʟkܺ1)`:uQź]ﲘBSZ֭ )ZSZq>n1VW>ԩ*풫$9u^Ӻ5?9E <.rg'#XzbnGF;,z]kyamTLz/7_f,} ^峋<jB"Eg/ݤU _li%"B@yuL8#%ql>41;-)ht`5<XڞC+I=,JUK6ds6^,CߗclZuCi0Vtâ]f1`E Q+X6~)5&8Ikd_F.=Nx+X ܏ uyL sL P1' *YR&HhF!q-!I0e**tb g9!ސĠ#LcF%]yU 1%g0̂oL5/]Tt&=u¸1gRn rRP"_D{^%g?hdgӍ :`; *c "3<53=gzӡmImV4͕nZΆx`>QaKj=߃\4iTv>:v| V9Ug6ɧhY`~W2H鹸VRr6|1ߛ4[Y φǷ~?|{^Pr':}d5~qhw=|Q Fy v"ygH:GD^GKʘW]o>y"V&>m+b童sŻNv!-Z:%c‰{,/8;z>6 rghs#D)UmB;h;u%ȹTֲcQ7h^ RԯwhwZ#*zgKR9j<-]&Ky44\?\TG{QߓHѷ2>{JCn…J9"niS5N:~JvϐTqR?AX&K7,LNwy/ЏDCL|]af3yJ$aBeQBMٻp5J(Eh#fFRZƉR3+L<(_dS3&l;+$@])[+_ b#Idt䫻=u#w{MLee'8WM%zZ܏5)u֒L_mҦ+jkt u![f)U{o>38çqZ=zɅMwtQ'sNgD ,UZzs^C($W1RˠԂD\&JV{ʠ (=@nΩ=KE13׌5k);1$!IF&Zϔ[֭fVȧߝޥі^$tA`3byO\(|q\CTQȼ{ݫeaLCivWHLn%3W^ͳワ-ctrL+k,c8%`ʹa 2 +|$;H>Z v<Yv=X /ѻ |ӽeCY,Pccq64f?߽?o6x=}Er8%`X;`i'C/^x/ٹ""^(!NHg*b; aga2F؅*,G qyo|`DWn}xr7?{Wƍ K_R{[tMU(WWwImv/_vK6m/N*!%  gTŶh4Ow[k ᴴZP9ͽ!Z;(%SPHByY`3B?&2N u)7*D(%*JK3kLT;1dρd.Vq"Pi- #$F(teXDBh݅@%9Tl;CBRVUh) P6Pv4 n5 :C!*.N eWr$LnO%(Yr,ٻuPD"Ws8X21}7P4C t_~Vqko?2lB yyL#4QSVMѕji4VH!|Xh=tGF׏3J)hyPMbg݃}C3kM@W- sUy]acUwHAJ>!ZݍZJ" M~o ŵ<쿟ǿ)L*8/Gr^E89pMV " " " .jJMYxw^1-(ʒBi)q;Ǒ zA BBH3%sF8,g;I]L`؟nnK_x[PnwNk/m7#1bľNri}o>~q̡UhS8ψa9 5 ]š#.0_ B;}o7;=PT@xxz#3T" cae?eJ!OTIXO 4Pqn2mx,aHJf(EiUso^Yt`}-̆!Ic!X1y=ЯzV1>KZ03A : 9'$sV](Ԓ=GMv@jHOobE>*" "Z}dh72H&T@j`< Ͳ-Ř2u,W=#<$ZgRWQw@>$&Xe sdkK8 IwzklXJ Eu7vZCj3owAJNL&zl;WnzpUNcj}ϕ~Z:N8PÆʲRl0YBt3=y0s&;,c@n{n1g<2Y( ƼdPYZJp?;_M$k2}HAѴޏɝ7d ג?v?LAѯ u ܟ{9\Y]MV'>oN˔n~[ueZ?x5̃5gcIyY^EQb#2Dp垲 Ѫܷgu#7|}N~Y۰DIC^vҩ} "v ֭+RT;rۈ+a09hАtJw7uAt}GudZuh="Ѻ!/\E:X˙e6(],ΉNo/seru}s]=87&pX Ws}z9/NW>gkg, i)uO`IuChdA~/*j8L Y(}Rhr iB.yZ 'Ō7?,垕BVwrK %/{4YX<[.uPvA:s|[ζv$zEe[Pt:@ͧ4.wF-:$J/FEk/Ñ`m#kحq;Dt}dێQ-[!7vAOZ$JetSrR' % )- 2#9% KF 1+؈1z#-RA0X9g )҂8NhRS]Jp$:Lj3ꑰ~ @YKEI|N*6M:,D'GLi$g@VH+RcKfݺ@ -ed:/z` AKSV Cqrd8w&<Z5!`Yጊ{Y$3u-i zp;i|zpNFOH CQ] .ΛRVb8nTA)ExK?NEG/+%25ǧ-. 좾RRJz RfzD,Z0O)+)7pCA2>:S8#a 1t6Si'eI 3h\ʠ"xbF˗bgl`r5_;&m׵cgNсxrVt]X!$ԃ}>uPĩ"(fQxpQ 5-qZ"%`h#w`cpa)2Nj[ _.bE,qYh,{U[ wC@pwߟ~ ,!.~ݛO_~@63 M'a^ܬ˰oe-(xɧ7 jƁQ>\=cJ,kGi.j~Ɗ6^dpfof'KAػmİN>кRzFGu]) :B,C9e e8y>s BgH% (#Da.,!Cl8LO2ϒ0=i."A0=9A>x,9dL=j@L2ӳ|zpNF_t"3辘"9A0=YaTsq:r`g_VQVc` }CF˛ =Z"gZDW%ޟj&YۨsJZ&^+^}hzgzʕ_18H.0w1}{@j#K$g??Eɱm)9G,eIZïXiշjAejI5@wode6Z_{wVV099kcw;^:κC{zzhMuN5|l'9wAө}v,Z͞ͻUZXwnl'9VQmۗ &[ {9+trHqPiTz1v6)1`)bl~os}j~=9i@sEE={bo` y ꜃RF}ȅPN.y㭊{ǭȚL%Na,k\1H 4rC3 1jM}6݂FύB-K-]YdGi,UY`YdKJeW5h i_Yᛲ +쪈e- UI-il6Jc][ɭxVj58p<d6ݦ$CᇨIz8.6CVhc*zY{Jvd H` :1Aji[ňCʤ=i<f&:}XfD&Ė(hm,X4Erl@8#I9kZ僑 2 3̮fyZnmu$hkxS͟ܬ=+ zѻ}ATJ{T'LL m #s xLQ=X v =J~43-ؐ&*4\*DGE;=&YJ٨S"`$p0]U+ޕ5VlZb$5i>}(4l;O|ʉ'̋ RbG@<( .FhPŵRtVVX'm/fz86ƾ.?hJB@.WyU &asH"Ղ n\4"ӱg ĹXLRDL\xtt_ᴹr*4:mᲜ.͎;陉'@p|NcIfy2֎R674_],U zQ1fp)xH3hjIͅ=f^BZ|,WƙD m,ctK2>!muoOUV)6jtyoO |]^#djJp$kfz1'jquHpDt֩ 整qg煉$Îp:D1DK(I5K݆qLu1qSrLLV$PqZ2uP˺g2)x09E,2nL <h*ĝ1Wlqx5}P h1HKȹT61Rlq{s!ɻ,9qQT \hcs4zt[z;<$3Änc^'RfNq8 Bi~7:Lb,hp. FB=#Kyɧc6gkOrr!& Z6ZMndԚơ a}NUX7l#KTӚCpɀij|ʀh06LVB8*(c+Ew@c+;goӘaʿZz~\Nf$Ҹ9uNрjkb00@G9r'ʞ{Ф,R8բ8 c h%HW1]ޖ4@I`SThCΣOwpR޳ L tR=#;gmV!ջ>ǎ:@`<.B]zfq2o񷀲8D!1ڧZ D$ @b'HZMߏg{8-wQEΤ8%V}OPgde&F[&^_~6V!ڲgdeߛfe|ٵXYe RSɄ'*YYObe?Vt4gsǣ7<@M L8\.Ytݮ 9 ǂ!66"1d ˱;燇к}!M&k 8 6$:[.']M"+O6Y|ohzw96Bc-uw:lA_'SP40$ZCk2=9xVc䟬>[_j _^`viJzra դZ+ Nщc0[ʠmP30[ϫڹ|>޾'U MƦmHĻf)xZ1g nNjP\gnN=[ Mʦ,;wӒwAө}v,nw\n-,;7ѽmJrӍ.<e%M$:/Tj!G[6[.r84lX`e Dj+< m/|k-`8Ckd̂)n'&QO2A85Wk<<6qΆRD[)O<9F \=sNM^p(Y2”u1)0 y!?,L_N5@Ƹ=BmF!8IP½.Hb"R2.UD 9VT(}J7߮ť$R 1R" fJ69KZeYcQhB9Zғ>R й4f0ZZ&ŗg8FGEa! @ P`B%ka2ݣ q9rT'%夆t QZ\\hLok԰xk5Ou&agv aX۔Z\&ӔTus!z#4NHvPKr͵1> BxcЯHVDAF.="Gv(R-ܞ*ZtrjN y|w I`8[H1UĈ@ zPZnY:zR|\'ndb8rP.[[ Q*kam[Kڭ7_ZY,'8,Q,*̔kh15}W Bř=-. .,c|#\$2W(jqƼqv|2Jz鱄l 4rLiGi82KTb8Mx;g~>p_NZiɓˁyF#`+,eZ݂X->q p500de.+ xCYVǛ 8#(_ws&;N%/m9Lr $k7uP"Ŭ|+9 GE>F!CEMʏUlQy3G,6}!muwdB#5%:]on^w"٫#ryu1^(>*Ɣ*ًXJU3Cآ'Kҙ:B?t4RKg}Zh5J^5}e[TS1IKJll!aV}B}{p5c<=7-tۡ9B={U?MxCo),7| x&e!_3(jj}?ݼ ͕)5@dzG"i䟿 ?O.5c G&W_`*xu,jCYV˵9/Ql|#gqB`TĜp>\t hn}s7II z XX`.{.21r$wSr9[n]s%ߚ8)M MIً(0cގPv'zV5'9c`Uk+p6҇W!X ;є-l h18 J4ihSvf&­w UX,w[k?Sʎb9 qfl6rh[Z@Q{r٫hA@9㉥\gFzw Pl; ? O =vih!ŜL.Q:Th; *L {* 6A4YDi"񈮎MJ2-z`Nb0Sclt!([7qYfgegoc,޷>7|>rf\/,UZGN('݀oW/,[5~O uGe&nJoS~iබ!" VFܗ>~7!=Z9g*\ֻ;9n>.UI9!Wf0E`'W5o 'Mo~^6:FBFeBCnٻV$WyY4 yvkYMkv5O- >4is)s)*9BPȈ?YK3 HN>Ia9kVM_RE/r\Jgux31dXGuXVG^@˻mWtn:ެ/mw7mye2Zf}̺/o.It(Q}H@b李Ȼ:Ljc-8k}~~nXv&]APrg2c(Ϻ<{VZ?Xfx/6It hRo%ٳ㝅fpDZ֬ aAomcel-@(>, bx?=A؎ T[=51V6'Yy RDWxI;|dJtvY>1HjdbuT̅G_X' V7Q'gM}LX+.@蒲4GA` 'U ^4EIR`ȁ@΃IhnTGՈJ K5⇤ WCpq5ޙ4UNO:(\G!4O i4ܗ{ݕ -N\v=G߮l1\(& 0M#2k>^ʘDp>H8#`Nx&}hfh1&pN_צ{X!(H#zePhj MBm:Ll|q;9ml4Lt!:M4,$YVA)aH =E}ƸПd?߬Qb4"srqpQ*IcL%'Y|3v46|YshaVļJF測i7d rΚ&|2 tىن1k,DFP {?s-ބoTv0&e7l&vvz9=nM9#XQkb^rϬ4 D ފ;fq`]p, O4DlD:x(~(۞| X}$lюM&QMq[ BTq/Jݛ+BK\nO W`$&}p2:!M_i`є?+{[GnaкHC؅Ղ$o5s'羰JnwOko"q)b==tML2I HdT2P[iVll_yv5&IK{fUccDZyXH-D!cۿ?5\M\I~N2!y^(S$9+q%VF;RFJmG+J/FͽYbL_f5q!@>:'s׸bJ@"Y'e )DGxۂ4gKfvm![b]qJ0'i XarnƦ¹E]c}nXuW,i#^iM;:g=D:in c'wBJyK<#tҭ;wTH8w>>@άttk*N!*φg4%1Wi4K>d.$FC5ԝ&A&[ooɎqQr÷i {UwT}(J˖pCRPW[C%6>w14qDAű aA(ZYP7ֵp$QUЕXg'tI;`%ducy%+6t2n3tDl'ݥ@' B7)h3jA'+/NOuP $,IKgаhD[K{&|2I"dѺ[:f h%g[:9o#ql:3=:bZV_؅r|0ss%݀}~2>2Z @.{ȭ÷Fx>{.~*{ỗNZۇ/Isi.Ki$XOaCs.}wAy@Q/>N=5Y\w]N6 eAH{Gw%UOt4( Z0P?#:ԽonC*}-vF vR랬͉tP'g3s|bg&B-rlt{z=m/I Xs0 >Zhn&l y2N7 Aٲ\ F2t0=(|X>(ΘqљhK,mS}ZF%Z8XVQb?z0raq~̌?ϴ&9Vz7j" VKrqnڴ4 *0R۽A5s^\a !T?, { M O',%)2?!TNr~6%?rB/% OJhx vOtS4 q _K]!Bw\&Қ%pYaq 4^BUb.2T^1SahvŮPtVz2F)(% ' "wmH_v;|_e ,s3~l>%<0-ɦdf[X'cIzXEUVǺ5!.鏱a4s,]E,Pժׅr3*$ f.Ew~ӇJLY%(&Ee Xl~cљ˧Jh敶Jc5Wֲ2 <(%P׽ݴתb vq0?_ IGUU9oLkzG-3nb_gPomH5z)$'.U{AX Bf B!ȔRQb!)'.%bMʂbE "» ~ybkkDcK#8 `:ذdՀ_v !GoLMa$f 4囷e贵2U%w `5|\"(E"}Qy \1q`|~)#ؽLbGq+g]Nn7v4v" pz>Svev] P#&"ʝQQΰd: &|*$.+MiR%j+ԉ bBB#Aăӻ#9[*)H*zħcf3qr>I~`j#Y0.VRJD}/LadE<~S vA2FV\w܇]ξKo7YRHdҢk 2+nĮI|Uv9w I331NܐŜ(PDRŏ'gRviPi^hmU G`1R͘f##J^)Ƅ> rRo[BQ J2Y%6L9l]&bQގލu j.*k#Ķ랹DȁWvw7D?0BT8S,HBJ|<L]o\0<7\* A0D2;nen52WIw[-zhΐE Tm? &qa9D430̣rDf7Lnk;/Eύǯ^6/33AE~8n1:8b}5O@7;`X p,\2Vca*Fi.QFjKc9j+TYD%Lpvg+r{+=ؙKMɘU eyI\R(z(" M[mAj*JLwavbӫ/wnL M䞂=j-?ZH0W[&{kc!:+\}9~9ƚ 2x]$lC?iMbrbnfw0a 6B ]ܙ:,s&6\y5S%~GZZfS~ىYLjI:N׏& n-"Fj^i27u6sge!g |=֓6>fӱ-?Ǯdaw$j!GxD's?B[7{wޫݯki*ٷޫӞNM%N]ZPM߲:\b6ˬIa$; +}U$g+ƭTRԖVJ I-HEQ륢"`r5:FyWBEdye.RB`l=N9<أR+";RtVwiN5gDF}ÑXcT_R 88[ Z)<+T;IrH+}UN4/f/TH}VzzV3 u%cȞrj78[)Z)yVl9-8bg38mY)a1Ǘg5 9{:m+@y?RWbl^RYrόnrߍ`pWe3,՚tq湅9I V^{?uӞLntt~r{+\9[UQa sAGu23޴Bz`+-앁Hw5f Zcv2q3TܫЮH&UY8XmJB8'DPG'g Dc ? $V!~lA_x࿾ktvwg VZ #JR{a:ŋ例se,Ϲq}4PF\!F?0|@X-~a_Xgh3N[QSF[(,*3>is;a5ԄR;;m@?kD e~G8{Q IJ78{-УKe&D$MQQf$;>X2e[s+&q3`cO%Y&Vq UatHoec$:8DWY93#LmH1:=1#S)Y.;^Tӎ*Ќ"Nm+qŰHc $뛀㽟uJ0g.z2APA"}uU뼾_.gdJƃ߀zSQ!3!uV )& n3j,6R .سmFLa܆2KLq70=8xv2X0!*ja>WŸބ;XeUI WP)oG+o}\x86FY3={;jjJ-aqW$Ha\Y|BԴ뱫?:ףW/ AH74O1`<εm K;H@wT9EJO C_qc5_$Dl2 ۛpغ-CHޏJ%Y*""7]wMj5Db"ex#9[nKVS2,%GA[]yzV]ҡ 'Uo_s9P[/>r`\CYn>IeZ+Up+(Q}J5yWPXKd%*B%.Z;s@u5՘{%_)ުU%a90Y ]{BZ 6+2saL!֑[jy;f7ضyޙ|bfk|"cTh}_ nc+]^y7uyt']i 0G|MwMfW])U#O7oJɹIӞ¡{ſNRCGu{g`Ê6ǂ;3hALR>ܴLy\C`׵$ (n<9q@V9M!]yG2[M4)qN JBw%[k~}8F&1)8s+bfyVYIsAֽL`11ejt:8 -V5K;_Öu8]dHyiD]GGC]PD݀FuLw=S&T _Z39%P~6Z)$aq9vxZ9`,vs`x%`V1A+&%_N "`C!*ľ !0 XD&g\-+9c LFxW!@"S"6RdfcT_RM ?N>͞|+ n*L!P4r^HəbLY"pSM5WPX.GHM`t`tu]D_(BosO|'Tn@bR9JH0&)y&u7@) 2IfaЬs]@Fa]uVBcvQ ~8y2Y>%ઊ*8R(P.L һbvER"$%ʹYfJԡ@[_>X~֚ 8M `ju<Ve†v%lHjY5;[C ɎW&ϖTD]B?4YUMjf0_f'5ogMXRTm\j˥\Kߚ׳~n.mL%2)cF_@Ʈ$W 칵]SQwB W1HqF|]S]把[9`G#pְpjk~Ǿ Q|G6`퇏c/갽}[m V>VKnANw Ƿw->9 @tZQA>-'ٽ1>ݽ`(w= ~z'q(NQWi6FUCv?4DxwlOq˝j޽) 3: w%QE/½)q"7Xo|PPHɧlP64?WpjGBjaO^<;?E+fvj!Sw\ݨ:!vM9Iexvcj[ΗJ:A hpBVjN,;=+v זѻ@EQ儛`O@~!D)ǯjG)ǿhv*'%UYN:6\i1θjii|F?YuZYuZYuZYu˪~)֔& E.mL( ΅S@}*)UhnE}y.?vw2լqְm!pY{=kag>ލKڽm[E+-#*Zi$;]T\X76tHhWk3LJ(vު,2fAD%FE'HNtk켬h{)^3ϋHz*>$D-M&:U܉+6/@G8(6mÑTP| ,v?NPù8]+3 FIrNW̦ME+mc')Z<(E+DI,~K%OW\i UҾڐTwdwHMޮq֗vx-oeCn;xܩa20J+V4TF |k^ҶqK5og fT5͚Ď-.6M-Wj]L_%4U&5G,A[DLNK\y“sADfhdw |%0" X=ސgz3h-/Km9+r:DZ(&o 8 m47S3|nXZR dFRMv42I'b ͭvj@/LVRCbkYNʗſ=}륿asi dDdd 2EF*" t/e# LgWO#t1j`vJp˪\cSJ ϕ@,$Ԍ4qA2֕dk`y]X̫MEeN1 KMk"#(/*M:ޗBU[} 5Zx%S<1XȑK1`t7mBp[k!_Y[=ÏmV럛i>@Ҏ HPH+ebIi~,l[4_aO WUTC!5kYa _gxhneS>(&oVqMHeӂ逴31gA+=F2'> gϮj$wf` PFj/kRFfT$^IaR*XUnؚMQdrJ$IGfm&6'K~z֚#7̬vxXǁG'MmWL#krCjөVH Eq|qkr2WvZ(x!"zAUێwO̳26n5|+b"l%363t`؂*(QN>S_X=~xV%1#PI39WgCku -8)9RZ@a'W'9RJJ3 ?0Sp$qxʹclA|Φ"z@,fH Âh?,|!8=4b9')!*2P J&/:(ۑii5%{lh.BQ'7SHo41Xy7Adr&![Gd6eIQ>O^=N%51|^\xZ7t3'D)LLhbNW8dMK(Ee[zJZU+;qܵT a%Ԭ?8&qe$So{GI_o cwv=HTbmK3(KPI4M/z CִC LtS^<%ZF-"6Jv oEԺ 84pmSGGhp ҄)I &E!Qa7otl}ũ Xwb/sͣڵhxVd1&>ڸt}.c=cµ~Y?rV[_1k^5w~hHJ5 ]!X [<$̨dY4C%vL~32ݨ'@I[_|m&5VV^gE%rv>ϗCZ6o[+6Sc0%XM͙gǙ  r-BL: ܛ2N#)a!|zqvdKF*c %̆^ r V;jkqzf"ObEU~݀ywb<`}Ǖ,!sF!'<@ت=\OI8&Kۇ2PbȪ|)t-|s -+'zi%OcBFmoYkhc[W놨 Ajb` +AJ_50a~H( ӏIY*02J:AtPs-_u />{Bb篯.kc"`9b ~,FNƖrgȆ)w>@W|NFl0mAyi::ѽ.ZI_wYco3⣹ f3)5-N8t9D/2uz2b^?ήqW! @' g D{r{Op1 OUi{]䣣lroeʥ/J [sbyZ?5N4nFS=UpOJ_~UVa:0^[vxЦ:v8M8F@LY}leX'4#Fy~vaOE%0obt -aڭGo:7C!>CQ/6%yԠnVe:PH 头(>U؎kܬn]Z|۬7kl`H[{Uh>5GVX'WW8CKY#3\;@%%Xᗞηh[֍×K_C,D05^ɰw>XHE6׍pDPtƴZƝqet O_xyb fCNZi_\幀1No}Fѱc8Wh + 5>HKF,l_^湐5BטFMQB JB5iLDǙ sU) r~YjiEސy>G64_fiXc˦h|+3Y-N޻ Pt]Fi%m`;svq}4s8( kiz.Tz*N0uq?Mn1.. J+3fe|'gh/͆2~1zغ[F7~6:)1@#4?I$@V/}?w֋I#, PRHA?Y$bG"Fkhƃ4; bsi-vŴ—52!hU F9i; N݀vBrʙ]YzlD~HZ{,SMJt*R%гjtv RSՀ̩sB)eQY{VMt2¤"m{ic3A|Xͼ $-1MܿZ bģ|<B&آ h̨J~5Nf hc\3-=լ [u{u3bȚZLW1jRֆh ^ Z:'_-Keq!+_v3`ɦhi <\ch,وivy ~O#Ke ܢw,tz"{鱯Mu`t6%i-:L" UWZ1Bܷ~VQ-v|:7"Uax :X:KW%h5WPFJ#M7k^X^ XùxAL~HM?^aLIBe9t i΂#X;u|O/g7~ Pa!(*Lh\Jd"`(P[]2pr a6i~ܐյr q0Ahq6 f2Y l5Ch.f{JcI|0e_ o;eYwa}j74ΨSnml,V]n8=pY`N|V$Q2)π|Vg_bb1= 9'Txq ¸8OJZ1QJ=7h)Ye,z ![ǃx'FZx7,{7ջ;|BM*{mv(↴5-T.:54k7\+xLe޿G 8;%Vw2Z>ߺ×5f)uJGy? 3z5=)gNxK3h= 94Wʿ>T,eoRN#'U~M'̩9 5ؕ x8٧e}(WA--0od@| 3ka0)IY8 9_v_~z??|s_bƚrLg3-">\t7F_OXmu6>?-7@.nu/a] OLF Ҡgtpt_N70RJE$A]6'QVB[K>&v{Owr6)R*/zVu jQK붮0RS*Lh$KGGPVTrQC֊*\l~?{!qҩoylaݙ#E2ӴɂE )%hǝ3?&%_r'}y ʂ%Je7:-n/ZgoI7oчrNlEjKNwց/Ʈ$U{-4erc"hn)?/֒fR۔~{ٿe҉SqW!ڤR|M kvK"ä5/[Z_ڋ#uF}c6\m;[V1q#YA/_ !g$gLf`4nGYVtdr~IɦnvM= v6cbWUUUH^u,g H`]k>躼V_FwE Ɍ)hc41F,*MR%rh? N 6a{W>9|UFlmeI6v[jK/1fEfґQ+-~p?ts%j&&Hi5gaā*XSdPDBDK FJKퟣ]f/P 4Ȧ!gskvc\Fh!GYb3 `(3 1p6&LR5hB82)1(ML$a(*EsiD1 A1VšHJ3}ebru3 y3%|{L*@ddYg]]VmI4.cTo6U/lep 5J3ZT~'3`VFPJ#IjjH}SF-4B:TLgya(ʐ?ސ%,U[2 hc쑎߁1 fp z+{Z}%gǥB,ώi!p׀*t;BrAH1"6#@,SWGbS+BBt."%d|e? \1;^'] Ģ}Yg.rWq1-e!*^BS j=%K];ˍ=%#y/O$}HWR2.E(+AJov4JH(N>=(R9Cp4n'3&\')K"&ƱF (w)yh0BȇT FDpz}O?L~'mI=M˘Rbt 83)Fb0SsIlFqYdJh<lS샰{J;CK-u )9x (P7$8#rdpGe"F1&~y2$e,AMZ8Ck J2H9,I,X0h_R!M#0 ASC.;j\;Y* X:J0 $}ˀo0=zlv~9[-xs69fiat/[ 믺Coț~!E۽˙?itkE "@[ 5_!>X|Ȼx<"Fƥݻ!!|~oluG !Zؾ?.'.}ߗO]<}@_^R6KRTAOoY11@yVud#CyorQA$ >1 `>BT>ފ=r#,|kx})V`0§OT)زW ÉNDF@AZ/sISnB#:8PT[ɓǪA ܶ:C @O  xie.Q "Ђ'PKDjZ7qT"-J6$AC7vD0& հS}(pM `[9Dx״!B?4xF"N<ɞȕLA!* 2ZA`WPÜ=PǪE]W\ZD"!~eH_NFϟ[_HFְ`M}  Vvy8t%(p2D΃!?ҺoH;y~Wos$0C$f>p,^ߟ4< r`Ɵ݀We{QغGzŔX#}kHv`6ܝEȊu˗yAjؙ*rvw׹B=^0[0ar{VWJG:"bӐ$e' я $?Y3c«!˅|?i 4Tacn2~w޻f4)~~`{7|;~6%Ú?ʋU4r*J aT=bϦ+d*B m J5yK<4_~fA}$T.NO0l%ԦhƜ<*Ujz2fjKFVӣG\%哒W枏Ղq?f4-y<-!I--t/ٴ[ bQPtIh<{qOfުF5 D3@FUWRW^_iWOd4UQ8§"qXA c1xxUc5p4 tL`wvT$YQC| 86=V 2ч%{_ &J=vקbYׁ 2"s1%Tt>̛z,j WjTiKD9ޤNxʐ(u-ڂ m9noj4mkXJƯ47%C\v2a)Kac!Z,A>buV L.j/HU qZn-H\[E &AֹU8_d=.)H\3]x'PJE8қn]CE^M *%`%ռJ-fXNU?CTTYuCGs[VS,Tkϐe@q7Zov  e;!~V{ 7W.Q2;;ME[,!*ڭi)}l?6.P5!!\Dcd 8!j7VPb":諸ݺ;xܚvݚW.eDõ{aٚjB@ N4&NɃjyc|7,s>eg&mnt3 ɀA?՛vc Ɠy'ji]2rqv/_w_|Vgŀ/H'rtq#rfrU.;7H1 Ry._zet_'X IoAE~ez:޾ӹH,APКɮM5=w$ͽAvF_:go!dk<~`Ms6G.ٜ7i4>PŝZ ]BID-b8KY',Ju` f-GǪdI OՐRD]Ze +/63Z} 2,P&R&`?^gRz)SZ8nr]Iܕ$J`+BT{d*ܶ;NrhHm[ORuS[d4T5^rkZSY'UG")/:yuACCCU :Z;[¢IQ>\* Tխ׋T ՝ch苞S@nogx`Q,p>T_r%֬w{|>8:[.[h)6[F=k.)])-/aK(|4'7˒:3;Yŧ({ߟB-536*&ـ t\*$xOW")J1=seȹ0m}+`mlz;xdS-5EO6&E)6#ՑxEm'G+EMp.Y|cE5%(k=_4^3ᘗmMw,vqI)vy%0|kM%e /J{FLJB+!`N,:w.>v亂OǾQގ~^T"jۛMKdyŰU¬!(y1NoñKq9S@8?h%CՑ@PDꙠCqz/%r1&Iw@iy`&A 9Τϰds ɻ\ xJim/-a\y|NS$K~ٲZGj]m7m2ΈM79_Qoۺw5w# F >ߖꬿʇSg j^NatW;Í0,'t &J[=\/K$҅G;Z*8G*8o-\\*8?]HLAtIWxB0WD9 aj`hYC sL:KQHJij,@!).wऎb-ȔK/QSTQOX\!vsD9Csu_3SXt+I%WNuLdġbM}A֞M㺩.$bΔ)X\ Sn8Gbՙ@f 6&f]LIRw^3ᘘ:IC9~M5:,y=m+l:ڸ۴ߍ6UD*[Yyzśݙ!{[o"뤱ܸ wo{t ]-p\[oD78_8O$N{).X^OZ!`_}ivar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005154566015145074226017717 0ustar rootrootFeb 17 13:04:19 crc systemd[1]: Starting Kubernetes Kubelet... Feb 17 13:04:19 crc restorecon[4676]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:19 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:20 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 17 13:04:21 crc restorecon[4676]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 17 13:04:21 crc restorecon[4676]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Feb 17 13:04:21 crc kubenswrapper[4955]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 17 13:04:21 crc kubenswrapper[4955]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Feb 17 13:04:21 crc kubenswrapper[4955]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 17 13:04:21 crc kubenswrapper[4955]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 17 13:04:21 crc kubenswrapper[4955]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Feb 17 13:04:21 crc kubenswrapper[4955]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.937616 4955 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.943626 4955 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.943660 4955 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.943673 4955 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.943686 4955 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.943696 4955 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.943705 4955 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.943714 4955 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.943722 4955 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.943730 4955 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.943739 4955 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.943747 4955 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.943768 4955 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.943777 4955 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.943814 4955 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.943823 4955 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.943831 4955 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.943838 4955 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.943847 4955 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.943856 4955 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.943865 4955 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.943875 4955 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.943884 4955 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.943892 4955 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.943899 4955 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.943907 4955 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.943915 4955 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.943923 4955 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.943933 4955 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.943940 4955 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.943949 4955 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.943957 4955 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.943964 4955 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.943972 4955 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.943979 4955 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.943987 4955 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.943995 4955 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.944003 4955 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.944011 4955 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.944021 4955 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.944031 4955 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.944041 4955 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.944049 4955 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.944057 4955 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.944067 4955 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.944076 4955 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.944084 4955 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.944094 4955 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.944104 4955 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.944113 4955 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.944121 4955 feature_gate.go:330] unrecognized feature gate: Example Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.944129 4955 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.944137 4955 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.944144 4955 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.944156 4955 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.944163 4955 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.944170 4955 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.944178 4955 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.944186 4955 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.944193 4955 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.944200 4955 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.944208 4955 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.944219 4955 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.944228 4955 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.944236 4955 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.944244 4955 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.944251 4955 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.944261 4955 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.944269 4955 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.944276 4955 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.944284 4955 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.944291 4955 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945157 4955 flags.go:64] FLAG: --address="0.0.0.0" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945180 4955 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945197 4955 flags.go:64] FLAG: --anonymous-auth="true" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945208 4955 flags.go:64] FLAG: --application-metrics-count-limit="100" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945219 4955 flags.go:64] FLAG: --authentication-token-webhook="false" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945228 4955 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945241 4955 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945252 4955 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945261 4955 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945270 4955 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945279 4955 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945289 4955 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945298 4955 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945307 4955 flags.go:64] FLAG: --cgroup-root="" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945316 4955 flags.go:64] FLAG: --cgroups-per-qos="true" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945326 4955 flags.go:64] FLAG: --client-ca-file="" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945335 4955 flags.go:64] FLAG: --cloud-config="" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945343 4955 flags.go:64] FLAG: --cloud-provider="" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945352 4955 flags.go:64] FLAG: --cluster-dns="[]" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945362 4955 flags.go:64] FLAG: --cluster-domain="" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945371 4955 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945380 4955 flags.go:64] FLAG: --config-dir="" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945390 4955 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945400 4955 flags.go:64] FLAG: --container-log-max-files="5" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945411 4955 flags.go:64] FLAG: --container-log-max-size="10Mi" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945420 4955 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945429 4955 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945438 4955 flags.go:64] FLAG: --containerd-namespace="k8s.io" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945448 4955 flags.go:64] FLAG: --contention-profiling="false" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945457 4955 flags.go:64] FLAG: --cpu-cfs-quota="true" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945466 4955 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945475 4955 flags.go:64] FLAG: --cpu-manager-policy="none" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945484 4955 flags.go:64] FLAG: --cpu-manager-policy-options="" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945494 4955 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945503 4955 flags.go:64] FLAG: --enable-controller-attach-detach="true" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945512 4955 flags.go:64] FLAG: --enable-debugging-handlers="true" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945521 4955 flags.go:64] FLAG: --enable-load-reader="false" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945530 4955 flags.go:64] FLAG: --enable-server="true" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945539 4955 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945550 4955 flags.go:64] FLAG: --event-burst="100" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945559 4955 flags.go:64] FLAG: --event-qps="50" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945568 4955 flags.go:64] FLAG: --event-storage-age-limit="default=0" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945577 4955 flags.go:64] FLAG: --event-storage-event-limit="default=0" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945586 4955 flags.go:64] FLAG: --eviction-hard="" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945596 4955 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945605 4955 flags.go:64] FLAG: --eviction-minimum-reclaim="" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945614 4955 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945623 4955 flags.go:64] FLAG: --eviction-soft="" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945632 4955 flags.go:64] FLAG: --eviction-soft-grace-period="" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945640 4955 flags.go:64] FLAG: --exit-on-lock-contention="false" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945649 4955 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945658 4955 flags.go:64] FLAG: --experimental-mounter-path="" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945667 4955 flags.go:64] FLAG: --fail-cgroupv1="false" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945677 4955 flags.go:64] FLAG: --fail-swap-on="true" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945687 4955 flags.go:64] FLAG: --feature-gates="" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945698 4955 flags.go:64] FLAG: --file-check-frequency="20s" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945707 4955 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945717 4955 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945726 4955 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945735 4955 flags.go:64] FLAG: --healthz-port="10248" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945744 4955 flags.go:64] FLAG: --help="false" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945752 4955 flags.go:64] FLAG: --hostname-override="" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945760 4955 flags.go:64] FLAG: --housekeeping-interval="10s" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945770 4955 flags.go:64] FLAG: --http-check-frequency="20s" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945803 4955 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945812 4955 flags.go:64] FLAG: --image-credential-provider-config="" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945821 4955 flags.go:64] FLAG: --image-gc-high-threshold="85" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945830 4955 flags.go:64] FLAG: --image-gc-low-threshold="80" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945839 4955 flags.go:64] FLAG: --image-service-endpoint="" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945848 4955 flags.go:64] FLAG: --kernel-memcg-notification="false" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945856 4955 flags.go:64] FLAG: --kube-api-burst="100" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945865 4955 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945874 4955 flags.go:64] FLAG: --kube-api-qps="50" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945883 4955 flags.go:64] FLAG: --kube-reserved="" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945892 4955 flags.go:64] FLAG: --kube-reserved-cgroup="" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945900 4955 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945909 4955 flags.go:64] FLAG: --kubelet-cgroups="" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945918 4955 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945927 4955 flags.go:64] FLAG: --lock-file="" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945936 4955 flags.go:64] FLAG: --log-cadvisor-usage="false" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945945 4955 flags.go:64] FLAG: --log-flush-frequency="5s" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945954 4955 flags.go:64] FLAG: --log-json-info-buffer-size="0" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945968 4955 flags.go:64] FLAG: --log-json-split-stream="false" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945976 4955 flags.go:64] FLAG: --log-text-info-buffer-size="0" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945985 4955 flags.go:64] FLAG: --log-text-split-stream="false" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.945994 4955 flags.go:64] FLAG: --logging-format="text" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946003 4955 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946013 4955 flags.go:64] FLAG: --make-iptables-util-chains="true" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946021 4955 flags.go:64] FLAG: --manifest-url="" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946030 4955 flags.go:64] FLAG: --manifest-url-header="" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946042 4955 flags.go:64] FLAG: --max-housekeeping-interval="15s" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946051 4955 flags.go:64] FLAG: --max-open-files="1000000" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946061 4955 flags.go:64] FLAG: --max-pods="110" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946070 4955 flags.go:64] FLAG: --maximum-dead-containers="-1" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946078 4955 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946089 4955 flags.go:64] FLAG: --memory-manager-policy="None" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946098 4955 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946107 4955 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946116 4955 flags.go:64] FLAG: --node-ip="192.168.126.11" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946126 4955 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946146 4955 flags.go:64] FLAG: --node-status-max-images="50" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946155 4955 flags.go:64] FLAG: --node-status-update-frequency="10s" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946164 4955 flags.go:64] FLAG: --oom-score-adj="-999" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946173 4955 flags.go:64] FLAG: --pod-cidr="" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946181 4955 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946195 4955 flags.go:64] FLAG: --pod-manifest-path="" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946204 4955 flags.go:64] FLAG: --pod-max-pids="-1" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946212 4955 flags.go:64] FLAG: --pods-per-core="0" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946221 4955 flags.go:64] FLAG: --port="10250" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946230 4955 flags.go:64] FLAG: --protect-kernel-defaults="false" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946239 4955 flags.go:64] FLAG: --provider-id="" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946248 4955 flags.go:64] FLAG: --qos-reserved="" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946256 4955 flags.go:64] FLAG: --read-only-port="10255" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946265 4955 flags.go:64] FLAG: --register-node="true" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946274 4955 flags.go:64] FLAG: --register-schedulable="true" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946284 4955 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946298 4955 flags.go:64] FLAG: --registry-burst="10" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946307 4955 flags.go:64] FLAG: --registry-qps="5" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946316 4955 flags.go:64] FLAG: --reserved-cpus="" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946326 4955 flags.go:64] FLAG: --reserved-memory="" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946338 4955 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946348 4955 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946358 4955 flags.go:64] FLAG: --rotate-certificates="false" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946367 4955 flags.go:64] FLAG: --rotate-server-certificates="false" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946377 4955 flags.go:64] FLAG: --runonce="false" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946387 4955 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946396 4955 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946405 4955 flags.go:64] FLAG: --seccomp-default="false" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946414 4955 flags.go:64] FLAG: --serialize-image-pulls="true" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946422 4955 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946431 4955 flags.go:64] FLAG: --storage-driver-db="cadvisor" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946440 4955 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946450 4955 flags.go:64] FLAG: --storage-driver-password="root" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946458 4955 flags.go:64] FLAG: --storage-driver-secure="false" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946467 4955 flags.go:64] FLAG: --storage-driver-table="stats" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946476 4955 flags.go:64] FLAG: --storage-driver-user="root" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946486 4955 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946496 4955 flags.go:64] FLAG: --sync-frequency="1m0s" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946505 4955 flags.go:64] FLAG: --system-cgroups="" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946514 4955 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946527 4955 flags.go:64] FLAG: --system-reserved-cgroup="" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946536 4955 flags.go:64] FLAG: --tls-cert-file="" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946544 4955 flags.go:64] FLAG: --tls-cipher-suites="[]" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946555 4955 flags.go:64] FLAG: --tls-min-version="" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946563 4955 flags.go:64] FLAG: --tls-private-key-file="" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946573 4955 flags.go:64] FLAG: --topology-manager-policy="none" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946583 4955 flags.go:64] FLAG: --topology-manager-policy-options="" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946593 4955 flags.go:64] FLAG: --topology-manager-scope="container" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946602 4955 flags.go:64] FLAG: --v="2" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946612 4955 flags.go:64] FLAG: --version="false" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946632 4955 flags.go:64] FLAG: --vmodule="" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946645 4955 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.946655 4955 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.946888 4955 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.946899 4955 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.946907 4955 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.946916 4955 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.946924 4955 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.946935 4955 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.946944 4955 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.946953 4955 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.946962 4955 feature_gate.go:330] unrecognized feature gate: Example Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.946971 4955 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.946979 4955 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.946989 4955 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.946999 4955 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947008 4955 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947016 4955 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947025 4955 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947034 4955 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947042 4955 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947050 4955 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947058 4955 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947067 4955 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947074 4955 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947082 4955 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947089 4955 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947097 4955 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947105 4955 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947113 4955 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947120 4955 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947128 4955 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947138 4955 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947150 4955 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947159 4955 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947167 4955 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947175 4955 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947184 4955 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947192 4955 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947201 4955 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947208 4955 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947216 4955 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947223 4955 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947231 4955 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947240 4955 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947248 4955 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947255 4955 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947263 4955 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947271 4955 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947279 4955 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947286 4955 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947294 4955 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947301 4955 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947309 4955 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947319 4955 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947328 4955 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947336 4955 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947343 4955 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947352 4955 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947360 4955 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947368 4955 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947376 4955 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947383 4955 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947393 4955 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947405 4955 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947416 4955 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947424 4955 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947433 4955 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947441 4955 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947448 4955 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947456 4955 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947464 4955 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947472 4955 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.947479 4955 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.947491 4955 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.964063 4955 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.964121 4955 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964231 4955 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964242 4955 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964249 4955 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964254 4955 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964260 4955 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964267 4955 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964273 4955 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964278 4955 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964283 4955 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964288 4955 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964294 4955 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964299 4955 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964305 4955 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964313 4955 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964318 4955 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964324 4955 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964330 4955 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964335 4955 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964340 4955 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964345 4955 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964350 4955 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964355 4955 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964360 4955 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964365 4955 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964370 4955 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964376 4955 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964385 4955 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964420 4955 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964425 4955 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964431 4955 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964437 4955 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964451 4955 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964457 4955 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964463 4955 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964469 4955 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964474 4955 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964479 4955 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964485 4955 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964491 4955 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964496 4955 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964501 4955 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964506 4955 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964512 4955 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964517 4955 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964522 4955 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964527 4955 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964533 4955 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964538 4955 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964544 4955 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964549 4955 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964555 4955 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964562 4955 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964568 4955 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964574 4955 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964579 4955 feature_gate.go:330] unrecognized feature gate: Example Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964585 4955 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964590 4955 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964595 4955 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964600 4955 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964605 4955 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964610 4955 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964615 4955 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964620 4955 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964626 4955 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964631 4955 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964636 4955 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964641 4955 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964646 4955 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964653 4955 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964660 4955 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964666 4955 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.964690 4955 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964896 4955 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964909 4955 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964915 4955 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964922 4955 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964927 4955 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964932 4955 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964938 4955 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964944 4955 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964950 4955 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964956 4955 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964962 4955 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964968 4955 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964974 4955 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964980 4955 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964986 4955 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964992 4955 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.964998 4955 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965003 4955 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965009 4955 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965015 4955 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965020 4955 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965027 4955 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965032 4955 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965038 4955 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965043 4955 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965049 4955 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965054 4955 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965059 4955 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965065 4955 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965070 4955 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965076 4955 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965081 4955 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965087 4955 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965094 4955 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965102 4955 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965108 4955 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965115 4955 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965123 4955 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965129 4955 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965136 4955 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965144 4955 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965150 4955 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965157 4955 feature_gate.go:330] unrecognized feature gate: Example Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965162 4955 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965169 4955 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965175 4955 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965182 4955 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965188 4955 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965194 4955 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965200 4955 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965206 4955 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965212 4955 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965220 4955 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965227 4955 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965234 4955 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965241 4955 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965247 4955 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965252 4955 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965258 4955 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965264 4955 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965270 4955 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965275 4955 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965281 4955 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965287 4955 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965292 4955 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965297 4955 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965302 4955 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965308 4955 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965314 4955 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965319 4955 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 17 13:04:21 crc kubenswrapper[4955]: W0217 13:04:21.965330 4955 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.965339 4955 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.965636 4955 server.go:940] "Client rotation is on, will bootstrap in background" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.971013 4955 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.971141 4955 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.972772 4955 server.go:997] "Starting client certificate rotation" Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.972822 4955 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.973850 4955 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-20 02:22:24.491623056 +0000 UTC Feb 17 13:04:21 crc kubenswrapper[4955]: I0217 13:04:21.973975 4955 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.004283 4955 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.009520 4955 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 17 13:04:22 crc kubenswrapper[4955]: E0217 13:04:22.010887 4955 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.46:6443: connect: connection refused" logger="UnhandledError" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.035764 4955 log.go:25] "Validated CRI v1 runtime API" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.076488 4955 log.go:25] "Validated CRI v1 image API" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.078972 4955 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.086832 4955 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-02-17-12-59-58-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.086876 4955 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.118405 4955 manager.go:217] Machine: {Timestamp:2026-02-17 13:04:22.114005057 +0000 UTC m=+0.636734680 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:6d4602d9-c718-4393-8f91-3c381f800c9e BootID:7605a498-360b-4ce4-b7a1-dfbc354d798d Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:86:e1:3c Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:86:e1:3c Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:39:c0:ee Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:30:47:96 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:aa:ec:8c Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:2f:56:a4 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:de:00:33:0c:e5:3e Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:7e:55:12:93:b5:14 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.119346 4955 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.119684 4955 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.120348 4955 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.120759 4955 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.120846 4955 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.121223 4955 topology_manager.go:138] "Creating topology manager with none policy" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.121244 4955 container_manager_linux.go:303] "Creating device plugin manager" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.121911 4955 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.121973 4955 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.123584 4955 state_mem.go:36] "Initialized new in-memory state store" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.124255 4955 server.go:1245] "Using root directory" path="/var/lib/kubelet" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.128023 4955 kubelet.go:418] "Attempting to sync node with API server" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.128061 4955 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.128164 4955 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.128191 4955 kubelet.go:324] "Adding apiserver pod source" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.128213 4955 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 17 13:04:22 crc kubenswrapper[4955]: W0217 13:04:22.133409 4955 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.46:6443: connect: connection refused Feb 17 13:04:22 crc kubenswrapper[4955]: E0217 13:04:22.133543 4955 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.46:6443: connect: connection refused" logger="UnhandledError" Feb 17 13:04:22 crc kubenswrapper[4955]: W0217 13:04:22.133619 4955 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.46:6443: connect: connection refused Feb 17 13:04:22 crc kubenswrapper[4955]: E0217 13:04:22.133858 4955 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.46:6443: connect: connection refused" logger="UnhandledError" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.134421 4955 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.135650 4955 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.138064 4955 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.139834 4955 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.139871 4955 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.139879 4955 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.139888 4955 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.139904 4955 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.139913 4955 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.139925 4955 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.139940 4955 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.139954 4955 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.139964 4955 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.139976 4955 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.139987 4955 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.141030 4955 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.141612 4955 server.go:1280] "Started kubelet" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.143297 4955 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.143320 4955 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Feb 17 13:04:22 crc systemd[1]: Started Kubernetes Kubelet. Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.144462 4955 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.46:6443: connect: connection refused Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.144616 4955 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.151608 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.151661 4955 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.152386 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 01:06:31.814560886 +0000 UTC Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.153276 4955 volume_manager.go:287] "The desired_state_of_world populator starts" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.153357 4955 volume_manager.go:289] "Starting Kubelet Volume Manager" Feb 17 13:04:22 crc kubenswrapper[4955]: E0217 13:04:22.153365 4955 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.153465 4955 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Feb 17 13:04:22 crc kubenswrapper[4955]: W0217 13:04:22.154369 4955 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.46:6443: connect: connection refused Feb 17 13:04:22 crc kubenswrapper[4955]: E0217 13:04:22.154479 4955 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.46:6443: connect: connection refused" logger="UnhandledError" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.154724 4955 server.go:460] "Adding debug handlers to kubelet server" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.155692 4955 factory.go:55] Registering systemd factory Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.155768 4955 factory.go:221] Registration of the systemd container factory successfully Feb 17 13:04:22 crc kubenswrapper[4955]: E0217 13:04:22.155679 4955 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.46:6443: connect: connection refused" interval="200ms" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.156491 4955 factory.go:153] Registering CRI-O factory Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.156532 4955 factory.go:221] Registration of the crio container factory successfully Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.156687 4955 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.156757 4955 factory.go:103] Registering Raw factory Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.156859 4955 manager.go:1196] Started watching for new ooms in manager Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.158539 4955 manager.go:319] Starting recovery of all containers Feb 17 13:04:22 crc kubenswrapper[4955]: E0217 13:04:22.163725 4955 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.46:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.18950a661db63189 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-17 13:04:22.141579657 +0000 UTC m=+0.664309200,LastTimestamp:2026-02-17 13:04:22.141579657 +0000 UTC m=+0.664309200,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.178743 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.179132 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.179270 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.179394 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.179512 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.179649 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.179780 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.179965 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.180105 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.180223 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.180355 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.180485 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.180615 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.180807 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.182605 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.182873 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.183052 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.183244 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.183465 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.183649 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.183770 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.183953 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.184093 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.184214 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.184330 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.184457 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.184587 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.184707 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.184901 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.185027 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.185149 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.185288 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.185421 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.185563 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.185699 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.185859 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.185995 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.186120 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.186255 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.186436 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.186554 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.186668 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.186810 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.186950 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.187172 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.187331 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.187463 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.187596 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.187715 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.185209 4955 manager.go:324] Recovery completed Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.187871 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.188130 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.188249 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.188404 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.188548 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.188671 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.188847 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.188976 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.189105 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.189221 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.189353 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.189476 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.189591 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.189705 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.189893 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.190033 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.190153 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.190275 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.190391 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.190506 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.190641 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.190760 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.190936 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.191057 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.191172 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.191315 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.191448 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.191568 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.191682 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.191824 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.191955 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.192092 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.192211 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.192335 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.192452 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.192566 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.192690 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.192909 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.193040 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.193152 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.193264 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.193377 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.193512 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.193641 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.193758 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.193916 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.194853 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.194912 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.194942 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.194965 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.194986 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195008 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195029 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195050 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195072 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195107 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195144 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195169 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195197 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195220 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195243 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195266 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195289 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195312 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195335 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195356 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195376 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195395 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195414 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195433 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195452 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195472 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195491 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195513 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195535 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195555 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195577 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195600 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195620 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195640 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195660 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195679 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195698 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195720 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195740 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195759 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195778 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195821 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195840 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195861 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195891 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195909 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195930 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195954 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195973 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.195993 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.196013 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.196034 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.196055 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.200383 4955 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.200440 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.200464 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.200484 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.200504 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.200525 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.200545 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.200563 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.200581 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.200599 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.200620 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.200639 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.200656 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.200675 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.200693 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.200712 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.200732 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.200754 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.200859 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.200879 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.200898 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.200917 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.200940 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.200959 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.200981 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.201000 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.201067 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.201088 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.201109 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.201130 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.201150 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.201169 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.201189 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.201209 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.201228 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.201247 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.201266 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.201287 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.201305 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.201324 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.201344 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.201393 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.201416 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.201438 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.201457 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.201476 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.201495 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.201515 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.201534 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.201554 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.201574 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.201593 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.201613 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.201633 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.201653 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.201673 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.201692 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.201711 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.201731 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.201754 4955 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.201772 4955 reconstruct.go:97] "Volume reconstruction finished" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.201828 4955 reconciler.go:26] "Reconciler: start to sync state" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.206748 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.210565 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.210615 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.210628 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.211917 4955 cpu_manager.go:225] "Starting CPU manager" policy="none" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.211958 4955 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.211982 4955 state_mem.go:36] "Initialized new in-memory state store" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.218508 4955 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.221563 4955 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.221609 4955 status_manager.go:217] "Starting to sync pod status with apiserver" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.221638 4955 kubelet.go:2335] "Starting kubelet main sync loop" Feb 17 13:04:22 crc kubenswrapper[4955]: E0217 13:04:22.221682 4955 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Feb 17 13:04:22 crc kubenswrapper[4955]: W0217 13:04:22.225469 4955 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.46:6443: connect: connection refused Feb 17 13:04:22 crc kubenswrapper[4955]: E0217 13:04:22.225757 4955 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.46:6443: connect: connection refused" logger="UnhandledError" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.229936 4955 policy_none.go:49] "None policy: Start" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.231763 4955 memory_manager.go:170] "Starting memorymanager" policy="None" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.231848 4955 state_mem.go:35] "Initializing new in-memory state store" Feb 17 13:04:22 crc kubenswrapper[4955]: E0217 13:04:22.253662 4955 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.310969 4955 manager.go:334] "Starting Device Plugin manager" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.311120 4955 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.311146 4955 server.go:79] "Starting device plugin registration server" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.312179 4955 eviction_manager.go:189] "Eviction manager: starting control loop" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.312212 4955 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.312642 4955 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.312763 4955 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.312770 4955 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.321993 4955 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.322158 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:22 crc kubenswrapper[4955]: E0217 13:04:22.323942 4955 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.325721 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.325763 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.325776 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.326019 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.326929 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.326968 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.327829 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.327862 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.327873 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.328086 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.328476 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.328508 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.328521 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.328662 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.328694 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.332515 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.332537 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.332545 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.332723 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.332736 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.332745 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.332883 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.333136 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.333185 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.333883 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.333918 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.333927 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.334047 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.334221 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.334271 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.334289 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.334506 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.334535 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.335411 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.335447 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.335463 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.335665 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.335741 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.335772 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.335819 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.335903 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.337500 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.337553 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.337587 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:22 crc kubenswrapper[4955]: E0217 13:04:22.357352 4955 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.46:6443: connect: connection refused" interval="400ms" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.407919 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.408181 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.408278 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.408411 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.408506 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.408677 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.408752 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.408821 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.408848 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.408945 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.408991 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.409015 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.409170 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.409241 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.409278 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.412470 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.414395 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.414533 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.414643 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.414753 4955 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 17 13:04:22 crc kubenswrapper[4955]: E0217 13:04:22.415454 4955 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.46:6443: connect: connection refused" node="crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.510915 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.511018 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.511092 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.511135 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.511171 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.511205 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.511238 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.511269 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.511303 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.511333 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.511335 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.511441 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.511492 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.511363 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.511568 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.511627 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.511641 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.511670 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.511718 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.511762 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.511741 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.511832 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.511902 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.511909 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.512027 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.511835 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.512088 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.512131 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.512146 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.511870 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.615828 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.625031 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.625101 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.625124 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.625166 4955 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 17 13:04:22 crc kubenswrapper[4955]: E0217 13:04:22.625951 4955 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.46:6443: connect: connection refused" node="crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.658947 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.681823 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: W0217 13:04:22.710249 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-0857e69a69bc07f9d6c190b9d1e02cd21525407a74167867be409c37950c18f5 WatchSource:0}: Error finding container 0857e69a69bc07f9d6c190b9d1e02cd21525407a74167867be409c37950c18f5: Status 404 returned error can't find the container with id 0857e69a69bc07f9d6c190b9d1e02cd21525407a74167867be409c37950c18f5 Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.713820 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: W0217 13:04:22.718335 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-6ff0218431858f3a6b3096c2b86156ff45651621abbc10395d8369b823e20575 WatchSource:0}: Error finding container 6ff0218431858f3a6b3096c2b86156ff45651621abbc10395d8369b823e20575: Status 404 returned error can't find the container with id 6ff0218431858f3a6b3096c2b86156ff45651621abbc10395d8369b823e20575 Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.720833 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: I0217 13:04:22.726436 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 17 13:04:22 crc kubenswrapper[4955]: W0217 13:04:22.740939 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-0bc5e6f5b1da75646a60f46158e7cbe7a3278f25b6347b69f4ca60dc0c4a74d0 WatchSource:0}: Error finding container 0bc5e6f5b1da75646a60f46158e7cbe7a3278f25b6347b69f4ca60dc0c4a74d0: Status 404 returned error can't find the container with id 0bc5e6f5b1da75646a60f46158e7cbe7a3278f25b6347b69f4ca60dc0c4a74d0 Feb 17 13:04:22 crc kubenswrapper[4955]: W0217 13:04:22.742138 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-82e73b9d7dcf8e41baaad5b89e05c59d380842fd82180c243c12141dda4e822a WatchSource:0}: Error finding container 82e73b9d7dcf8e41baaad5b89e05c59d380842fd82180c243c12141dda4e822a: Status 404 returned error can't find the container with id 82e73b9d7dcf8e41baaad5b89e05c59d380842fd82180c243c12141dda4e822a Feb 17 13:04:22 crc kubenswrapper[4955]: E0217 13:04:22.759548 4955 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.46:6443: connect: connection refused" interval="800ms" Feb 17 13:04:22 crc kubenswrapper[4955]: W0217 13:04:22.767722 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-4345fd05137717522be2e12930fa518c0e9ed7126d469c9d59ed2a9ca270ab8d WatchSource:0}: Error finding container 4345fd05137717522be2e12930fa518c0e9ed7126d469c9d59ed2a9ca270ab8d: Status 404 returned error can't find the container with id 4345fd05137717522be2e12930fa518c0e9ed7126d469c9d59ed2a9ca270ab8d Feb 17 13:04:22 crc kubenswrapper[4955]: W0217 13:04:22.959977 4955 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.46:6443: connect: connection refused Feb 17 13:04:22 crc kubenswrapper[4955]: E0217 13:04:22.960090 4955 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.46:6443: connect: connection refused" logger="UnhandledError" Feb 17 13:04:23 crc kubenswrapper[4955]: I0217 13:04:23.026084 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:23 crc kubenswrapper[4955]: I0217 13:04:23.028272 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:23 crc kubenswrapper[4955]: I0217 13:04:23.028344 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:23 crc kubenswrapper[4955]: I0217 13:04:23.028368 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:23 crc kubenswrapper[4955]: I0217 13:04:23.028414 4955 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 17 13:04:23 crc kubenswrapper[4955]: E0217 13:04:23.029022 4955 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.46:6443: connect: connection refused" node="crc" Feb 17 13:04:23 crc kubenswrapper[4955]: W0217 13:04:23.143460 4955 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.46:6443: connect: connection refused Feb 17 13:04:23 crc kubenswrapper[4955]: E0217 13:04:23.143630 4955 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.46:6443: connect: connection refused" logger="UnhandledError" Feb 17 13:04:23 crc kubenswrapper[4955]: I0217 13:04:23.146423 4955 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.46:6443: connect: connection refused Feb 17 13:04:23 crc kubenswrapper[4955]: I0217 13:04:23.153493 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 16:35:26.58474147 +0000 UTC Feb 17 13:04:23 crc kubenswrapper[4955]: I0217 13:04:23.230958 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"82e73b9d7dcf8e41baaad5b89e05c59d380842fd82180c243c12141dda4e822a"} Feb 17 13:04:23 crc kubenswrapper[4955]: I0217 13:04:23.232400 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0bc5e6f5b1da75646a60f46158e7cbe7a3278f25b6347b69f4ca60dc0c4a74d0"} Feb 17 13:04:23 crc kubenswrapper[4955]: I0217 13:04:23.233533 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6ff0218431858f3a6b3096c2b86156ff45651621abbc10395d8369b823e20575"} Feb 17 13:04:23 crc kubenswrapper[4955]: I0217 13:04:23.234511 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"0857e69a69bc07f9d6c190b9d1e02cd21525407a74167867be409c37950c18f5"} Feb 17 13:04:23 crc kubenswrapper[4955]: I0217 13:04:23.235639 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"4345fd05137717522be2e12930fa518c0e9ed7126d469c9d59ed2a9ca270ab8d"} Feb 17 13:04:23 crc kubenswrapper[4955]: W0217 13:04:23.271085 4955 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.46:6443: connect: connection refused Feb 17 13:04:23 crc kubenswrapper[4955]: E0217 13:04:23.271209 4955 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.46:6443: connect: connection refused" logger="UnhandledError" Feb 17 13:04:23 crc kubenswrapper[4955]: E0217 13:04:23.560895 4955 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.46:6443: connect: connection refused" interval="1.6s" Feb 17 13:04:23 crc kubenswrapper[4955]: W0217 13:04:23.585979 4955 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.46:6443: connect: connection refused Feb 17 13:04:23 crc kubenswrapper[4955]: E0217 13:04:23.586153 4955 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.46:6443: connect: connection refused" logger="UnhandledError" Feb 17 13:04:23 crc kubenswrapper[4955]: I0217 13:04:23.829878 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:23 crc kubenswrapper[4955]: I0217 13:04:23.832125 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:23 crc kubenswrapper[4955]: I0217 13:04:23.832206 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:23 crc kubenswrapper[4955]: I0217 13:04:23.832232 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:23 crc kubenswrapper[4955]: I0217 13:04:23.832290 4955 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 17 13:04:23 crc kubenswrapper[4955]: E0217 13:04:23.833550 4955 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.46:6443: connect: connection refused" node="crc" Feb 17 13:04:24 crc kubenswrapper[4955]: I0217 13:04:24.046362 4955 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 17 13:04:24 crc kubenswrapper[4955]: E0217 13:04:24.047569 4955 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.46:6443: connect: connection refused" logger="UnhandledError" Feb 17 13:04:24 crc kubenswrapper[4955]: I0217 13:04:24.146470 4955 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.46:6443: connect: connection refused Feb 17 13:04:24 crc kubenswrapper[4955]: I0217 13:04:24.154603 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 11:36:41.969526334 +0000 UTC Feb 17 13:04:24 crc kubenswrapper[4955]: I0217 13:04:24.242662 4955 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="25f110bf0e9d951f4d48a491daae897a8250cf77fe464f1914cb2f1cfcde1d72" exitCode=0 Feb 17 13:04:24 crc kubenswrapper[4955]: I0217 13:04:24.242839 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"25f110bf0e9d951f4d48a491daae897a8250cf77fe464f1914cb2f1cfcde1d72"} Feb 17 13:04:24 crc kubenswrapper[4955]: I0217 13:04:24.242921 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:24 crc kubenswrapper[4955]: I0217 13:04:24.245081 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:24 crc kubenswrapper[4955]: I0217 13:04:24.245121 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:24 crc kubenswrapper[4955]: I0217 13:04:24.245140 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:24 crc kubenswrapper[4955]: I0217 13:04:24.247073 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9f43c7fe16acfd4d39301ee64665d4cd1d885397f981d70f73215da7f3aabe72"} Feb 17 13:04:24 crc kubenswrapper[4955]: I0217 13:04:24.247128 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0261f41b82b492d324b5e50edd72cedd5e824de542a7b47104abd2ab757a7049"} Feb 17 13:04:24 crc kubenswrapper[4955]: I0217 13:04:24.247150 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5"} Feb 17 13:04:24 crc kubenswrapper[4955]: I0217 13:04:24.250519 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e"} Feb 17 13:04:24 crc kubenswrapper[4955]: I0217 13:04:24.250541 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:24 crc kubenswrapper[4955]: I0217 13:04:24.250458 4955 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e" exitCode=0 Feb 17 13:04:24 crc kubenswrapper[4955]: I0217 13:04:24.252378 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:24 crc kubenswrapper[4955]: I0217 13:04:24.252440 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:24 crc kubenswrapper[4955]: I0217 13:04:24.252466 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:24 crc kubenswrapper[4955]: I0217 13:04:24.256593 4955 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1" exitCode=0 Feb 17 13:04:24 crc kubenswrapper[4955]: I0217 13:04:24.256727 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:24 crc kubenswrapper[4955]: I0217 13:04:24.256763 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1"} Feb 17 13:04:24 crc kubenswrapper[4955]: I0217 13:04:24.258714 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:24 crc kubenswrapper[4955]: I0217 13:04:24.258922 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:24 crc kubenswrapper[4955]: I0217 13:04:24.258992 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:24 crc kubenswrapper[4955]: I0217 13:04:24.259021 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:24 crc kubenswrapper[4955]: I0217 13:04:24.259824 4955 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="de99882bedf59812423fca8186a527f03eec5abe5777ff95b421b9d0330747d3" exitCode=0 Feb 17 13:04:24 crc kubenswrapper[4955]: I0217 13:04:24.259885 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"de99882bedf59812423fca8186a527f03eec5abe5777ff95b421b9d0330747d3"} Feb 17 13:04:24 crc kubenswrapper[4955]: I0217 13:04:24.259920 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:24 crc kubenswrapper[4955]: I0217 13:04:24.259914 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:24 crc kubenswrapper[4955]: I0217 13:04:24.259977 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:24 crc kubenswrapper[4955]: I0217 13:04:24.259997 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:24 crc kubenswrapper[4955]: I0217 13:04:24.262436 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:24 crc kubenswrapper[4955]: I0217 13:04:24.262483 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:24 crc kubenswrapper[4955]: I0217 13:04:24.262522 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:24 crc kubenswrapper[4955]: W0217 13:04:24.981444 4955 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.46:6443: connect: connection refused Feb 17 13:04:24 crc kubenswrapper[4955]: E0217 13:04:24.981566 4955 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.46:6443: connect: connection refused" logger="UnhandledError" Feb 17 13:04:25 crc kubenswrapper[4955]: I0217 13:04:25.145727 4955 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.46:6443: connect: connection refused Feb 17 13:04:25 crc kubenswrapper[4955]: I0217 13:04:25.154822 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 14:27:35.678189334 +0000 UTC Feb 17 13:04:25 crc kubenswrapper[4955]: E0217 13:04:25.162497 4955 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.46:6443: connect: connection refused" interval="3.2s" Feb 17 13:04:25 crc kubenswrapper[4955]: I0217 13:04:25.263707 4955 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a" exitCode=0 Feb 17 13:04:25 crc kubenswrapper[4955]: I0217 13:04:25.263757 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a"} Feb 17 13:04:25 crc kubenswrapper[4955]: I0217 13:04:25.263930 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:25 crc kubenswrapper[4955]: I0217 13:04:25.266664 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:25 crc kubenswrapper[4955]: I0217 13:04:25.266702 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:25 crc kubenswrapper[4955]: I0217 13:04:25.266717 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:25 crc kubenswrapper[4955]: I0217 13:04:25.270026 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:25 crc kubenswrapper[4955]: I0217 13:04:25.270027 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"cd99814f5b3ea3e8ad492e07f6050234b9394d3f775077207e41d44db4d63898"} Feb 17 13:04:25 crc kubenswrapper[4955]: I0217 13:04:25.270826 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:25 crc kubenswrapper[4955]: I0217 13:04:25.270868 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:25 crc kubenswrapper[4955]: I0217 13:04:25.270879 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:25 crc kubenswrapper[4955]: I0217 13:04:25.274158 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"0bbbc1ab7b069d1abd31929ec33bd494a7f403437f0be70da380ea0a5ff2e711"} Feb 17 13:04:25 crc kubenswrapper[4955]: I0217 13:04:25.274197 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"9c8e77568b5ec88d39510e2cdf7b9a3438a9c7cd0ad83cb350c9d7c927f0dbc4"} Feb 17 13:04:25 crc kubenswrapper[4955]: I0217 13:04:25.274211 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"3285726bf736fdd9d99c9974bef45c180410c4e3b795acbbf6c4ac576be3ee9b"} Feb 17 13:04:25 crc kubenswrapper[4955]: I0217 13:04:25.274306 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:25 crc kubenswrapper[4955]: I0217 13:04:25.275011 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:25 crc kubenswrapper[4955]: I0217 13:04:25.275033 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:25 crc kubenswrapper[4955]: I0217 13:04:25.275045 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:25 crc kubenswrapper[4955]: I0217 13:04:25.279953 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c871c1b454e4edd2f41076ecfd6a66592023d7161f094fbb4fd197fbf9c881f3"} Feb 17 13:04:25 crc kubenswrapper[4955]: I0217 13:04:25.279969 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:25 crc kubenswrapper[4955]: I0217 13:04:25.281632 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:25 crc kubenswrapper[4955]: I0217 13:04:25.281685 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:25 crc kubenswrapper[4955]: I0217 13:04:25.281697 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:25 crc kubenswrapper[4955]: I0217 13:04:25.284470 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3"} Feb 17 13:04:25 crc kubenswrapper[4955]: I0217 13:04:25.284511 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611"} Feb 17 13:04:25 crc kubenswrapper[4955]: I0217 13:04:25.284524 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5"} Feb 17 13:04:25 crc kubenswrapper[4955]: I0217 13:04:25.284537 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9"} Feb 17 13:04:25 crc kubenswrapper[4955]: I0217 13:04:25.449508 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:25 crc kubenswrapper[4955]: I0217 13:04:25.452514 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:25 crc kubenswrapper[4955]: I0217 13:04:25.452545 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:25 crc kubenswrapper[4955]: I0217 13:04:25.452555 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:25 crc kubenswrapper[4955]: I0217 13:04:25.452578 4955 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 17 13:04:25 crc kubenswrapper[4955]: E0217 13:04:25.453585 4955 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.46:6443: connect: connection refused" node="crc" Feb 17 13:04:25 crc kubenswrapper[4955]: W0217 13:04:25.545670 4955 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.46:6443: connect: connection refused Feb 17 13:04:25 crc kubenswrapper[4955]: E0217 13:04:25.545764 4955 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.46:6443: connect: connection refused" logger="UnhandledError" Feb 17 13:04:25 crc kubenswrapper[4955]: W0217 13:04:25.696622 4955 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.46:6443: connect: connection refused Feb 17 13:04:25 crc kubenswrapper[4955]: E0217 13:04:25.696718 4955 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.46:6443: connect: connection refused" logger="UnhandledError" Feb 17 13:04:25 crc kubenswrapper[4955]: W0217 13:04:25.718823 4955 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.46:6443: connect: connection refused Feb 17 13:04:25 crc kubenswrapper[4955]: E0217 13:04:25.718924 4955 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.46:6443: connect: connection refused" logger="UnhandledError" Feb 17 13:04:25 crc kubenswrapper[4955]: I0217 13:04:25.767229 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 17 13:04:26 crc kubenswrapper[4955]: I0217 13:04:26.155543 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 04:57:19.492568669 +0000 UTC Feb 17 13:04:26 crc kubenswrapper[4955]: I0217 13:04:26.295187 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472"} Feb 17 13:04:26 crc kubenswrapper[4955]: I0217 13:04:26.295302 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:26 crc kubenswrapper[4955]: I0217 13:04:26.302271 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:26 crc kubenswrapper[4955]: I0217 13:04:26.302343 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:26 crc kubenswrapper[4955]: I0217 13:04:26.302365 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:26 crc kubenswrapper[4955]: I0217 13:04:26.305305 4955 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350" exitCode=0 Feb 17 13:04:26 crc kubenswrapper[4955]: I0217 13:04:26.305448 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:26 crc kubenswrapper[4955]: I0217 13:04:26.305524 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350"} Feb 17 13:04:26 crc kubenswrapper[4955]: I0217 13:04:26.305578 4955 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 17 13:04:26 crc kubenswrapper[4955]: I0217 13:04:26.305625 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:26 crc kubenswrapper[4955]: I0217 13:04:26.305766 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:26 crc kubenswrapper[4955]: I0217 13:04:26.305638 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:26 crc kubenswrapper[4955]: I0217 13:04:26.306862 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:26 crc kubenswrapper[4955]: I0217 13:04:26.306910 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:26 crc kubenswrapper[4955]: I0217 13:04:26.306927 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:26 crc kubenswrapper[4955]: I0217 13:04:26.307538 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:26 crc kubenswrapper[4955]: I0217 13:04:26.307589 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:26 crc kubenswrapper[4955]: I0217 13:04:26.307606 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:26 crc kubenswrapper[4955]: I0217 13:04:26.307540 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:26 crc kubenswrapper[4955]: I0217 13:04:26.307855 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:26 crc kubenswrapper[4955]: I0217 13:04:26.307882 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:26 crc kubenswrapper[4955]: I0217 13:04:26.309215 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:26 crc kubenswrapper[4955]: I0217 13:04:26.309255 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:26 crc kubenswrapper[4955]: I0217 13:04:26.309275 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:26 crc kubenswrapper[4955]: I0217 13:04:26.573646 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 17 13:04:27 crc kubenswrapper[4955]: I0217 13:04:27.156480 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 04:07:05.154457009 +0000 UTC Feb 17 13:04:27 crc kubenswrapper[4955]: I0217 13:04:27.313993 4955 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 17 13:04:27 crc kubenswrapper[4955]: I0217 13:04:27.314043 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:27 crc kubenswrapper[4955]: I0217 13:04:27.314029 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"96e38466a067b20ca4323ab6983cb324bb63179f942d2c37c0daa8b03767e7ac"} Feb 17 13:04:27 crc kubenswrapper[4955]: I0217 13:04:27.314181 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:27 crc kubenswrapper[4955]: I0217 13:04:27.314233 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"7d6919c91918e8bacafd1442b06d14ffd9b547f0f37b2298e2d8d3acc5962d04"} Feb 17 13:04:27 crc kubenswrapper[4955]: I0217 13:04:27.314278 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"49f1fee7ac5f8d1442d9f63b0e8d285187129eb3a3e3b50a49a45c7a7a39ae8d"} Feb 17 13:04:27 crc kubenswrapper[4955]: I0217 13:04:27.315116 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:27 crc kubenswrapper[4955]: I0217 13:04:27.315166 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:27 crc kubenswrapper[4955]: I0217 13:04:27.315129 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:27 crc kubenswrapper[4955]: I0217 13:04:27.315255 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:27 crc kubenswrapper[4955]: I0217 13:04:27.315274 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:27 crc kubenswrapper[4955]: I0217 13:04:27.315191 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:28 crc kubenswrapper[4955]: I0217 13:04:28.106683 4955 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 17 13:04:28 crc kubenswrapper[4955]: I0217 13:04:28.157185 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 22:08:41.869746179 +0000 UTC Feb 17 13:04:28 crc kubenswrapper[4955]: I0217 13:04:28.308295 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 13:04:28 crc kubenswrapper[4955]: I0217 13:04:28.324592 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a001dfb595f39d1da33b265474b6cca51c62f522ff061c9e7c52f93806034f4d"} Feb 17 13:04:28 crc kubenswrapper[4955]: I0217 13:04:28.324684 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4830450cbada3d24c9747e9de5f48e53e118549f1dd67edf7fff8b2adc47128f"} Feb 17 13:04:28 crc kubenswrapper[4955]: I0217 13:04:28.324817 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:28 crc kubenswrapper[4955]: I0217 13:04:28.324848 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:28 crc kubenswrapper[4955]: I0217 13:04:28.324861 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:28 crc kubenswrapper[4955]: I0217 13:04:28.326704 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:28 crc kubenswrapper[4955]: I0217 13:04:28.326755 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:28 crc kubenswrapper[4955]: I0217 13:04:28.326772 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:28 crc kubenswrapper[4955]: I0217 13:04:28.326725 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:28 crc kubenswrapper[4955]: I0217 13:04:28.326892 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:28 crc kubenswrapper[4955]: I0217 13:04:28.326909 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:28 crc kubenswrapper[4955]: I0217 13:04:28.326916 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:28 crc kubenswrapper[4955]: I0217 13:04:28.326957 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:28 crc kubenswrapper[4955]: I0217 13:04:28.327016 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:28 crc kubenswrapper[4955]: I0217 13:04:28.578977 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 17 13:04:28 crc kubenswrapper[4955]: I0217 13:04:28.587176 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 17 13:04:28 crc kubenswrapper[4955]: I0217 13:04:28.653990 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:28 crc kubenswrapper[4955]: I0217 13:04:28.656148 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:28 crc kubenswrapper[4955]: I0217 13:04:28.656205 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:28 crc kubenswrapper[4955]: I0217 13:04:28.656253 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:28 crc kubenswrapper[4955]: I0217 13:04:28.656297 4955 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 17 13:04:29 crc kubenswrapper[4955]: I0217 13:04:29.157750 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 15:23:13.66097759 +0000 UTC Feb 17 13:04:29 crc kubenswrapper[4955]: I0217 13:04:29.189158 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 17 13:04:29 crc kubenswrapper[4955]: I0217 13:04:29.189417 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:29 crc kubenswrapper[4955]: I0217 13:04:29.191213 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:29 crc kubenswrapper[4955]: I0217 13:04:29.191264 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:29 crc kubenswrapper[4955]: I0217 13:04:29.191288 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:29 crc kubenswrapper[4955]: I0217 13:04:29.327274 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:29 crc kubenswrapper[4955]: I0217 13:04:29.327394 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:29 crc kubenswrapper[4955]: I0217 13:04:29.330280 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:29 crc kubenswrapper[4955]: I0217 13:04:29.330337 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:29 crc kubenswrapper[4955]: I0217 13:04:29.330357 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:29 crc kubenswrapper[4955]: I0217 13:04:29.331898 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:29 crc kubenswrapper[4955]: I0217 13:04:29.331967 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:29 crc kubenswrapper[4955]: I0217 13:04:29.332001 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:29 crc kubenswrapper[4955]: I0217 13:04:29.499846 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 13:04:29 crc kubenswrapper[4955]: I0217 13:04:29.500073 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:29 crc kubenswrapper[4955]: I0217 13:04:29.501684 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:29 crc kubenswrapper[4955]: I0217 13:04:29.501746 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:29 crc kubenswrapper[4955]: I0217 13:04:29.501772 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:29 crc kubenswrapper[4955]: I0217 13:04:29.537325 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Feb 17 13:04:30 crc kubenswrapper[4955]: I0217 13:04:30.158499 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 05:58:18.153962331 +0000 UTC Feb 17 13:04:30 crc kubenswrapper[4955]: I0217 13:04:30.330178 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:30 crc kubenswrapper[4955]: I0217 13:04:30.330246 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:30 crc kubenswrapper[4955]: I0217 13:04:30.331825 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:30 crc kubenswrapper[4955]: I0217 13:04:30.331894 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:30 crc kubenswrapper[4955]: I0217 13:04:30.331919 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:30 crc kubenswrapper[4955]: I0217 13:04:30.331933 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:30 crc kubenswrapper[4955]: I0217 13:04:30.331978 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:30 crc kubenswrapper[4955]: I0217 13:04:30.332000 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:30 crc kubenswrapper[4955]: I0217 13:04:30.740761 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 13:04:30 crc kubenswrapper[4955]: I0217 13:04:30.741047 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:30 crc kubenswrapper[4955]: I0217 13:04:30.742691 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:30 crc kubenswrapper[4955]: I0217 13:04:30.742747 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:30 crc kubenswrapper[4955]: I0217 13:04:30.742773 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:31 crc kubenswrapper[4955]: I0217 13:04:31.159731 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 00:22:04.362926564 +0000 UTC Feb 17 13:04:32 crc kubenswrapper[4955]: I0217 13:04:32.160198 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 00:58:52.901775768 +0000 UTC Feb 17 13:04:32 crc kubenswrapper[4955]: E0217 13:04:32.324087 4955 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 17 13:04:33 crc kubenswrapper[4955]: I0217 13:04:33.161017 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-03 00:34:42.931202351 +0000 UTC Feb 17 13:04:33 crc kubenswrapper[4955]: I0217 13:04:33.865075 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 17 13:04:33 crc kubenswrapper[4955]: I0217 13:04:33.865339 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:33 crc kubenswrapper[4955]: I0217 13:04:33.866850 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:33 crc kubenswrapper[4955]: I0217 13:04:33.866892 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:33 crc kubenswrapper[4955]: I0217 13:04:33.866910 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:33 crc kubenswrapper[4955]: I0217 13:04:33.872338 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 17 13:04:34 crc kubenswrapper[4955]: I0217 13:04:34.162236 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 08:51:14.355883121 +0000 UTC Feb 17 13:04:34 crc kubenswrapper[4955]: I0217 13:04:34.340896 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:34 crc kubenswrapper[4955]: I0217 13:04:34.345693 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:34 crc kubenswrapper[4955]: I0217 13:04:34.345762 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:34 crc kubenswrapper[4955]: I0217 13:04:34.345841 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:35 crc kubenswrapper[4955]: I0217 13:04:35.162460 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 04:10:45.550583748 +0000 UTC Feb 17 13:04:35 crc kubenswrapper[4955]: I0217 13:04:35.615035 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Feb 17 13:04:35 crc kubenswrapper[4955]: I0217 13:04:35.615324 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:35 crc kubenswrapper[4955]: I0217 13:04:35.616849 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:35 crc kubenswrapper[4955]: I0217 13:04:35.616897 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:35 crc kubenswrapper[4955]: I0217 13:04:35.616914 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:36 crc kubenswrapper[4955]: I0217 13:04:36.146860 4955 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Feb 17 13:04:36 crc kubenswrapper[4955]: I0217 13:04:36.163333 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 22:15:31.81057064 +0000 UTC Feb 17 13:04:36 crc kubenswrapper[4955]: I0217 13:04:36.865804 4955 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 17 13:04:36 crc kubenswrapper[4955]: I0217 13:04:36.865893 4955 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 17 13:04:37 crc kubenswrapper[4955]: I0217 13:04:37.163573 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 17:05:18.381346706 +0000 UTC Feb 17 13:04:37 crc kubenswrapper[4955]: I0217 13:04:37.375759 4955 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 17 13:04:37 crc kubenswrapper[4955]: I0217 13:04:37.375887 4955 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 17 13:04:37 crc kubenswrapper[4955]: I0217 13:04:37.384843 4955 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 17 13:04:37 crc kubenswrapper[4955]: I0217 13:04:37.384881 4955 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 17 13:04:38 crc kubenswrapper[4955]: I0217 13:04:38.164070 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 18:31:00.866694202 +0000 UTC Feb 17 13:04:39 crc kubenswrapper[4955]: I0217 13:04:39.164799 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 20:24:44.855851406 +0000 UTC Feb 17 13:04:39 crc kubenswrapper[4955]: I0217 13:04:39.508959 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 13:04:39 crc kubenswrapper[4955]: I0217 13:04:39.510907 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:39 crc kubenswrapper[4955]: I0217 13:04:39.513011 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:39 crc kubenswrapper[4955]: I0217 13:04:39.513090 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:39 crc kubenswrapper[4955]: I0217 13:04:39.513109 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:39 crc kubenswrapper[4955]: I0217 13:04:39.518332 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 13:04:40 crc kubenswrapper[4955]: I0217 13:04:40.165999 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 06:06:00.937085449 +0000 UTC Feb 17 13:04:40 crc kubenswrapper[4955]: I0217 13:04:40.374119 4955 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 17 13:04:40 crc kubenswrapper[4955]: I0217 13:04:40.374185 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:40 crc kubenswrapper[4955]: I0217 13:04:40.375289 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:40 crc kubenswrapper[4955]: I0217 13:04:40.375363 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:40 crc kubenswrapper[4955]: I0217 13:04:40.375380 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:41 crc kubenswrapper[4955]: I0217 13:04:41.167426 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 12:37:57.869129391 +0000 UTC Feb 17 13:04:42 crc kubenswrapper[4955]: I0217 13:04:42.168532 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 01:07:46.944706165 +0000 UTC Feb 17 13:04:42 crc kubenswrapper[4955]: E0217 13:04:42.324431 4955 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 17 13:04:42 crc kubenswrapper[4955]: E0217 13:04:42.352585 4955 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Feb 17 13:04:42 crc kubenswrapper[4955]: I0217 13:04:42.355482 4955 trace.go:236] Trace[1012022717]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (17-Feb-2026 13:04:31.330) (total time: 11024ms): Feb 17 13:04:42 crc kubenswrapper[4955]: Trace[1012022717]: ---"Objects listed" error: 11024ms (13:04:42.355) Feb 17 13:04:42 crc kubenswrapper[4955]: Trace[1012022717]: [11.024639059s] [11.024639059s] END Feb 17 13:04:42 crc kubenswrapper[4955]: I0217 13:04:42.355527 4955 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 17 13:04:42 crc kubenswrapper[4955]: I0217 13:04:42.356882 4955 trace.go:236] Trace[553023536]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (17-Feb-2026 13:04:30.076) (total time: 12280ms): Feb 17 13:04:42 crc kubenswrapper[4955]: Trace[553023536]: ---"Objects listed" error: 12280ms (13:04:42.356) Feb 17 13:04:42 crc kubenswrapper[4955]: Trace[553023536]: [12.2803986s] [12.2803986s] END Feb 17 13:04:42 crc kubenswrapper[4955]: I0217 13:04:42.356909 4955 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 17 13:04:42 crc kubenswrapper[4955]: E0217 13:04:42.357732 4955 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Feb 17 13:04:42 crc kubenswrapper[4955]: I0217 13:04:42.359202 4955 trace.go:236] Trace[1448135921]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (17-Feb-2026 13:04:31.269) (total time: 11089ms): Feb 17 13:04:42 crc kubenswrapper[4955]: Trace[1448135921]: ---"Objects listed" error: 11089ms (13:04:42.359) Feb 17 13:04:42 crc kubenswrapper[4955]: Trace[1448135921]: [11.089624669s] [11.089624669s] END Feb 17 13:04:42 crc kubenswrapper[4955]: I0217 13:04:42.359248 4955 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 17 13:04:42 crc kubenswrapper[4955]: I0217 13:04:42.359835 4955 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Feb 17 13:04:42 crc kubenswrapper[4955]: I0217 13:04:42.360607 4955 trace.go:236] Trace[1436567725]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (17-Feb-2026 13:04:29.690) (total time: 12669ms): Feb 17 13:04:42 crc kubenswrapper[4955]: Trace[1436567725]: ---"Objects listed" error: 12669ms (13:04:42.360) Feb 17 13:04:42 crc kubenswrapper[4955]: Trace[1436567725]: [12.669986699s] [12.669986699s] END Feb 17 13:04:42 crc kubenswrapper[4955]: I0217 13:04:42.360642 4955 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 17 13:04:42 crc kubenswrapper[4955]: I0217 13:04:42.380413 4955 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 17 13:04:42 crc kubenswrapper[4955]: I0217 13:04:42.409359 4955 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:48214->192.168.126.11:17697: read: connection reset by peer" start-of-body= Feb 17 13:04:42 crc kubenswrapper[4955]: I0217 13:04:42.409359 4955 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:48202->192.168.126.11:17697: read: connection reset by peer" start-of-body= Feb 17 13:04:42 crc kubenswrapper[4955]: I0217 13:04:42.409471 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:48214->192.168.126.11:17697: read: connection reset by peer" Feb 17 13:04:42 crc kubenswrapper[4955]: I0217 13:04:42.409586 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:48202->192.168.126.11:17697: read: connection reset by peer" Feb 17 13:04:42 crc kubenswrapper[4955]: I0217 13:04:42.410008 4955 csr.go:261] certificate signing request csr-pzv4f is approved, waiting to be issued Feb 17 13:04:42 crc kubenswrapper[4955]: I0217 13:04:42.410133 4955 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Feb 17 13:04:42 crc kubenswrapper[4955]: I0217 13:04:42.410203 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Feb 17 13:04:42 crc kubenswrapper[4955]: I0217 13:04:42.424056 4955 csr.go:257] certificate signing request csr-pzv4f is issued Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.142188 4955 apiserver.go:52] "Watching apiserver" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.144987 4955 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.145319 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h"] Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.145751 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.145808 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.145878 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:04:43 crc kubenswrapper[4955]: E0217 13:04:43.145954 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.145968 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.145990 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 17 13:04:43 crc kubenswrapper[4955]: E0217 13:04:43.146016 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.146053 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 17 13:04:43 crc kubenswrapper[4955]: E0217 13:04:43.146071 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.148387 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.148849 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.149019 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.149375 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.149541 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.149602 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.149811 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.149887 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.150009 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.155828 4955 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.166195 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.166255 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.166289 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.166321 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.166348 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.166370 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.166392 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.166419 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.166444 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.166468 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.166492 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.166516 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.166540 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.166557 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.166580 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.166601 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.166626 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.166718 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.166751 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.166813 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.166836 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.166859 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.166878 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.166899 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.166942 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.166960 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.166978 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.166999 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167020 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167038 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167057 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167075 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167096 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167119 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167145 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167146 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167165 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167237 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167260 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167277 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167314 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167331 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167333 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167350 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167386 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167404 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167423 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167439 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167455 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167470 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167488 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167504 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167522 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167542 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167561 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167578 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167594 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167610 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167625 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167641 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167660 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167681 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167701 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167717 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167737 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167758 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167793 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167819 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167841 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167865 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167887 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167907 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167924 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167941 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167957 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167972 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167988 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.168005 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.168020 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.168037 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.168053 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.168069 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.168086 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.168103 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.168119 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.168138 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.168160 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.168375 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.168402 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.168422 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.168533 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.168584 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.168609 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.168629 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.168647 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.168668 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.168688 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.168735 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.168755 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.168832 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.168879 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.168902 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.168924 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.168947 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.169749 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.169809 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.169832 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.169855 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.169877 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.169897 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.169920 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.169941 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.170347 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.170396 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.170417 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.170436 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.170455 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.170474 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.170496 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.170518 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.170539 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.170604 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.170624 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.170643 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.170664 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.170719 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.170742 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.170761 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.170812 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.170835 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.170854 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.170873 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.170891 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.170914 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.170935 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.170952 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.170970 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.170989 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171008 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171054 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171073 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171092 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171109 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171131 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171151 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171170 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171188 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171240 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171258 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171275 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171296 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171313 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171332 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171350 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171367 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171385 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171405 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171428 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171446 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171472 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171491 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171511 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171530 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171547 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171565 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171584 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171603 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171621 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171642 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171661 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171677 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171694 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171710 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171728 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171745 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171768 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171807 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171824 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167485 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167623 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167769 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.167974 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.168460 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.168607 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.168751 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.168816 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.168800 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 12:21:55.327662664 +0000 UTC Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.169022 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.169459 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.169467 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.169624 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.169861 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.170728 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171992 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171401 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171537 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171696 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171831 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.172084 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.172091 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.172180 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.172269 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.172499 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.172508 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.172502 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.172702 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.173354 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.173454 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.173567 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.173702 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.173919 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.174057 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.174277 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.174334 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.174421 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.174542 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.174830 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.174861 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.171840 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.174908 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.174931 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.174949 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175028 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175047 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175066 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175086 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175105 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175110 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175123 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175140 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175157 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175176 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175194 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175210 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175227 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175243 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175282 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175301 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175337 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175357 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175383 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175404 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175426 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175447 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175469 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175487 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175504 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175523 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175523 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175544 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175562 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175583 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175602 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175619 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175682 4955 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175695 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175698 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175705 4955 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175732 4955 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175751 4955 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175764 4955 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175796 4955 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175812 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175824 4955 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175837 4955 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175849 4955 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175882 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175897 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175911 4955 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175923 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175932 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175927 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.175943 4955 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.176017 4955 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.176053 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.176071 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.176088 4955 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.176108 4955 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.176123 4955 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.176221 4955 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.176238 4955 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.176238 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.176252 4955 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.176276 4955 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.176292 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.176305 4955 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.176345 4955 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.176356 4955 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.176367 4955 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.176365 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.176379 4955 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.176389 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.176402 4955 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.176412 4955 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.176423 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.176435 4955 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.176447 4955 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.176457 4955 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.176467 4955 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.176477 4955 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.176573 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.176814 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.176860 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.177052 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.177099 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.177282 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.177409 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.177477 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.178038 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.178356 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.178594 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.179193 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.179275 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.179345 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.179461 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.179464 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.180451 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.180901 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.181170 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.181252 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.181494 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.181670 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.181695 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.182080 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.182104 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.182551 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.182574 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.182597 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.182658 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.183080 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.183089 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.183289 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.183531 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.183667 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.183703 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.183841 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.183986 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.184124 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.184189 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.184240 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.184400 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.184544 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.185016 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.184905 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.185308 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.185470 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.185534 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.185568 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.185706 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.185775 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.185928 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.186068 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: E0217 13:04:43.186251 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:04:43.686226028 +0000 UTC m=+22.208955771 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.186259 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.186565 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.186761 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.187040 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.187574 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.188061 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.188426 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.189249 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.189888 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.190423 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.190929 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.191385 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.191732 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.193154 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.193524 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.193824 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.194501 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.194860 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.194920 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.195143 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.195494 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.196238 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.196525 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.196764 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.197092 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.197362 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.197606 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.201956 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.202314 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.202971 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.202982 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.203166 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.203606 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.204053 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.204252 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.204844 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.204932 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.205074 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.205384 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.205969 4955 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.206480 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.206890 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.207045 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.207142 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.207327 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.207420 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.207696 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.207739 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.207845 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.214234 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.215020 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.215061 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.215577 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.215936 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.216303 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.216442 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.216585 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.216854 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: E0217 13:04:43.218217 4955 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 17 13:04:43 crc kubenswrapper[4955]: E0217 13:04:43.218361 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-17 13:04:43.718340253 +0000 UTC m=+22.241069856 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.221315 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.221721 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.221949 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.222775 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.223051 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.227849 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.230175 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.230859 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.231200 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.231248 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.231311 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.231335 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.232211 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.233793 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.230478 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: E0217 13:04:43.236231 4955 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 17 13:04:43 crc kubenswrapper[4955]: E0217 13:04:43.236260 4955 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 17 13:04:43 crc kubenswrapper[4955]: E0217 13:04:43.236277 4955 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.237128 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.237412 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.237725 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.238582 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.239358 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: E0217 13:04:43.239628 4955 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 17 13:04:43 crc kubenswrapper[4955]: E0217 13:04:43.239648 4955 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 17 13:04:43 crc kubenswrapper[4955]: E0217 13:04:43.239666 4955 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 13:04:43 crc kubenswrapper[4955]: E0217 13:04:43.239756 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-17 13:04:43.739726558 +0000 UTC m=+22.262456101 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 13:04:43 crc kubenswrapper[4955]: E0217 13:04:43.239832 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-17 13:04:43.739824061 +0000 UTC m=+22.262553604 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.240049 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 17 13:04:43 crc kubenswrapper[4955]: E0217 13:04:43.242464 4955 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 17 13:04:43 crc kubenswrapper[4955]: E0217 13:04:43.242566 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-17 13:04:43.742543457 +0000 UTC m=+22.265273000 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.243490 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.245190 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.247443 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.245429 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.250703 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.250797 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.251109 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.251119 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.251574 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.251677 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.251863 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.251969 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.252106 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.252510 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.252570 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.252717 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.253599 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.255236 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.255540 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.257215 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.262035 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.265260 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.273208 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.273983 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.274584 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.274654 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277018 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277054 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277080 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277135 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277156 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277173 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277183 4955 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277193 4955 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277204 4955 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277324 4955 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277338 4955 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277349 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277359 4955 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277368 4955 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277377 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277386 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277395 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277405 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277419 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277428 4955 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277437 4955 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277449 4955 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277492 4955 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277503 4955 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277515 4955 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277524 4955 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277534 4955 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277543 4955 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277554 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277563 4955 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277579 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277591 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277600 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277609 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277618 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277627 4955 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277635 4955 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277646 4955 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277655 4955 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277665 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277676 4955 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277691 4955 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277700 4955 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277708 4955 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277717 4955 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277725 4955 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277734 4955 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277743 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277752 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277761 4955 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277769 4955 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277777 4955 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277801 4955 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277810 4955 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277819 4955 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277827 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277835 4955 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277843 4955 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277852 4955 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277863 4955 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277874 4955 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277884 4955 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277893 4955 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277903 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277915 4955 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277924 4955 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277934 4955 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277945 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277954 4955 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277965 4955 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277974 4955 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277983 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.277993 4955 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278001 4955 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278010 4955 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278018 4955 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278028 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278038 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278047 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278056 4955 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278066 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278075 4955 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278086 4955 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278095 4955 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278104 4955 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278114 4955 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278123 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278132 4955 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278141 4955 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278150 4955 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278159 4955 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278170 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278179 4955 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278191 4955 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278200 4955 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278209 4955 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278218 4955 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278228 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278237 4955 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278246 4955 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278255 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278264 4955 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278274 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278283 4955 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278292 4955 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278302 4955 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278311 4955 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278320 4955 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278329 4955 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278339 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278349 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278358 4955 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278367 4955 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278376 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278386 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278394 4955 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278402 4955 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278458 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278470 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278479 4955 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278488 4955 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278497 4955 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278507 4955 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278537 4955 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278545 4955 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278557 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278568 4955 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278577 4955 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278587 4955 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278599 4955 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278609 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278619 4955 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278628 4955 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278639 4955 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278648 4955 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278657 4955 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278666 4955 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278675 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278685 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278694 4955 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278704 4955 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278712 4955 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278721 4955 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278732 4955 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278741 4955 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278750 4955 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278759 4955 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278769 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278796 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278806 4955 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278816 4955 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278825 4955 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278834 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.278843 4955 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.292350 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.296996 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.305571 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.307495 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.326307 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.334552 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.344219 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.355526 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.368897 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.378949 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.379439 4955 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.379461 4955 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.385774 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.388852 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.391904 4955 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472" exitCode=255 Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.391966 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472"} Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.393809 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"fb91e7f943fd299c092b22f66029faf568d1c0218db440b126b4afed12f209ee"} Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.406727 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.410487 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.410909 4955 scope.go:117] "RemoveContainer" containerID="d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.426261 4955 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-17 12:59:42 +0000 UTC, rotation deadline is 2026-11-19 06:16:14.946626062 +0000 UTC Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.426515 4955 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6593h11m31.520117273s for next certificate rotation Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.426928 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.448742 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.460878 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.474441 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.477584 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.487833 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.489291 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 17 13:04:43 crc kubenswrapper[4955]: W0217 13:04:43.507985 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-ae3e65d5825e413b4fe1b062bce962e7e861d6aa74283ed1bb82f974fd6bc828 WatchSource:0}: Error finding container ae3e65d5825e413b4fe1b062bce962e7e861d6aa74283ed1bb82f974fd6bc828: Status 404 returned error can't find the container with id ae3e65d5825e413b4fe1b062bce962e7e861d6aa74283ed1bb82f974fd6bc828 Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.782376 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.782473 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.782514 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.782547 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:04:43 crc kubenswrapper[4955]: E0217 13:04:43.782656 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:04:44.782608945 +0000 UTC m=+23.305338648 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:04:43 crc kubenswrapper[4955]: E0217 13:04:43.782670 4955 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 17 13:04:43 crc kubenswrapper[4955]: E0217 13:04:43.782749 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-17 13:04:44.782729139 +0000 UTC m=+23.305458692 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.782777 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:04:43 crc kubenswrapper[4955]: E0217 13:04:43.782811 4955 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 17 13:04:43 crc kubenswrapper[4955]: E0217 13:04:43.782855 4955 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 17 13:04:43 crc kubenswrapper[4955]: E0217 13:04:43.782864 4955 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 17 13:04:43 crc kubenswrapper[4955]: E0217 13:04:43.782814 4955 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 17 13:04:43 crc kubenswrapper[4955]: E0217 13:04:43.782889 4955 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 17 13:04:43 crc kubenswrapper[4955]: E0217 13:04:43.782899 4955 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 13:04:43 crc kubenswrapper[4955]: E0217 13:04:43.782903 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-17 13:04:44.782893603 +0000 UTC m=+23.305623156 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 17 13:04:43 crc kubenswrapper[4955]: E0217 13:04:43.782870 4955 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 13:04:43 crc kubenswrapper[4955]: E0217 13:04:43.782973 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-17 13:04:44.782949465 +0000 UTC m=+23.305679008 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 13:04:43 crc kubenswrapper[4955]: E0217 13:04:43.783022 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-17 13:04:44.782997406 +0000 UTC m=+23.305726949 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.873688 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.880619 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.885722 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.895409 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.906840 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.909612 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.918845 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05aa39d-94e6-49d4-885e-2c1f986310f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T13:04:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 13:04:35.885968 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 13:04:35.888176 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3462920193/tls.crt::/tmp/serving-cert-3462920193/tls.key\\\\\\\"\\\\nI0217 13:04:42.370546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 13:04:42.384242 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 13:04:42.384284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 13:04:42.384317 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 13:04:42.384327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 13:04:42.394555 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 13:04:42.394614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394628 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 13:04:42.394646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 13:04:42.394654 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 13:04:42.394662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 13:04:42.394900 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 13:04:42.396494 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.936015 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.951474 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.974557 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:43 crc kubenswrapper[4955]: I0217 13:04:43.989389 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.010311 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05aa39d-94e6-49d4-885e-2c1f986310f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T13:04:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 13:04:35.885968 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 13:04:35.888176 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3462920193/tls.crt::/tmp/serving-cert-3462920193/tls.key\\\\\\\"\\\\nI0217 13:04:42.370546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 13:04:42.384242 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 13:04:42.384284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 13:04:42.384317 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 13:04:42.384327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 13:04:42.394555 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 13:04:42.394614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394628 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 13:04:42.394646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 13:04:42.394654 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 13:04:42.394662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 13:04:42.394900 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 13:04:42.396494 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.020979 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.031632 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.044453 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.060453 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.099078 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5c34a3d-82ad-4d03-abdd-0a20818ac2fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0261f41b82b492d324b5e50edd72cedd5e824de542a7b47104abd2ab757a7049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f43c7fe16acfd4d39301ee64665d4cd1d885397f981d70f73215da7f3aabe72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c871c1b454e4edd2f41076ecfd6a66592023d7161f094fbb4fd197fbf9c881f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.117520 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.172728 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 05:07:13.398447869 +0000 UTC Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.189014 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-fp5j6"] Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.189606 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.190291 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-29qxq"] Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.190863 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.194420 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.194602 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.194775 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.194880 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.195292 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.195422 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.201808 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.202102 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.202420 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.203093 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.204399 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-9hx9f"] Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.204895 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.205881 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-msn6p"] Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.206145 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-msn6p" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.207296 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.207441 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.207974 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.208084 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.208427 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.220301 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05aa39d-94e6-49d4-885e-2c1f986310f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T13:04:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 13:04:35.885968 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 13:04:35.888176 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3462920193/tls.crt::/tmp/serving-cert-3462920193/tls.key\\\\\\\"\\\\nI0217 13:04:42.370546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 13:04:42.384242 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 13:04:42.384284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 13:04:42.384317 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 13:04:42.384327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 13:04:42.394555 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 13:04:42.394614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394628 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 13:04:42.394646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 13:04:42.394654 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 13:04:42.394662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 13:04:42.394900 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 13:04:42.396494 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.226053 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.226590 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.227900 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.228532 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.229556 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.230067 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.230689 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.232490 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.233041 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.233168 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.234203 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.234733 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.236768 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.237421 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.238015 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.238575 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.239164 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.240059 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.240621 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.241460 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.242643 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.243181 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.244471 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.244981 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.245888 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.246991 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.247698 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.252108 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.252603 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.253357 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.256349 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.256927 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.257437 4955 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.257554 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.258999 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.259512 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.261245 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.263603 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.264475 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.265691 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.266553 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.267934 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.268654 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.268719 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.270191 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.270925 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.272066 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.272680 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.273703 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.274291 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.275512 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.276173 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.277260 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.278203 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.278761 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.279870 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.280529 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.281702 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"924c66b7-7c3c-4db7-a427-3f152efbac4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fp5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.286749 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/5bf36887-2a2b-4193-abdb-a15beaf46300-rootfs\") pod \"machine-config-daemon-29qxq\" (UID: \"5bf36887-2a2b-4193-abdb-a15beaf46300\") " pod="openshift-machine-config-operator/machine-config-daemon-29qxq" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.286855 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/1e57bdf8-e537-402a-9815-609fecdd68ea-host-var-lib-kubelet\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.286885 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5bf36887-2a2b-4193-abdb-a15beaf46300-proxy-tls\") pod \"machine-config-daemon-29qxq\" (UID: \"5bf36887-2a2b-4193-abdb-a15beaf46300\") " pod="openshift-machine-config-operator/machine-config-daemon-29qxq" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.286909 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/1e57bdf8-e537-402a-9815-609fecdd68ea-hostroot\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.286953 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/924c66b7-7c3c-4db7-a427-3f152efbac4e-system-cni-dir\") pod \"multus-additional-cni-plugins-fp5j6\" (UID: \"924c66b7-7c3c-4db7-a427-3f152efbac4e\") " pod="openshift-multus/multus-additional-cni-plugins-fp5j6" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.286985 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q22d4\" (UniqueName: \"kubernetes.io/projected/1e57bdf8-e537-402a-9815-609fecdd68ea-kube-api-access-q22d4\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.287018 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/1e57bdf8-e537-402a-9815-609fecdd68ea-host-run-multus-certs\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.287038 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/924c66b7-7c3c-4db7-a427-3f152efbac4e-os-release\") pod \"multus-additional-cni-plugins-fp5j6\" (UID: \"924c66b7-7c3c-4db7-a427-3f152efbac4e\") " pod="openshift-multus/multus-additional-cni-plugins-fp5j6" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.287063 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/1e57bdf8-e537-402a-9815-609fecdd68ea-multus-socket-dir-parent\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.287092 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1e57bdf8-e537-402a-9815-609fecdd68ea-host-var-lib-cni-bin\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.287118 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1e57bdf8-e537-402a-9815-609fecdd68ea-multus-conf-dir\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.287144 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/924c66b7-7c3c-4db7-a427-3f152efbac4e-tuning-conf-dir\") pod \"multus-additional-cni-plugins-fp5j6\" (UID: \"924c66b7-7c3c-4db7-a427-3f152efbac4e\") " pod="openshift-multus/multus-additional-cni-plugins-fp5j6" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.287165 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1e57bdf8-e537-402a-9815-609fecdd68ea-cnibin\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.287192 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5bf36887-2a2b-4193-abdb-a15beaf46300-mcd-auth-proxy-config\") pod \"machine-config-daemon-29qxq\" (UID: \"5bf36887-2a2b-4193-abdb-a15beaf46300\") " pod="openshift-machine-config-operator/machine-config-daemon-29qxq" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.287227 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/924c66b7-7c3c-4db7-a427-3f152efbac4e-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-fp5j6\" (UID: \"924c66b7-7c3c-4db7-a427-3f152efbac4e\") " pod="openshift-multus/multus-additional-cni-plugins-fp5j6" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.287255 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1e57bdf8-e537-402a-9815-609fecdd68ea-os-release\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.287279 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/924c66b7-7c3c-4db7-a427-3f152efbac4e-cni-binary-copy\") pod \"multus-additional-cni-plugins-fp5j6\" (UID: \"924c66b7-7c3c-4db7-a427-3f152efbac4e\") " pod="openshift-multus/multus-additional-cni-plugins-fp5j6" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.287304 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6g4fg\" (UniqueName: \"kubernetes.io/projected/924c66b7-7c3c-4db7-a427-3f152efbac4e-kube-api-access-6g4fg\") pod \"multus-additional-cni-plugins-fp5j6\" (UID: \"924c66b7-7c3c-4db7-a427-3f152efbac4e\") " pod="openshift-multus/multus-additional-cni-plugins-fp5j6" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.287324 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/1e57bdf8-e537-402a-9815-609fecdd68ea-host-var-lib-cni-multus\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.287351 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc-hosts-file\") pod \"node-resolver-msn6p\" (UID: \"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\") " pod="openshift-dns/node-resolver-msn6p" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.287374 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2tvg\" (UniqueName: \"kubernetes.io/projected/8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc-kube-api-access-q2tvg\") pod \"node-resolver-msn6p\" (UID: \"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\") " pod="openshift-dns/node-resolver-msn6p" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.287401 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1e57bdf8-e537-402a-9815-609fecdd68ea-host-run-netns\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.287427 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1e57bdf8-e537-402a-9815-609fecdd68ea-multus-cni-dir\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.287447 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1e57bdf8-e537-402a-9815-609fecdd68ea-etc-kubernetes\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.287492 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/924c66b7-7c3c-4db7-a427-3f152efbac4e-cnibin\") pod \"multus-additional-cni-plugins-fp5j6\" (UID: \"924c66b7-7c3c-4db7-a427-3f152efbac4e\") " pod="openshift-multus/multus-additional-cni-plugins-fp5j6" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.287511 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1e57bdf8-e537-402a-9815-609fecdd68ea-system-cni-dir\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.287534 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/1e57bdf8-e537-402a-9815-609fecdd68ea-multus-daemon-config\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.287556 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1e57bdf8-e537-402a-9815-609fecdd68ea-cni-binary-copy\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.287582 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/1e57bdf8-e537-402a-9815-609fecdd68ea-host-run-k8s-cni-cncf-io\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.287618 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l299f\" (UniqueName: \"kubernetes.io/projected/5bf36887-2a2b-4193-abdb-a15beaf46300-kube-api-access-l299f\") pod \"machine-config-daemon-29qxq\" (UID: \"5bf36887-2a2b-4193-abdb-a15beaf46300\") " pod="openshift-machine-config-operator/machine-config-daemon-29qxq" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.298343 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5c34a3d-82ad-4d03-abdd-0a20818ac2fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0261f41b82b492d324b5e50edd72cedd5e824de542a7b47104abd2ab757a7049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f43c7fe16acfd4d39301ee64665d4cd1d885397f981d70f73215da7f3aabe72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c871c1b454e4edd2f41076ecfd6a66592023d7161f094fbb4fd197fbf9c881f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.310595 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.321204 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.331585 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.341163 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.351125 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.361291 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.370393 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.378997 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bf36887-2a2b-4193-abdb-a15beaf46300\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-29qxq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.388638 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/924c66b7-7c3c-4db7-a427-3f152efbac4e-os-release\") pod \"multus-additional-cni-plugins-fp5j6\" (UID: \"924c66b7-7c3c-4db7-a427-3f152efbac4e\") " pod="openshift-multus/multus-additional-cni-plugins-fp5j6" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.388686 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/1e57bdf8-e537-402a-9815-609fecdd68ea-multus-socket-dir-parent\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.388713 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1e57bdf8-e537-402a-9815-609fecdd68ea-host-var-lib-cni-bin\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.388750 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1e57bdf8-e537-402a-9815-609fecdd68ea-multus-conf-dir\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.388806 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/924c66b7-7c3c-4db7-a427-3f152efbac4e-tuning-conf-dir\") pod \"multus-additional-cni-plugins-fp5j6\" (UID: \"924c66b7-7c3c-4db7-a427-3f152efbac4e\") " pod="openshift-multus/multus-additional-cni-plugins-fp5j6" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.388842 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1e57bdf8-e537-402a-9815-609fecdd68ea-cnibin\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.388875 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1e57bdf8-e537-402a-9815-609fecdd68ea-host-var-lib-cni-bin\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.388880 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5bf36887-2a2b-4193-abdb-a15beaf46300-mcd-auth-proxy-config\") pod \"machine-config-daemon-29qxq\" (UID: \"5bf36887-2a2b-4193-abdb-a15beaf46300\") " pod="openshift-machine-config-operator/machine-config-daemon-29qxq" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.388946 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/924c66b7-7c3c-4db7-a427-3f152efbac4e-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-fp5j6\" (UID: \"924c66b7-7c3c-4db7-a427-3f152efbac4e\") " pod="openshift-multus/multus-additional-cni-plugins-fp5j6" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.388968 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1e57bdf8-e537-402a-9815-609fecdd68ea-os-release\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.388961 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1e57bdf8-e537-402a-9815-609fecdd68ea-multus-conf-dir\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.388993 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/924c66b7-7c3c-4db7-a427-3f152efbac4e-cni-binary-copy\") pod \"multus-additional-cni-plugins-fp5j6\" (UID: \"924c66b7-7c3c-4db7-a427-3f152efbac4e\") " pod="openshift-multus/multus-additional-cni-plugins-fp5j6" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.389095 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6g4fg\" (UniqueName: \"kubernetes.io/projected/924c66b7-7c3c-4db7-a427-3f152efbac4e-kube-api-access-6g4fg\") pod \"multus-additional-cni-plugins-fp5j6\" (UID: \"924c66b7-7c3c-4db7-a427-3f152efbac4e\") " pod="openshift-multus/multus-additional-cni-plugins-fp5j6" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.389101 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/1e57bdf8-e537-402a-9815-609fecdd68ea-multus-socket-dir-parent\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.389149 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/1e57bdf8-e537-402a-9815-609fecdd68ea-host-var-lib-cni-multus\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.389173 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/924c66b7-7c3c-4db7-a427-3f152efbac4e-os-release\") pod \"multus-additional-cni-plugins-fp5j6\" (UID: \"924c66b7-7c3c-4db7-a427-3f152efbac4e\") " pod="openshift-multus/multus-additional-cni-plugins-fp5j6" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.389124 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/1e57bdf8-e537-402a-9815-609fecdd68ea-host-var-lib-cni-multus\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.389349 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1e57bdf8-e537-402a-9815-609fecdd68ea-os-release\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.389517 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc-hosts-file\") pod \"node-resolver-msn6p\" (UID: \"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\") " pod="openshift-dns/node-resolver-msn6p" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.389579 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2tvg\" (UniqueName: \"kubernetes.io/projected/8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc-kube-api-access-q2tvg\") pod \"node-resolver-msn6p\" (UID: \"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\") " pod="openshift-dns/node-resolver-msn6p" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.389606 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1e57bdf8-e537-402a-9815-609fecdd68ea-host-run-netns\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.389630 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1e57bdf8-e537-402a-9815-609fecdd68ea-multus-cni-dir\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.389650 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1e57bdf8-e537-402a-9815-609fecdd68ea-etc-kubernetes\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.389693 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/924c66b7-7c3c-4db7-a427-3f152efbac4e-cnibin\") pod \"multus-additional-cni-plugins-fp5j6\" (UID: \"924c66b7-7c3c-4db7-a427-3f152efbac4e\") " pod="openshift-multus/multus-additional-cni-plugins-fp5j6" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.389715 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1e57bdf8-e537-402a-9815-609fecdd68ea-system-cni-dir\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.389719 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1e57bdf8-e537-402a-9815-609fecdd68ea-host-run-netns\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.389735 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/1e57bdf8-e537-402a-9815-609fecdd68ea-multus-daemon-config\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.389760 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1e57bdf8-e537-402a-9815-609fecdd68ea-etc-kubernetes\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.389762 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1e57bdf8-e537-402a-9815-609fecdd68ea-cni-binary-copy\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.389814 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/1e57bdf8-e537-402a-9815-609fecdd68ea-host-run-k8s-cni-cncf-io\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.389864 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l299f\" (UniqueName: \"kubernetes.io/projected/5bf36887-2a2b-4193-abdb-a15beaf46300-kube-api-access-l299f\") pod \"machine-config-daemon-29qxq\" (UID: \"5bf36887-2a2b-4193-abdb-a15beaf46300\") " pod="openshift-machine-config-operator/machine-config-daemon-29qxq" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.389890 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/5bf36887-2a2b-4193-abdb-a15beaf46300-rootfs\") pod \"machine-config-daemon-29qxq\" (UID: \"5bf36887-2a2b-4193-abdb-a15beaf46300\") " pod="openshift-machine-config-operator/machine-config-daemon-29qxq" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.389912 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/1e57bdf8-e537-402a-9815-609fecdd68ea-host-var-lib-kubelet\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.389930 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5bf36887-2a2b-4193-abdb-a15beaf46300-proxy-tls\") pod \"machine-config-daemon-29qxq\" (UID: \"5bf36887-2a2b-4193-abdb-a15beaf46300\") " pod="openshift-machine-config-operator/machine-config-daemon-29qxq" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.389948 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/1e57bdf8-e537-402a-9815-609fecdd68ea-hostroot\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.389987 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/924c66b7-7c3c-4db7-a427-3f152efbac4e-system-cni-dir\") pod \"multus-additional-cni-plugins-fp5j6\" (UID: \"924c66b7-7c3c-4db7-a427-3f152efbac4e\") " pod="openshift-multus/multus-additional-cni-plugins-fp5j6" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.390008 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q22d4\" (UniqueName: \"kubernetes.io/projected/1e57bdf8-e537-402a-9815-609fecdd68ea-kube-api-access-q22d4\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.390033 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/1e57bdf8-e537-402a-9815-609fecdd68ea-host-run-multus-certs\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.390029 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1e57bdf8-e537-402a-9815-609fecdd68ea-system-cni-dir\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.390072 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/924c66b7-7c3c-4db7-a427-3f152efbac4e-cni-binary-copy\") pod \"multus-additional-cni-plugins-fp5j6\" (UID: \"924c66b7-7c3c-4db7-a427-3f152efbac4e\") " pod="openshift-multus/multus-additional-cni-plugins-fp5j6" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.390100 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/924c66b7-7c3c-4db7-a427-3f152efbac4e-cnibin\") pod \"multus-additional-cni-plugins-fp5j6\" (UID: \"924c66b7-7c3c-4db7-a427-3f152efbac4e\") " pod="openshift-multus/multus-additional-cni-plugins-fp5j6" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.390112 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1e57bdf8-e537-402a-9815-609fecdd68ea-multus-cni-dir\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.390186 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/1e57bdf8-e537-402a-9815-609fecdd68ea-host-run-multus-certs\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.390209 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/1e57bdf8-e537-402a-9815-609fecdd68ea-host-run-k8s-cni-cncf-io\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.390273 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/5bf36887-2a2b-4193-abdb-a15beaf46300-rootfs\") pod \"machine-config-daemon-29qxq\" (UID: \"5bf36887-2a2b-4193-abdb-a15beaf46300\") " pod="openshift-machine-config-operator/machine-config-daemon-29qxq" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.390288 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/1e57bdf8-e537-402a-9815-609fecdd68ea-host-var-lib-kubelet\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.390310 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/1e57bdf8-e537-402a-9815-609fecdd68ea-hostroot\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.390324 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/924c66b7-7c3c-4db7-a427-3f152efbac4e-system-cni-dir\") pod \"multus-additional-cni-plugins-fp5j6\" (UID: \"924c66b7-7c3c-4db7-a427-3f152efbac4e\") " pod="openshift-multus/multus-additional-cni-plugins-fp5j6" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.390351 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5bf36887-2a2b-4193-abdb-a15beaf46300-mcd-auth-proxy-config\") pod \"machine-config-daemon-29qxq\" (UID: \"5bf36887-2a2b-4193-abdb-a15beaf46300\") " pod="openshift-machine-config-operator/machine-config-daemon-29qxq" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.390528 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc-hosts-file\") pod \"node-resolver-msn6p\" (UID: \"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\") " pod="openshift-dns/node-resolver-msn6p" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.390646 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/924c66b7-7c3c-4db7-a427-3f152efbac4e-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-fp5j6\" (UID: \"924c66b7-7c3c-4db7-a427-3f152efbac4e\") " pod="openshift-multus/multus-additional-cni-plugins-fp5j6" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.390673 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1e57bdf8-e537-402a-9815-609fecdd68ea-cni-binary-copy\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.390721 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1e57bdf8-e537-402a-9815-609fecdd68ea-cnibin\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.390968 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/924c66b7-7c3c-4db7-a427-3f152efbac4e-tuning-conf-dir\") pod \"multus-additional-cni-plugins-fp5j6\" (UID: \"924c66b7-7c3c-4db7-a427-3f152efbac4e\") " pod="openshift-multus/multus-additional-cni-plugins-fp5j6" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.390993 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/1e57bdf8-e537-402a-9815-609fecdd68ea-multus-daemon-config\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.392008 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9hx9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e57bdf8-e537-402a-9815-609fecdd68ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q22d4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9hx9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.396559 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5bf36887-2a2b-4193-abdb-a15beaf46300-proxy-tls\") pod \"machine-config-daemon-29qxq\" (UID: \"5bf36887-2a2b-4193-abdb-a15beaf46300\") " pod="openshift-machine-config-operator/machine-config-daemon-29qxq" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.399471 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.402886 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac"} Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.403249 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.403182 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-msn6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2tvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-msn6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.407367 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"3d0966253f623158053125dffca956000824438bdc548eba6dafe9a162d5674a"} Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.407483 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"9d07bfb361f14d817913fa2cabd30649447790068545dda2f3a2f71213ea0e39"} Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.407606 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2tvg\" (UniqueName: \"kubernetes.io/projected/8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc-kube-api-access-q2tvg\") pod \"node-resolver-msn6p\" (UID: \"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\") " pod="openshift-dns/node-resolver-msn6p" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.408894 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"ae3e65d5825e413b4fe1b062bce962e7e861d6aa74283ed1bb82f974fd6bc828"} Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.410012 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q22d4\" (UniqueName: \"kubernetes.io/projected/1e57bdf8-e537-402a-9815-609fecdd68ea-kube-api-access-q22d4\") pod \"multus-9hx9f\" (UID: \"1e57bdf8-e537-402a-9815-609fecdd68ea\") " pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.410106 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"b9ecd4380a6ae8ff62e27e3946815170c2f02b4c3921bda595467345289c9e0a"} Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.410141 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"816826c3ed50a65e07e040f77fd22f6d2315ee5f7cb34b2affb91b86bbe3cc9e"} Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.412041 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l299f\" (UniqueName: \"kubernetes.io/projected/5bf36887-2a2b-4193-abdb-a15beaf46300-kube-api-access-l299f\") pod \"machine-config-daemon-29qxq\" (UID: \"5bf36887-2a2b-4193-abdb-a15beaf46300\") " pod="openshift-machine-config-operator/machine-config-daemon-29qxq" Feb 17 13:04:44 crc kubenswrapper[4955]: E0217 13:04:44.416018 4955 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.418579 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05aa39d-94e6-49d4-885e-2c1f986310f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T13:04:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 13:04:35.885968 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 13:04:35.888176 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3462920193/tls.crt::/tmp/serving-cert-3462920193/tls.key\\\\\\\"\\\\nI0217 13:04:42.370546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 13:04:42.384242 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 13:04:42.384284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 13:04:42.384317 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 13:04:42.384327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 13:04:42.394555 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 13:04:42.394614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394628 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 13:04:42.394646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 13:04:42.394654 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 13:04:42.394662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 13:04:42.394900 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 13:04:42.396494 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.423060 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6g4fg\" (UniqueName: \"kubernetes.io/projected/924c66b7-7c3c-4db7-a427-3f152efbac4e-kube-api-access-6g4fg\") pod \"multus-additional-cni-plugins-fp5j6\" (UID: \"924c66b7-7c3c-4db7-a427-3f152efbac4e\") " pod="openshift-multus/multus-additional-cni-plugins-fp5j6" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.438358 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5c34a3d-82ad-4d03-abdd-0a20818ac2fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0261f41b82b492d324b5e50edd72cedd5e824de542a7b47104abd2ab757a7049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f43c7fe16acfd4d39301ee64665d4cd1d885397f981d70f73215da7f3aabe72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c871c1b454e4edd2f41076ecfd6a66592023d7161f094fbb4fd197fbf9c881f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.450167 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.461416 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"924c66b7-7c3c-4db7-a427-3f152efbac4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fp5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.471195 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.480707 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5c34a3d-82ad-4d03-abdd-0a20818ac2fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0261f41b82b492d324b5e50edd72cedd5e824de542a7b47104abd2ab757a7049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f43c7fe16acfd4d39301ee64665d4cd1d885397f981d70f73215da7f3aabe72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c871c1b454e4edd2f41076ecfd6a66592023d7161f094fbb4fd197fbf9c881f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.488844 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.502338 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.502397 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"924c66b7-7c3c-4db7-a427-3f152efbac4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fp5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.508914 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.514521 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d0966253f623158053125dffca956000824438bdc548eba6dafe9a162d5674a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d07bfb361f14d817913fa2cabd30649447790068545dda2f3a2f71213ea0e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.519309 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-9hx9f" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.525502 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9ecd4380a6ae8ff62e27e3946815170c2f02b4c3921bda595467345289c9e0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.526734 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-msn6p" Feb 17 13:04:44 crc kubenswrapper[4955]: W0217 13:04:44.528539 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5bf36887_2a2b_4193_abdb_a15beaf46300.slice/crio-96d63e3e3231c92e1a8e43745a36c32ff2504f17c34230695fd13fb158ac6102 WatchSource:0}: Error finding container 96d63e3e3231c92e1a8e43745a36c32ff2504f17c34230695fd13fb158ac6102: Status 404 returned error can't find the container with id 96d63e3e3231c92e1a8e43745a36c32ff2504f17c34230695fd13fb158ac6102 Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.538094 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.550144 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05aa39d-94e6-49d4-885e-2c1f986310f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T13:04:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 13:04:35.885968 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 13:04:35.888176 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3462920193/tls.crt::/tmp/serving-cert-3462920193/tls.key\\\\\\\"\\\\nI0217 13:04:42.370546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 13:04:42.384242 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 13:04:42.384284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 13:04:42.384317 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 13:04:42.384327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 13:04:42.394555 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 13:04:42.394614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394628 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 13:04:42.394646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 13:04:42.394654 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 13:04:42.394662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 13:04:42.394900 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 13:04:42.396494 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.561025 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-dcwgg"] Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.562071 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.562472 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.568168 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.568278 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.568296 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.578374 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.578768 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.579001 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.584414 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.591739 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-run-systemd\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.591775 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-run-ovn\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.591827 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.591849 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/dcbc7a69-58e7-4b7a-bb84-96276819ce88-ovn-node-metrics-cert\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.591874 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-etc-openvswitch\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.591893 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-run-openvswitch\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.591919 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-host-slash\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.591938 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-log-socket\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.592004 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-host-run-netns\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.592033 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-host-cni-netd\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.592053 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/dcbc7a69-58e7-4b7a-bb84-96276819ce88-env-overrides\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.592040 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.592098 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-systemd-units\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.592167 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-node-log\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.592295 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/dcbc7a69-58e7-4b7a-bb84-96276819ce88-ovnkube-config\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.592377 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-host-kubelet\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.592395 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-host-cni-bin\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.592559 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-var-lib-openvswitch\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.592687 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-host-run-ovn-kubernetes\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.592764 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2m6wz\" (UniqueName: \"kubernetes.io/projected/dcbc7a69-58e7-4b7a-bb84-96276819ce88-kube-api-access-2m6wz\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.592871 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/dcbc7a69-58e7-4b7a-bb84-96276819ce88-ovnkube-script-lib\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.605472 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bf36887-2a2b-4193-abdb-a15beaf46300\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-29qxq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: W0217 13:04:44.625116 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8df54e9e_2c74_4b95_bbc8_0fdad6efc5fc.slice/crio-cd88fadc21c16c4bf45e91fb92ab5757e772be72833c1554783041a93c74e699 WatchSource:0}: Error finding container cd88fadc21c16c4bf45e91fb92ab5757e772be72833c1554783041a93c74e699: Status 404 returned error can't find the container with id cd88fadc21c16c4bf45e91fb92ab5757e772be72833c1554783041a93c74e699 Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.632433 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9hx9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e57bdf8-e537-402a-9815-609fecdd68ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q22d4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9hx9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.677316 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-msn6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2tvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-msn6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.695218 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-run-systemd\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.695266 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-run-ovn\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.695318 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.695342 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/dcbc7a69-58e7-4b7a-bb84-96276819ce88-ovn-node-metrics-cert\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.695361 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-etc-openvswitch\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.695369 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-run-systemd\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.695412 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-run-openvswitch\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.695380 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-run-openvswitch\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.695462 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-etc-openvswitch\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.695475 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-host-slash\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.695489 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.695414 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-run-ovn\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.695504 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-log-socket\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.695532 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-host-slash\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.695564 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-host-run-netns\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.695590 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-host-cni-netd\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.695608 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/dcbc7a69-58e7-4b7a-bb84-96276819ce88-env-overrides\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.695639 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-systemd-units\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.695656 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-node-log\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.695682 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/dcbc7a69-58e7-4b7a-bb84-96276819ce88-ovnkube-config\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.695707 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-host-kubelet\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.695726 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-host-cni-bin\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.695758 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-var-lib-openvswitch\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.695795 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-host-run-ovn-kubernetes\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.695817 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2m6wz\" (UniqueName: \"kubernetes.io/projected/dcbc7a69-58e7-4b7a-bb84-96276819ce88-kube-api-access-2m6wz\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.695838 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/dcbc7a69-58e7-4b7a-bb84-96276819ce88-ovnkube-script-lib\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.696175 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-node-log\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.696215 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-log-socket\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.696241 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-host-run-netns\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.696267 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-host-cni-netd\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.696828 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-host-run-ovn-kubernetes\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.696871 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-var-lib-openvswitch\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.696996 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/dcbc7a69-58e7-4b7a-bb84-96276819ce88-env-overrides\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.697042 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-systemd-units\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.697075 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-host-kubelet\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.697215 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-host-cni-bin\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.697392 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/dcbc7a69-58e7-4b7a-bb84-96276819ce88-ovnkube-config\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.697423 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/dcbc7a69-58e7-4b7a-bb84-96276819ce88-ovnkube-script-lib\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.704444 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/dcbc7a69-58e7-4b7a-bb84-96276819ce88-ovn-node-metrics-cert\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.714090 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05aa39d-94e6-49d4-885e-2c1f986310f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T13:04:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 13:04:35.885968 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 13:04:35.888176 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3462920193/tls.crt::/tmp/serving-cert-3462920193/tls.key\\\\\\\"\\\\nI0217 13:04:42.370546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 13:04:42.384242 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 13:04:42.384284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 13:04:42.384317 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 13:04:42.384327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 13:04:42.394555 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 13:04:42.394614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394628 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 13:04:42.394646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 13:04:42.394654 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 13:04:42.394662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 13:04:42.394900 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 13:04:42.396494 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.738694 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2m6wz\" (UniqueName: \"kubernetes.io/projected/dcbc7a69-58e7-4b7a-bb84-96276819ce88-kube-api-access-2m6wz\") pod \"ovnkube-node-dcwgg\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.769852 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.802426 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.802572 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.802622 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:04:44 crc kubenswrapper[4955]: E0217 13:04:44.802678 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:04:46.802639981 +0000 UTC m=+25.325369704 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.802727 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.802817 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:04:44 crc kubenswrapper[4955]: E0217 13:04:44.802829 4955 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 17 13:04:44 crc kubenswrapper[4955]: E0217 13:04:44.802856 4955 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 17 13:04:44 crc kubenswrapper[4955]: E0217 13:04:44.802871 4955 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 13:04:44 crc kubenswrapper[4955]: E0217 13:04:44.802945 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-17 13:04:46.802921809 +0000 UTC m=+25.325651552 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 13:04:44 crc kubenswrapper[4955]: E0217 13:04:44.803027 4955 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 17 13:04:44 crc kubenswrapper[4955]: E0217 13:04:44.803061 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-17 13:04:46.803051463 +0000 UTC m=+25.325781196 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 17 13:04:44 crc kubenswrapper[4955]: E0217 13:04:44.803029 4955 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 17 13:04:44 crc kubenswrapper[4955]: E0217 13:04:44.803073 4955 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 17 13:04:44 crc kubenswrapper[4955]: E0217 13:04:44.803091 4955 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 17 13:04:44 crc kubenswrapper[4955]: E0217 13:04:44.803106 4955 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 13:04:44 crc kubenswrapper[4955]: E0217 13:04:44.803119 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-17 13:04:46.803110004 +0000 UTC m=+25.325839727 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 17 13:04:44 crc kubenswrapper[4955]: E0217 13:04:44.803145 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-17 13:04:46.803132705 +0000 UTC m=+25.325862248 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.816320 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.850688 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bf36887-2a2b-4193-abdb-a15beaf46300\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-29qxq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.894619 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9hx9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e57bdf8-e537-402a-9815-609fecdd68ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q22d4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9hx9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:44Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.906076 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:44 crc kubenswrapper[4955]: W0217 13:04:44.922865 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddcbc7a69_58e7_4b7a_bb84_96276819ce88.slice/crio-4d43867b072716a140df4d0f846947f25ee3e32f7c51e8569e825b40bbc38dda WatchSource:0}: Error finding container 4d43867b072716a140df4d0f846947f25ee3e32f7c51e8569e825b40bbc38dda: Status 404 returned error can't find the container with id 4d43867b072716a140df4d0f846947f25ee3e32f7c51e8569e825b40bbc38dda Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.932768 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-msn6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2tvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-msn6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:44Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:44 crc kubenswrapper[4955]: I0217 13:04:44.974591 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5c34a3d-82ad-4d03-abdd-0a20818ac2fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0261f41b82b492d324b5e50edd72cedd5e824de542a7b47104abd2ab757a7049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f43c7fe16acfd4d39301ee64665d4cd1d885397f981d70f73215da7f3aabe72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c871c1b454e4edd2f41076ecfd6a66592023d7161f094fbb4fd197fbf9c881f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:44Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.016128 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.060250 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"924c66b7-7c3c-4db7-a427-3f152efbac4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fp5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.103699 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcbc7a69-58e7-4b7a-bb84-96276819ce88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dcwgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.137155 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d0966253f623158053125dffca956000824438bdc548eba6dafe9a162d5674a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d07bfb361f14d817913fa2cabd30649447790068545dda2f3a2f71213ea0e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.173908 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 04:13:58.319901391 +0000 UTC Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.175882 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9ecd4380a6ae8ff62e27e3946815170c2f02b4c3921bda595467345289c9e0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.217660 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.223092 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.223115 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:04:45 crc kubenswrapper[4955]: E0217 13:04:45.223305 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.223115 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:04:45 crc kubenswrapper[4955]: E0217 13:04:45.223589 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:04:45 crc kubenswrapper[4955]: E0217 13:04:45.223474 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.416450 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-9hx9f" event={"ID":"1e57bdf8-e537-402a-9815-609fecdd68ea","Type":"ContainerStarted","Data":"dbc9014d80aee39922161014aac8fafa7d93be3e4092fe705c76da4a487a5e5d"} Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.416906 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-9hx9f" event={"ID":"1e57bdf8-e537-402a-9815-609fecdd68ea","Type":"ContainerStarted","Data":"6eb1a79f1e2c9176aa54fb9dafc6df7408b61d8a585eb40fe6f4c574d1899425"} Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.419303 4955 generic.go:334] "Generic (PLEG): container finished" podID="924c66b7-7c3c-4db7-a427-3f152efbac4e" containerID="e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9" exitCode=0 Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.419431 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" event={"ID":"924c66b7-7c3c-4db7-a427-3f152efbac4e","Type":"ContainerDied","Data":"e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9"} Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.419468 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" event={"ID":"924c66b7-7c3c-4db7-a427-3f152efbac4e","Type":"ContainerStarted","Data":"e0ec60c22af8e0ca57ee6ecf1f9155c792913fa6e364a8835042250d827dcc6d"} Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.422068 4955 generic.go:334] "Generic (PLEG): container finished" podID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerID="e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0" exitCode=0 Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.422248 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" event={"ID":"dcbc7a69-58e7-4b7a-bb84-96276819ce88","Type":"ContainerDied","Data":"e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0"} Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.422327 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" event={"ID":"dcbc7a69-58e7-4b7a-bb84-96276819ce88","Type":"ContainerStarted","Data":"4d43867b072716a140df4d0f846947f25ee3e32f7c51e8569e825b40bbc38dda"} Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.426738 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-msn6p" event={"ID":"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc","Type":"ContainerStarted","Data":"277a12d62c90ef420c31d8dc71f0d7750dfd5f5ab7e90fec0dcd00e54b60245d"} Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.426861 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-msn6p" event={"ID":"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc","Type":"ContainerStarted","Data":"cd88fadc21c16c4bf45e91fb92ab5757e772be72833c1554783041a93c74e699"} Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.433283 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" event={"ID":"5bf36887-2a2b-4193-abdb-a15beaf46300","Type":"ContainerStarted","Data":"11a8924eb6ce41d629feb2b42c7d4778a3ca9110a59c3ac25acbba4367272374"} Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.433354 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" event={"ID":"5bf36887-2a2b-4193-abdb-a15beaf46300","Type":"ContainerStarted","Data":"9bec78f9e9f6416d6b600f6231f0d24652473b48b95bf0e0675cc85c9208d124"} Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.433376 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" event={"ID":"5bf36887-2a2b-4193-abdb-a15beaf46300","Type":"ContainerStarted","Data":"96d63e3e3231c92e1a8e43745a36c32ff2504f17c34230695fd13fb158ac6102"} Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.444482 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9ecd4380a6ae8ff62e27e3946815170c2f02b4c3921bda595467345289c9e0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.462464 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.482337 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.500521 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.517714 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bf36887-2a2b-4193-abdb-a15beaf46300\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-29qxq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.536168 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9hx9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e57bdf8-e537-402a-9815-609fecdd68ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbc9014d80aee39922161014aac8fafa7d93be3e4092fe705c76da4a487a5e5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q22d4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9hx9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.548811 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-msn6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2tvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-msn6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.567710 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05aa39d-94e6-49d4-885e-2c1f986310f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T13:04:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 13:04:35.885968 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 13:04:35.888176 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3462920193/tls.crt::/tmp/serving-cert-3462920193/tls.key\\\\\\\"\\\\nI0217 13:04:42.370546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 13:04:42.384242 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 13:04:42.384284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 13:04:42.384317 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 13:04:42.384327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 13:04:42.394555 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 13:04:42.394614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394628 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 13:04:42.394646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 13:04:42.394654 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 13:04:42.394662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 13:04:42.394900 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 13:04:42.396494 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.585396 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5c34a3d-82ad-4d03-abdd-0a20818ac2fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0261f41b82b492d324b5e50edd72cedd5e824de542a7b47104abd2ab757a7049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f43c7fe16acfd4d39301ee64665d4cd1d885397f981d70f73215da7f3aabe72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c871c1b454e4edd2f41076ecfd6a66592023d7161f094fbb4fd197fbf9c881f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.660383 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.675747 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.682859 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"924c66b7-7c3c-4db7-a427-3f152efbac4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fp5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.691686 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.715477 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.748293 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcbc7a69-58e7-4b7a-bb84-96276819ce88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dcwgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.788009 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d0966253f623158053125dffca956000824438bdc548eba6dafe9a162d5674a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d07bfb361f14d817913fa2cabd30649447790068545dda2f3a2f71213ea0e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.815835 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9ecd4380a6ae8ff62e27e3946815170c2f02b4c3921bda595467345289c9e0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.841532 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.876411 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.918054 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.960228 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bf36887-2a2b-4193-abdb-a15beaf46300\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11a8924eb6ce41d629feb2b42c7d4778a3ca9110a59c3ac25acbba4367272374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bec78f9e9f6416d6b600f6231f0d24652473b48b95bf0e0675cc85c9208d124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-29qxq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:45 crc kubenswrapper[4955]: I0217 13:04:45.997293 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9hx9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e57bdf8-e537-402a-9815-609fecdd68ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbc9014d80aee39922161014aac8fafa7d93be3e4092fe705c76da4a487a5e5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q22d4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9hx9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:46 crc kubenswrapper[4955]: I0217 13:04:46.034908 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-msn6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://277a12d62c90ef420c31d8dc71f0d7750dfd5f5ab7e90fec0dcd00e54b60245d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2tvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-msn6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:46Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:46 crc kubenswrapper[4955]: I0217 13:04:46.074992 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05aa39d-94e6-49d4-885e-2c1f986310f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T13:04:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 13:04:35.885968 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 13:04:35.888176 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3462920193/tls.crt::/tmp/serving-cert-3462920193/tls.key\\\\\\\"\\\\nI0217 13:04:42.370546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 13:04:42.384242 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 13:04:42.384284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 13:04:42.384317 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 13:04:42.384327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 13:04:42.394555 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 13:04:42.394614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394628 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 13:04:42.394646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 13:04:42.394654 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 13:04:42.394662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 13:04:42.394900 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 13:04:42.396494 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:46Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:46 crc kubenswrapper[4955]: I0217 13:04:46.120412 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5c34a3d-82ad-4d03-abdd-0a20818ac2fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0261f41b82b492d324b5e50edd72cedd5e824de542a7b47104abd2ab757a7049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f43c7fe16acfd4d39301ee64665d4cd1d885397f981d70f73215da7f3aabe72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c871c1b454e4edd2f41076ecfd6a66592023d7161f094fbb4fd197fbf9c881f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:46Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:46 crc kubenswrapper[4955]: I0217 13:04:46.170732 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:46Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:46 crc kubenswrapper[4955]: I0217 13:04:46.174849 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 08:50:22.426955426 +0000 UTC Feb 17 13:04:46 crc kubenswrapper[4955]: I0217 13:04:46.199066 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"924c66b7-7c3c-4db7-a427-3f152efbac4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fp5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:46Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:46 crc kubenswrapper[4955]: I0217 13:04:46.238050 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcbc7a69-58e7-4b7a-bb84-96276819ce88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dcwgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:46Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:46 crc kubenswrapper[4955]: I0217 13:04:46.288293 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b49c9af-148a-48ec-af37-f555f9186337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d6919c91918e8bacafd1442b06d14ffd9b547f0f37b2298e2d8d3acc5962d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e38466a067b20ca4323ab6983cb324bb63179f942d2c37c0daa8b03767e7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4830450cbada3d24c9747e9de5f48e53e118549f1dd67edf7fff8b2adc47128f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001dfb595f39d1da33b265474b6cca51c62f522ff061c9e7c52f93806034f4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f1fee7ac5f8d1442d9f63b0e8d285187129eb3a3e3b50a49a45c7a7a39ae8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:46Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:46 crc kubenswrapper[4955]: I0217 13:04:46.314952 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d0966253f623158053125dffca956000824438bdc548eba6dafe9a162d5674a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d07bfb361f14d817913fa2cabd30649447790068545dda2f3a2f71213ea0e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:46Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:46 crc kubenswrapper[4955]: I0217 13:04:46.437874 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"c0f4056e4b5a554f05004c194485d25231b67024e4376a73c715fcb32004a3bf"} Feb 17 13:04:46 crc kubenswrapper[4955]: I0217 13:04:46.442960 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" event={"ID":"dcbc7a69-58e7-4b7a-bb84-96276819ce88","Type":"ContainerStarted","Data":"23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44"} Feb 17 13:04:46 crc kubenswrapper[4955]: I0217 13:04:46.443010 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" event={"ID":"dcbc7a69-58e7-4b7a-bb84-96276819ce88","Type":"ContainerStarted","Data":"11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4"} Feb 17 13:04:46 crc kubenswrapper[4955]: I0217 13:04:46.443023 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" event={"ID":"dcbc7a69-58e7-4b7a-bb84-96276819ce88","Type":"ContainerStarted","Data":"ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff"} Feb 17 13:04:46 crc kubenswrapper[4955]: I0217 13:04:46.443036 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" event={"ID":"dcbc7a69-58e7-4b7a-bb84-96276819ce88","Type":"ContainerStarted","Data":"d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748"} Feb 17 13:04:46 crc kubenswrapper[4955]: I0217 13:04:46.443046 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" event={"ID":"dcbc7a69-58e7-4b7a-bb84-96276819ce88","Type":"ContainerStarted","Data":"b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636"} Feb 17 13:04:46 crc kubenswrapper[4955]: I0217 13:04:46.443056 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" event={"ID":"dcbc7a69-58e7-4b7a-bb84-96276819ce88","Type":"ContainerStarted","Data":"89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18"} Feb 17 13:04:46 crc kubenswrapper[4955]: I0217 13:04:46.445570 4955 generic.go:334] "Generic (PLEG): container finished" podID="924c66b7-7c3c-4db7-a427-3f152efbac4e" containerID="32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e" exitCode=0 Feb 17 13:04:46 crc kubenswrapper[4955]: I0217 13:04:46.445660 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" event={"ID":"924c66b7-7c3c-4db7-a427-3f152efbac4e","Type":"ContainerDied","Data":"32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e"} Feb 17 13:04:46 crc kubenswrapper[4955]: E0217 13:04:46.462927 4955 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"etcd-crc\" already exists" pod="openshift-etcd/etcd-crc" Feb 17 13:04:46 crc kubenswrapper[4955]: I0217 13:04:46.462923 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9ecd4380a6ae8ff62e27e3946815170c2f02b4c3921bda595467345289c9e0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:46Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:46 crc kubenswrapper[4955]: I0217 13:04:46.476923 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0f4056e4b5a554f05004c194485d25231b67024e4376a73c715fcb32004a3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:46Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:46 crc kubenswrapper[4955]: I0217 13:04:46.498402 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05aa39d-94e6-49d4-885e-2c1f986310f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T13:04:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 13:04:35.885968 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 13:04:35.888176 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3462920193/tls.crt::/tmp/serving-cert-3462920193/tls.key\\\\\\\"\\\\nI0217 13:04:42.370546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 13:04:42.384242 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 13:04:42.384284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 13:04:42.384317 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 13:04:42.384327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 13:04:42.394555 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 13:04:42.394614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394628 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 13:04:42.394646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 13:04:42.394654 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 13:04:42.394662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 13:04:42.394900 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 13:04:42.396494 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:46Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:46 crc kubenswrapper[4955]: I0217 13:04:46.518361 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:46Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:46 crc kubenswrapper[4955]: I0217 13:04:46.532812 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:46Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:46 crc kubenswrapper[4955]: I0217 13:04:46.573394 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bf36887-2a2b-4193-abdb-a15beaf46300\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11a8924eb6ce41d629feb2b42c7d4778a3ca9110a59c3ac25acbba4367272374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bec78f9e9f6416d6b600f6231f0d24652473b48b95bf0e0675cc85c9208d124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-29qxq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:46Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:46 crc kubenswrapper[4955]: I0217 13:04:46.616102 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9hx9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e57bdf8-e537-402a-9815-609fecdd68ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbc9014d80aee39922161014aac8fafa7d93be3e4092fe705c76da4a487a5e5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q22d4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9hx9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:46Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:46 crc kubenswrapper[4955]: I0217 13:04:46.653177 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-msn6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://277a12d62c90ef420c31d8dc71f0d7750dfd5f5ab7e90fec0dcd00e54b60245d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2tvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-msn6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:46Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:46 crc kubenswrapper[4955]: I0217 13:04:46.705033 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b49c9af-148a-48ec-af37-f555f9186337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d6919c91918e8bacafd1442b06d14ffd9b547f0f37b2298e2d8d3acc5962d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e38466a067b20ca4323ab6983cb324bb63179f942d2c37c0daa8b03767e7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4830450cbada3d24c9747e9de5f48e53e118549f1dd67edf7fff8b2adc47128f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001dfb595f39d1da33b265474b6cca51c62f522ff061c9e7c52f93806034f4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f1fee7ac5f8d1442d9f63b0e8d285187129eb3a3e3b50a49a45c7a7a39ae8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:46Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:46 crc kubenswrapper[4955]: I0217 13:04:46.735687 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5c34a3d-82ad-4d03-abdd-0a20818ac2fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0261f41b82b492d324b5e50edd72cedd5e824de542a7b47104abd2ab757a7049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f43c7fe16acfd4d39301ee64665d4cd1d885397f981d70f73215da7f3aabe72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c871c1b454e4edd2f41076ecfd6a66592023d7161f094fbb4fd197fbf9c881f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:46Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:46 crc kubenswrapper[4955]: I0217 13:04:46.777677 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:46Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:46 crc kubenswrapper[4955]: I0217 13:04:46.830291 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"924c66b7-7c3c-4db7-a427-3f152efbac4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fp5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:46Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:46 crc kubenswrapper[4955]: I0217 13:04:46.845394 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:04:46 crc kubenswrapper[4955]: E0217 13:04:46.845757 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:04:50.845737507 +0000 UTC m=+29.368467050 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:04:46 crc kubenswrapper[4955]: I0217 13:04:46.845988 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:04:46 crc kubenswrapper[4955]: I0217 13:04:46.846181 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:04:46 crc kubenswrapper[4955]: I0217 13:04:46.846263 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:04:46 crc kubenswrapper[4955]: I0217 13:04:46.846344 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:04:46 crc kubenswrapper[4955]: E0217 13:04:46.846152 4955 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 17 13:04:46 crc kubenswrapper[4955]: E0217 13:04:46.846581 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-17 13:04:50.84657182 +0000 UTC m=+29.369301363 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 17 13:04:46 crc kubenswrapper[4955]: E0217 13:04:46.846490 4955 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 17 13:04:46 crc kubenswrapper[4955]: E0217 13:04:46.846975 4955 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 17 13:04:46 crc kubenswrapper[4955]: E0217 13:04:46.847039 4955 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 13:04:46 crc kubenswrapper[4955]: E0217 13:04:46.847102 4955 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 17 13:04:46 crc kubenswrapper[4955]: E0217 13:04:46.847148 4955 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 17 13:04:46 crc kubenswrapper[4955]: E0217 13:04:46.847165 4955 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 13:04:46 crc kubenswrapper[4955]: E0217 13:04:46.847112 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-17 13:04:50.847102735 +0000 UTC m=+29.369832278 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 13:04:46 crc kubenswrapper[4955]: E0217 13:04:46.847281 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-17 13:04:50.847240489 +0000 UTC m=+29.369970032 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 13:04:46 crc kubenswrapper[4955]: E0217 13:04:46.847004 4955 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 17 13:04:46 crc kubenswrapper[4955]: E0217 13:04:46.847318 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-17 13:04:50.847309421 +0000 UTC m=+29.370039154 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 17 13:04:46 crc kubenswrapper[4955]: I0217 13:04:46.875497 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcbc7a69-58e7-4b7a-bb84-96276819ce88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dcwgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:46Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:46 crc kubenswrapper[4955]: I0217 13:04:46.891420 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d0966253f623158053125dffca956000824438bdc548eba6dafe9a162d5674a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d07bfb361f14d817913fa2cabd30649447790068545dda2f3a2f71213ea0e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:46Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:46 crc kubenswrapper[4955]: I0217 13:04:46.935416 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9ecd4380a6ae8ff62e27e3946815170c2f02b4c3921bda595467345289c9e0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:46Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:46 crc kubenswrapper[4955]: I0217 13:04:46.976254 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0f4056e4b5a554f05004c194485d25231b67024e4376a73c715fcb32004a3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:46Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.012602 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-msn6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://277a12d62c90ef420c31d8dc71f0d7750dfd5f5ab7e90fec0dcd00e54b60245d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2tvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-msn6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:47Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.055687 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05aa39d-94e6-49d4-885e-2c1f986310f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T13:04:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 13:04:35.885968 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 13:04:35.888176 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3462920193/tls.crt::/tmp/serving-cert-3462920193/tls.key\\\\\\\"\\\\nI0217 13:04:42.370546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 13:04:42.384242 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 13:04:42.384284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 13:04:42.384317 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 13:04:42.384327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 13:04:42.394555 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 13:04:42.394614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394628 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 13:04:42.394646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 13:04:42.394654 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 13:04:42.394662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 13:04:42.394900 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 13:04:42.396494 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:47Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.093794 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:47Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.137469 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:47Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.174748 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bf36887-2a2b-4193-abdb-a15beaf46300\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11a8924eb6ce41d629feb2b42c7d4778a3ca9110a59c3ac25acbba4367272374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bec78f9e9f6416d6b600f6231f0d24652473b48b95bf0e0675cc85c9208d124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-29qxq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:47Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.175708 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 13:45:07.191371227 +0000 UTC Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.222184 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9hx9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e57bdf8-e537-402a-9815-609fecdd68ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbc9014d80aee39922161014aac8fafa7d93be3e4092fe705c76da4a487a5e5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q22d4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9hx9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:47Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.222315 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.222303 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:04:47 crc kubenswrapper[4955]: E0217 13:04:47.222494 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.222303 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:04:47 crc kubenswrapper[4955]: E0217 13:04:47.222678 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:04:47 crc kubenswrapper[4955]: E0217 13:04:47.222914 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.244345 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-2gsx6"] Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.244834 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-2gsx6" Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.265318 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.272814 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b49c9af-148a-48ec-af37-f555f9186337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d6919c91918e8bacafd1442b06d14ffd9b547f0f37b2298e2d8d3acc5962d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e38466a067b20ca4323ab6983cb324bb63179f942d2c37c0daa8b03767e7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4830450cbada3d24c9747e9de5f48e53e118549f1dd67edf7fff8b2adc47128f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001dfb595f39d1da33b265474b6cca51c62f522ff061c9e7c52f93806034f4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f1fee7ac5f8d1442d9f63b0e8d285187129eb3a3e3b50a49a45c7a7a39ae8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:47Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.284969 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.304232 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.324484 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.351948 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1face148-45d1-44b3-8775-209f9f8651fc-host\") pod \"node-ca-2gsx6\" (UID: \"1face148-45d1-44b3-8775-209f9f8651fc\") " pod="openshift-image-registry/node-ca-2gsx6" Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.352011 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/1face148-45d1-44b3-8775-209f9f8651fc-serviceca\") pod \"node-ca-2gsx6\" (UID: \"1face148-45d1-44b3-8775-209f9f8651fc\") " pod="openshift-image-registry/node-ca-2gsx6" Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.352093 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnlcr\" (UniqueName: \"kubernetes.io/projected/1face148-45d1-44b3-8775-209f9f8651fc-kube-api-access-xnlcr\") pod \"node-ca-2gsx6\" (UID: \"1face148-45d1-44b3-8775-209f9f8651fc\") " pod="openshift-image-registry/node-ca-2gsx6" Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.375767 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5c34a3d-82ad-4d03-abdd-0a20818ac2fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0261f41b82b492d324b5e50edd72cedd5e824de542a7b47104abd2ab757a7049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f43c7fe16acfd4d39301ee64665d4cd1d885397f981d70f73215da7f3aabe72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c871c1b454e4edd2f41076ecfd6a66592023d7161f094fbb4fd197fbf9c881f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:47Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.415588 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:47Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.452644 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnlcr\" (UniqueName: \"kubernetes.io/projected/1face148-45d1-44b3-8775-209f9f8651fc-kube-api-access-xnlcr\") pod \"node-ca-2gsx6\" (UID: \"1face148-45d1-44b3-8775-209f9f8651fc\") " pod="openshift-image-registry/node-ca-2gsx6" Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.452731 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1face148-45d1-44b3-8775-209f9f8651fc-host\") pod \"node-ca-2gsx6\" (UID: \"1face148-45d1-44b3-8775-209f9f8651fc\") " pod="openshift-image-registry/node-ca-2gsx6" Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.452759 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/1face148-45d1-44b3-8775-209f9f8651fc-serviceca\") pod \"node-ca-2gsx6\" (UID: \"1face148-45d1-44b3-8775-209f9f8651fc\") " pod="openshift-image-registry/node-ca-2gsx6" Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.453578 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1face148-45d1-44b3-8775-209f9f8651fc-host\") pod \"node-ca-2gsx6\" (UID: \"1face148-45d1-44b3-8775-209f9f8651fc\") " pod="openshift-image-registry/node-ca-2gsx6" Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.454390 4955 generic.go:334] "Generic (PLEG): container finished" podID="924c66b7-7c3c-4db7-a427-3f152efbac4e" containerID="b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693" exitCode=0 Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.454517 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" event={"ID":"924c66b7-7c3c-4db7-a427-3f152efbac4e","Type":"ContainerDied","Data":"b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693"} Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.454885 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/1face148-45d1-44b3-8775-209f9f8651fc-serviceca\") pod \"node-ca-2gsx6\" (UID: \"1face148-45d1-44b3-8775-209f9f8651fc\") " pod="openshift-image-registry/node-ca-2gsx6" Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.462840 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"924c66b7-7c3c-4db7-a427-3f152efbac4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fp5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:47Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.488976 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnlcr\" (UniqueName: \"kubernetes.io/projected/1face148-45d1-44b3-8775-209f9f8651fc-kube-api-access-xnlcr\") pod \"node-ca-2gsx6\" (UID: \"1face148-45d1-44b3-8775-209f9f8651fc\") " pod="openshift-image-registry/node-ca-2gsx6" Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.537081 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcbc7a69-58e7-4b7a-bb84-96276819ce88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dcwgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:47Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.561200 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d0966253f623158053125dffca956000824438bdc548eba6dafe9a162d5674a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d07bfb361f14d817913fa2cabd30649447790068545dda2f3a2f71213ea0e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:47Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.562650 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-2gsx6" Feb 17 13:04:47 crc kubenswrapper[4955]: W0217 13:04:47.587850 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1face148_45d1_44b3_8775_209f9f8651fc.slice/crio-4167b51b16c3d1c29bd2e520a221edaf7d0c8b58958221ec617ed69675eee814 WatchSource:0}: Error finding container 4167b51b16c3d1c29bd2e520a221edaf7d0c8b58958221ec617ed69675eee814: Status 404 returned error can't find the container with id 4167b51b16c3d1c29bd2e520a221edaf7d0c8b58958221ec617ed69675eee814 Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.597373 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:47Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.632998 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bf36887-2a2b-4193-abdb-a15beaf46300\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11a8924eb6ce41d629feb2b42c7d4778a3ca9110a59c3ac25acbba4367272374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bec78f9e9f6416d6b600f6231f0d24652473b48b95bf0e0675cc85c9208d124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-29qxq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:47Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.677063 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9hx9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e57bdf8-e537-402a-9815-609fecdd68ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbc9014d80aee39922161014aac8fafa7d93be3e4092fe705c76da4a487a5e5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q22d4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9hx9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:47Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.712639 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-msn6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://277a12d62c90ef420c31d8dc71f0d7750dfd5f5ab7e90fec0dcd00e54b60245d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2tvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-msn6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:47Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.756120 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05aa39d-94e6-49d4-885e-2c1f986310f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T13:04:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 13:04:35.885968 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 13:04:35.888176 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3462920193/tls.crt::/tmp/serving-cert-3462920193/tls.key\\\\\\\"\\\\nI0217 13:04:42.370546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 13:04:42.384242 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 13:04:42.384284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 13:04:42.384317 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 13:04:42.384327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 13:04:42.394555 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 13:04:42.394614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394628 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 13:04:42.394646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 13:04:42.394654 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 13:04:42.394662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 13:04:42.394900 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 13:04:42.396494 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:47Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.796207 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:47Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.834036 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:47Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.884545 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"924c66b7-7c3c-4db7-a427-3f152efbac4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fp5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:47Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.928561 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcbc7a69-58e7-4b7a-bb84-96276819ce88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dcwgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:47Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:47 crc kubenswrapper[4955]: I0217 13:04:47.962243 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b49c9af-148a-48ec-af37-f555f9186337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d6919c91918e8bacafd1442b06d14ffd9b547f0f37b2298e2d8d3acc5962d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e38466a067b20ca4323ab6983cb324bb63179f942d2c37c0daa8b03767e7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4830450cbada3d24c9747e9de5f48e53e118549f1dd67edf7fff8b2adc47128f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001dfb595f39d1da33b265474b6cca51c62f522ff061c9e7c52f93806034f4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f1fee7ac5f8d1442d9f63b0e8d285187129eb3a3e3b50a49a45c7a7a39ae8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:47Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.001382 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5c34a3d-82ad-4d03-abdd-0a20818ac2fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0261f41b82b492d324b5e50edd72cedd5e824de542a7b47104abd2ab757a7049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f43c7fe16acfd4d39301ee64665d4cd1d885397f981d70f73215da7f3aabe72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c871c1b454e4edd2f41076ecfd6a66592023d7161f094fbb4fd197fbf9c881f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:47Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.034902 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2gsx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1face148-45d1-44b3-8775-209f9f8651fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnlcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2gsx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:48Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.077268 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d0966253f623158053125dffca956000824438bdc548eba6dafe9a162d5674a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d07bfb361f14d817913fa2cabd30649447790068545dda2f3a2f71213ea0e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:48Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.120972 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0f4056e4b5a554f05004c194485d25231b67024e4376a73c715fcb32004a3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:48Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.163630 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9ecd4380a6ae8ff62e27e3946815170c2f02b4c3921bda595467345289c9e0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:48Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.188994 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 20:24:08.452791892 +0000 UTC Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.461122 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-2gsx6" event={"ID":"1face148-45d1-44b3-8775-209f9f8651fc","Type":"ContainerStarted","Data":"82c5147255d021114a8639e9b60593f847896b613b5d862db588014f7485faa5"} Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.461202 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-2gsx6" event={"ID":"1face148-45d1-44b3-8775-209f9f8651fc","Type":"ContainerStarted","Data":"4167b51b16c3d1c29bd2e520a221edaf7d0c8b58958221ec617ed69675eee814"} Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.466147 4955 generic.go:334] "Generic (PLEG): container finished" podID="924c66b7-7c3c-4db7-a427-3f152efbac4e" containerID="35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461" exitCode=0 Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.466204 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" event={"ID":"924c66b7-7c3c-4db7-a427-3f152efbac4e","Type":"ContainerDied","Data":"35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461"} Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.508340 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b49c9af-148a-48ec-af37-f555f9186337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d6919c91918e8bacafd1442b06d14ffd9b547f0f37b2298e2d8d3acc5962d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e38466a067b20ca4323ab6983cb324bb63179f942d2c37c0daa8b03767e7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4830450cbada3d24c9747e9de5f48e53e118549f1dd67edf7fff8b2adc47128f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001dfb595f39d1da33b265474b6cca51c62f522ff061c9e7c52f93806034f4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f1fee7ac5f8d1442d9f63b0e8d285187129eb3a3e3b50a49a45c7a7a39ae8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:48Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.532259 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5c34a3d-82ad-4d03-abdd-0a20818ac2fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0261f41b82b492d324b5e50edd72cedd5e824de542a7b47104abd2ab757a7049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f43c7fe16acfd4d39301ee64665d4cd1d885397f981d70f73215da7f3aabe72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c871c1b454e4edd2f41076ecfd6a66592023d7161f094fbb4fd197fbf9c881f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:48Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.555448 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:48Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.586240 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"924c66b7-7c3c-4db7-a427-3f152efbac4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fp5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:48Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.612497 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcbc7a69-58e7-4b7a-bb84-96276819ce88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dcwgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:48Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.630489 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d0966253f623158053125dffca956000824438bdc548eba6dafe9a162d5674a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d07bfb361f14d817913fa2cabd30649447790068545dda2f3a2f71213ea0e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:48Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.647957 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2gsx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1face148-45d1-44b3-8775-209f9f8651fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82c5147255d021114a8639e9b60593f847896b613b5d862db588014f7485faa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnlcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2gsx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:48Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.666927 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9ecd4380a6ae8ff62e27e3946815170c2f02b4c3921bda595467345289c9e0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:48Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.686096 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0f4056e4b5a554f05004c194485d25231b67024e4376a73c715fcb32004a3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:48Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.718589 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05aa39d-94e6-49d4-885e-2c1f986310f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T13:04:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 13:04:35.885968 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 13:04:35.888176 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3462920193/tls.crt::/tmp/serving-cert-3462920193/tls.key\\\\\\\"\\\\nI0217 13:04:42.370546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 13:04:42.384242 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 13:04:42.384284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 13:04:42.384317 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 13:04:42.384327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 13:04:42.394555 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 13:04:42.394614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394628 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 13:04:42.394646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 13:04:42.394654 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 13:04:42.394662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 13:04:42.394900 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 13:04:42.396494 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:48Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.737536 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:48Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.755600 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:48Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.758596 4955 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.763068 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.763132 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.763151 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.763366 4955 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.772508 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bf36887-2a2b-4193-abdb-a15beaf46300\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11a8924eb6ce41d629feb2b42c7d4778a3ca9110a59c3ac25acbba4367272374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bec78f9e9f6416d6b600f6231f0d24652473b48b95bf0e0675cc85c9208d124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-29qxq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:48Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.772971 4955 kubelet_node_status.go:115] "Node was previously registered" node="crc" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.773370 4955 kubelet_node_status.go:79] "Successfully registered node" node="crc" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.775367 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.775406 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.775430 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.775454 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.775469 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:48Z","lastTransitionTime":"2026-02-17T13:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.791054 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9hx9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e57bdf8-e537-402a-9815-609fecdd68ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbc9014d80aee39922161014aac8fafa7d93be3e4092fe705c76da4a487a5e5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q22d4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9hx9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:48Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:48 crc kubenswrapper[4955]: E0217 13:04:48.793851 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7605a498-360b-4ce4-b7a1-dfbc354d798d\\\",\\\"systemUUID\\\":\\\"6d4602d9-c718-4393-8f91-3c381f800c9e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:48Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.799124 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.799173 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.799192 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.799219 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.799238 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:48Z","lastTransitionTime":"2026-02-17T13:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.806344 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-msn6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://277a12d62c90ef420c31d8dc71f0d7750dfd5f5ab7e90fec0dcd00e54b60245d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2tvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-msn6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:48Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:48 crc kubenswrapper[4955]: E0217 13:04:48.819271 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7605a498-360b-4ce4-b7a1-dfbc354d798d\\\",\\\"systemUUID\\\":\\\"6d4602d9-c718-4393-8f91-3c381f800c9e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:48Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.824050 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.824099 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.824112 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.824137 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.824154 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:48Z","lastTransitionTime":"2026-02-17T13:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.837423 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:48Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:48 crc kubenswrapper[4955]: E0217 13:04:48.841993 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7605a498-360b-4ce4-b7a1-dfbc354d798d\\\",\\\"systemUUID\\\":\\\"6d4602d9-c718-4393-8f91-3c381f800c9e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:48Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.847089 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.847149 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.847164 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.847187 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.847202 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:48Z","lastTransitionTime":"2026-02-17T13:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:48 crc kubenswrapper[4955]: E0217 13:04:48.866098 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7605a498-360b-4ce4-b7a1-dfbc354d798d\\\",\\\"systemUUID\\\":\\\"6d4602d9-c718-4393-8f91-3c381f800c9e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:48Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.871676 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.871747 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.871766 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.871814 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.871833 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:48Z","lastTransitionTime":"2026-02-17T13:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.876651 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bf36887-2a2b-4193-abdb-a15beaf46300\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11a8924eb6ce41d629feb2b42c7d4778a3ca9110a59c3ac25acbba4367272374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bec78f9e9f6416d6b600f6231f0d24652473b48b95bf0e0675cc85c9208d124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-29qxq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:48Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:48 crc kubenswrapper[4955]: E0217 13:04:48.885858 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7605a498-360b-4ce4-b7a1-dfbc354d798d\\\",\\\"systemUUID\\\":\\\"6d4602d9-c718-4393-8f91-3c381f800c9e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:48Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:48 crc kubenswrapper[4955]: E0217 13:04:48.886110 4955 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.888400 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.888466 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.888483 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.888505 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.888521 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:48Z","lastTransitionTime":"2026-02-17T13:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.916277 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9hx9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e57bdf8-e537-402a-9815-609fecdd68ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbc9014d80aee39922161014aac8fafa7d93be3e4092fe705c76da4a487a5e5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q22d4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9hx9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:48Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.957358 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-msn6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://277a12d62c90ef420c31d8dc71f0d7750dfd5f5ab7e90fec0dcd00e54b60245d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2tvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-msn6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:48Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.991468 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.991523 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.991540 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.991565 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:48 crc kubenswrapper[4955]: I0217 13:04:48.991579 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:48Z","lastTransitionTime":"2026-02-17T13:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.001160 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05aa39d-94e6-49d4-885e-2c1f986310f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T13:04:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 13:04:35.885968 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 13:04:35.888176 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3462920193/tls.crt::/tmp/serving-cert-3462920193/tls.key\\\\\\\"\\\\nI0217 13:04:42.370546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 13:04:42.384242 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 13:04:42.384284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 13:04:42.384317 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 13:04:42.384327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 13:04:42.394555 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 13:04:42.394614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394628 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 13:04:42.394646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 13:04:42.394654 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 13:04:42.394662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 13:04:42.394900 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 13:04:42.396494 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:48Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.042456 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:49Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.076109 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:49Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.095405 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.095470 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.095490 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.095520 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.095545 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:49Z","lastTransitionTime":"2026-02-17T13:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.125672 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"924c66b7-7c3c-4db7-a427-3f152efbac4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fp5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:49Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.171076 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcbc7a69-58e7-4b7a-bb84-96276819ce88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dcwgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:49Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.189146 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 05:27:46.53961287 +0000 UTC Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.198810 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.198860 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.198874 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.198899 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.198915 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:49Z","lastTransitionTime":"2026-02-17T13:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.209616 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b49c9af-148a-48ec-af37-f555f9186337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d6919c91918e8bacafd1442b06d14ffd9b547f0f37b2298e2d8d3acc5962d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e38466a067b20ca4323ab6983cb324bb63179f942d2c37c0daa8b03767e7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4830450cbada3d24c9747e9de5f48e53e118549f1dd67edf7fff8b2adc47128f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001dfb595f39d1da33b265474b6cca51c62f522ff061c9e7c52f93806034f4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f1fee7ac5f8d1442d9f63b0e8d285187129eb3a3e3b50a49a45c7a7a39ae8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:49Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.222237 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:04:49 crc kubenswrapper[4955]: E0217 13:04:49.222471 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.222521 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.222592 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:04:49 crc kubenswrapper[4955]: E0217 13:04:49.222755 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:04:49 crc kubenswrapper[4955]: E0217 13:04:49.222971 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.243912 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5c34a3d-82ad-4d03-abdd-0a20818ac2fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0261f41b82b492d324b5e50edd72cedd5e824de542a7b47104abd2ab757a7049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f43c7fe16acfd4d39301ee64665d4cd1d885397f981d70f73215da7f3aabe72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c871c1b454e4edd2f41076ecfd6a66592023d7161f094fbb4fd197fbf9c881f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:49Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.275821 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2gsx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1face148-45d1-44b3-8775-209f9f8651fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82c5147255d021114a8639e9b60593f847896b613b5d862db588014f7485faa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnlcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2gsx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:49Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.304036 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.304108 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.304131 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.304163 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.304187 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:49Z","lastTransitionTime":"2026-02-17T13:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.319452 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d0966253f623158053125dffca956000824438bdc548eba6dafe9a162d5674a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d07bfb361f14d817913fa2cabd30649447790068545dda2f3a2f71213ea0e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:49Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.352313 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0f4056e4b5a554f05004c194485d25231b67024e4376a73c715fcb32004a3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:49Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.394953 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9ecd4380a6ae8ff62e27e3946815170c2f02b4c3921bda595467345289c9e0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:49Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.406992 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.407025 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.407037 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.407057 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.407070 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:49Z","lastTransitionTime":"2026-02-17T13:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.476995 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" event={"ID":"dcbc7a69-58e7-4b7a-bb84-96276819ce88","Type":"ContainerStarted","Data":"e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648"} Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.480332 4955 generic.go:334] "Generic (PLEG): container finished" podID="924c66b7-7c3c-4db7-a427-3f152efbac4e" containerID="36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af" exitCode=0 Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.480363 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" event={"ID":"924c66b7-7c3c-4db7-a427-3f152efbac4e","Type":"ContainerDied","Data":"36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af"} Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.510947 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.511028 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.511054 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.511085 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.511111 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:49Z","lastTransitionTime":"2026-02-17T13:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.514280 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b49c9af-148a-48ec-af37-f555f9186337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d6919c91918e8bacafd1442b06d14ffd9b547f0f37b2298e2d8d3acc5962d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e38466a067b20ca4323ab6983cb324bb63179f942d2c37c0daa8b03767e7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4830450cbada3d24c9747e9de5f48e53e118549f1dd67edf7fff8b2adc47128f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001dfb595f39d1da33b265474b6cca51c62f522ff061c9e7c52f93806034f4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f1fee7ac5f8d1442d9f63b0e8d285187129eb3a3e3b50a49a45c7a7a39ae8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:49Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.534934 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5c34a3d-82ad-4d03-abdd-0a20818ac2fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0261f41b82b492d324b5e50edd72cedd5e824de542a7b47104abd2ab757a7049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f43c7fe16acfd4d39301ee64665d4cd1d885397f981d70f73215da7f3aabe72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c871c1b454e4edd2f41076ecfd6a66592023d7161f094fbb4fd197fbf9c881f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:49Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.554656 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:49Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.583047 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"924c66b7-7c3c-4db7-a427-3f152efbac4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fp5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:49Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.609352 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcbc7a69-58e7-4b7a-bb84-96276819ce88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dcwgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:49Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.615431 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.615476 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.615493 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.615524 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.615543 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:49Z","lastTransitionTime":"2026-02-17T13:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.636903 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d0966253f623158053125dffca956000824438bdc548eba6dafe9a162d5674a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d07bfb361f14d817913fa2cabd30649447790068545dda2f3a2f71213ea0e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:49Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.672080 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2gsx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1face148-45d1-44b3-8775-209f9f8651fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82c5147255d021114a8639e9b60593f847896b613b5d862db588014f7485faa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnlcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2gsx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:49Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.717945 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.718078 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.718088 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.718104 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.718151 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9ecd4380a6ae8ff62e27e3946815170c2f02b4c3921bda595467345289c9e0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:49Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.718222 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:49Z","lastTransitionTime":"2026-02-17T13:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.755733 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0f4056e4b5a554f05004c194485d25231b67024e4376a73c715fcb32004a3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:49Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.793378 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-msn6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://277a12d62c90ef420c31d8dc71f0d7750dfd5f5ab7e90fec0dcd00e54b60245d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2tvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-msn6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:49Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.822856 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.822916 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.822936 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.822965 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.822984 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:49Z","lastTransitionTime":"2026-02-17T13:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.835767 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05aa39d-94e6-49d4-885e-2c1f986310f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T13:04:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 13:04:35.885968 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 13:04:35.888176 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3462920193/tls.crt::/tmp/serving-cert-3462920193/tls.key\\\\\\\"\\\\nI0217 13:04:42.370546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 13:04:42.384242 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 13:04:42.384284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 13:04:42.384317 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 13:04:42.384327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 13:04:42.394555 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 13:04:42.394614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394628 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 13:04:42.394646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 13:04:42.394654 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 13:04:42.394662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 13:04:42.394900 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 13:04:42.396494 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:49Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.884893 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:49Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.926228 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.926292 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.926309 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.926339 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.926358 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:49Z","lastTransitionTime":"2026-02-17T13:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.927129 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:49Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.956626 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bf36887-2a2b-4193-abdb-a15beaf46300\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11a8924eb6ce41d629feb2b42c7d4778a3ca9110a59c3ac25acbba4367272374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bec78f9e9f6416d6b600f6231f0d24652473b48b95bf0e0675cc85c9208d124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-29qxq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:49Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:49 crc kubenswrapper[4955]: I0217 13:04:49.992802 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9hx9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e57bdf8-e537-402a-9815-609fecdd68ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbc9014d80aee39922161014aac8fafa7d93be3e4092fe705c76da4a487a5e5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q22d4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9hx9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:49Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.029675 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.029755 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.029775 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.029843 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.029862 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:50Z","lastTransitionTime":"2026-02-17T13:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.132645 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.132707 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.132724 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.132754 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.132773 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:50Z","lastTransitionTime":"2026-02-17T13:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.189505 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-30 12:24:44.024186016 +0000 UTC Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.236126 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.236164 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.236176 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.236193 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.236207 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:50Z","lastTransitionTime":"2026-02-17T13:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.340216 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.340299 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.340323 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.340362 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.340401 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:50Z","lastTransitionTime":"2026-02-17T13:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.444585 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.444664 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.444689 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.444724 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.444749 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:50Z","lastTransitionTime":"2026-02-17T13:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.490022 4955 generic.go:334] "Generic (PLEG): container finished" podID="924c66b7-7c3c-4db7-a427-3f152efbac4e" containerID="4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0" exitCode=0 Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.490110 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" event={"ID":"924c66b7-7c3c-4db7-a427-3f152efbac4e","Type":"ContainerDied","Data":"4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0"} Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.518503 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05aa39d-94e6-49d4-885e-2c1f986310f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T13:04:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 13:04:35.885968 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 13:04:35.888176 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3462920193/tls.crt::/tmp/serving-cert-3462920193/tls.key\\\\\\\"\\\\nI0217 13:04:42.370546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 13:04:42.384242 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 13:04:42.384284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 13:04:42.384317 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 13:04:42.384327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 13:04:42.394555 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 13:04:42.394614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394628 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 13:04:42.394646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 13:04:42.394654 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 13:04:42.394662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 13:04:42.394900 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 13:04:42.396494 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:50Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.548012 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.548064 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.548081 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.548106 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.548124 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:50Z","lastTransitionTime":"2026-02-17T13:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.549130 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:50Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.571614 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:50Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.592658 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bf36887-2a2b-4193-abdb-a15beaf46300\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11a8924eb6ce41d629feb2b42c7d4778a3ca9110a59c3ac25acbba4367272374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bec78f9e9f6416d6b600f6231f0d24652473b48b95bf0e0675cc85c9208d124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-29qxq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:50Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.616506 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9hx9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e57bdf8-e537-402a-9815-609fecdd68ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbc9014d80aee39922161014aac8fafa7d93be3e4092fe705c76da4a487a5e5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q22d4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9hx9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:50Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.640247 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-msn6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://277a12d62c90ef420c31d8dc71f0d7750dfd5f5ab7e90fec0dcd00e54b60245d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2tvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-msn6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:50Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.651743 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.651870 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.651940 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.651971 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.652029 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:50Z","lastTransitionTime":"2026-02-17T13:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.671527 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b49c9af-148a-48ec-af37-f555f9186337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d6919c91918e8bacafd1442b06d14ffd9b547f0f37b2298e2d8d3acc5962d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e38466a067b20ca4323ab6983cb324bb63179f942d2c37c0daa8b03767e7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4830450cbada3d24c9747e9de5f48e53e118549f1dd67edf7fff8b2adc47128f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001dfb595f39d1da33b265474b6cca51c62f522ff061c9e7c52f93806034f4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f1fee7ac5f8d1442d9f63b0e8d285187129eb3a3e3b50a49a45c7a7a39ae8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:50Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.693946 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5c34a3d-82ad-4d03-abdd-0a20818ac2fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0261f41b82b492d324b5e50edd72cedd5e824de542a7b47104abd2ab757a7049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f43c7fe16acfd4d39301ee64665d4cd1d885397f981d70f73215da7f3aabe72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c871c1b454e4edd2f41076ecfd6a66592023d7161f094fbb4fd197fbf9c881f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:50Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.714706 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:50Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.747252 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"924c66b7-7c3c-4db7-a427-3f152efbac4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fp5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:50Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.756162 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.756235 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.756248 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.756266 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.756279 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:50Z","lastTransitionTime":"2026-02-17T13:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.782542 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcbc7a69-58e7-4b7a-bb84-96276819ce88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dcwgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:50Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.804754 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d0966253f623158053125dffca956000824438bdc548eba6dafe9a162d5674a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d07bfb361f14d817913fa2cabd30649447790068545dda2f3a2f71213ea0e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:50Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.822180 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2gsx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1face148-45d1-44b3-8775-209f9f8651fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82c5147255d021114a8639e9b60593f847896b613b5d862db588014f7485faa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnlcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2gsx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:50Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.847518 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9ecd4380a6ae8ff62e27e3946815170c2f02b4c3921bda595467345289c9e0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:50Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.859729 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.859813 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.859834 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.859862 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.859884 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:50Z","lastTransitionTime":"2026-02-17T13:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.862905 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0f4056e4b5a554f05004c194485d25231b67024e4376a73c715fcb32004a3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:50Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.926987 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:04:50 crc kubenswrapper[4955]: E0217 13:04:50.927320 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:04:58.927271978 +0000 UTC m=+37.450001561 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.927419 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.927491 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.927549 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.927618 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:04:50 crc kubenswrapper[4955]: E0217 13:04:50.927740 4955 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 17 13:04:50 crc kubenswrapper[4955]: E0217 13:04:50.928487 4955 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 17 13:04:50 crc kubenswrapper[4955]: E0217 13:04:50.928517 4955 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 13:04:50 crc kubenswrapper[4955]: E0217 13:04:50.927850 4955 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 17 13:04:50 crc kubenswrapper[4955]: E0217 13:04:50.928623 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-17 13:04:58.928590095 +0000 UTC m=+37.451319678 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 13:04:50 crc kubenswrapper[4955]: E0217 13:04:50.927857 4955 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 17 13:04:50 crc kubenswrapper[4955]: E0217 13:04:50.928728 4955 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 17 13:04:50 crc kubenswrapper[4955]: E0217 13:04:50.928739 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-17 13:04:58.928697158 +0000 UTC m=+37.451426711 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 17 13:04:50 crc kubenswrapper[4955]: E0217 13:04:50.928753 4955 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 13:04:50 crc kubenswrapper[4955]: E0217 13:04:50.927938 4955 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 17 13:04:50 crc kubenswrapper[4955]: E0217 13:04:50.928836 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-17 13:04:58.928826331 +0000 UTC m=+37.451555884 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 17 13:04:50 crc kubenswrapper[4955]: E0217 13:04:50.928851 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-17 13:04:58.928844292 +0000 UTC m=+37.451573845 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.962897 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.962942 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.962957 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.962976 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:50 crc kubenswrapper[4955]: I0217 13:04:50.963006 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:50Z","lastTransitionTime":"2026-02-17T13:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.066223 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.066357 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.066384 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.066417 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.066440 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:51Z","lastTransitionTime":"2026-02-17T13:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.170853 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.170906 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.170921 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.170943 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.170957 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:51Z","lastTransitionTime":"2026-02-17T13:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.190514 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 19:55:45.883934297 +0000 UTC Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.222938 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.223000 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.222942 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:04:51 crc kubenswrapper[4955]: E0217 13:04:51.223099 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:04:51 crc kubenswrapper[4955]: E0217 13:04:51.223188 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:04:51 crc kubenswrapper[4955]: E0217 13:04:51.223488 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.273973 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.274035 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.274058 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.274088 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.274109 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:51Z","lastTransitionTime":"2026-02-17T13:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.377714 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.377809 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.377829 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.377857 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.377879 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:51Z","lastTransitionTime":"2026-02-17T13:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.481667 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.481716 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.481728 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.481750 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.481764 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:51Z","lastTransitionTime":"2026-02-17T13:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.501105 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" event={"ID":"924c66b7-7c3c-4db7-a427-3f152efbac4e","Type":"ContainerStarted","Data":"9beee007a179c3c2e04c2ce0de8ea6dbae9da3421ab24f23b718a23744c828a3"} Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.511991 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" event={"ID":"dcbc7a69-58e7-4b7a-bb84-96276819ce88","Type":"ContainerStarted","Data":"76543c6dcedac4d9e47a49af549fb14d31c1b0561a025f81788de7d80bb39e33"} Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.512448 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.512491 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.527196 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9ecd4380a6ae8ff62e27e3946815170c2f02b4c3921bda595467345289c9e0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:51Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.554475 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0f4056e4b5a554f05004c194485d25231b67024e4376a73c715fcb32004a3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:51Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.558076 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.559742 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.578278 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05aa39d-94e6-49d4-885e-2c1f986310f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T13:04:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 13:04:35.885968 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 13:04:35.888176 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3462920193/tls.crt::/tmp/serving-cert-3462920193/tls.key\\\\\\\"\\\\nI0217 13:04:42.370546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 13:04:42.384242 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 13:04:42.384284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 13:04:42.384317 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 13:04:42.384327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 13:04:42.394555 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 13:04:42.394614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394628 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 13:04:42.394646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 13:04:42.394654 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 13:04:42.394662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 13:04:42.394900 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 13:04:42.396494 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:51Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.585516 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.585583 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.585602 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.585627 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.585648 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:51Z","lastTransitionTime":"2026-02-17T13:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.596524 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:51Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.612950 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:51Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.627543 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bf36887-2a2b-4193-abdb-a15beaf46300\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11a8924eb6ce41d629feb2b42c7d4778a3ca9110a59c3ac25acbba4367272374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bec78f9e9f6416d6b600f6231f0d24652473b48b95bf0e0675cc85c9208d124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-29qxq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:51Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.655588 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9hx9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e57bdf8-e537-402a-9815-609fecdd68ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbc9014d80aee39922161014aac8fafa7d93be3e4092fe705c76da4a487a5e5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q22d4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9hx9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:51Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.669230 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-msn6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://277a12d62c90ef420c31d8dc71f0d7750dfd5f5ab7e90fec0dcd00e54b60245d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2tvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-msn6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:51Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.688512 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b49c9af-148a-48ec-af37-f555f9186337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d6919c91918e8bacafd1442b06d14ffd9b547f0f37b2298e2d8d3acc5962d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e38466a067b20ca4323ab6983cb324bb63179f942d2c37c0daa8b03767e7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4830450cbada3d24c9747e9de5f48e53e118549f1dd67edf7fff8b2adc47128f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001dfb595f39d1da33b265474b6cca51c62f522ff061c9e7c52f93806034f4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f1fee7ac5f8d1442d9f63b0e8d285187129eb3a3e3b50a49a45c7a7a39ae8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:51Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.691091 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.691191 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.691256 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.691284 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.691299 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:51Z","lastTransitionTime":"2026-02-17T13:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.706308 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5c34a3d-82ad-4d03-abdd-0a20818ac2fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0261f41b82b492d324b5e50edd72cedd5e824de542a7b47104abd2ab757a7049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f43c7fe16acfd4d39301ee64665d4cd1d885397f981d70f73215da7f3aabe72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c871c1b454e4edd2f41076ecfd6a66592023d7161f094fbb4fd197fbf9c881f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:51Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.730564 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:51Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.760314 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"924c66b7-7c3c-4db7-a427-3f152efbac4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9beee007a179c3c2e04c2ce0de8ea6dbae9da3421ab24f23b718a23744c828a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fp5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:51Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.793342 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcbc7a69-58e7-4b7a-bb84-96276819ce88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dcwgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:51Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.795705 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.795769 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.795816 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.795846 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.795866 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:51Z","lastTransitionTime":"2026-02-17T13:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.811201 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d0966253f623158053125dffca956000824438bdc548eba6dafe9a162d5674a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d07bfb361f14d817913fa2cabd30649447790068545dda2f3a2f71213ea0e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:51Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.826655 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2gsx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1face148-45d1-44b3-8775-209f9f8651fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82c5147255d021114a8639e9b60593f847896b613b5d862db588014f7485faa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnlcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2gsx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:51Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.860502 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b49c9af-148a-48ec-af37-f555f9186337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d6919c91918e8bacafd1442b06d14ffd9b547f0f37b2298e2d8d3acc5962d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e38466a067b20ca4323ab6983cb324bb63179f942d2c37c0daa8b03767e7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4830450cbada3d24c9747e9de5f48e53e118549f1dd67edf7fff8b2adc47128f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001dfb595f39d1da33b265474b6cca51c62f522ff061c9e7c52f93806034f4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f1fee7ac5f8d1442d9f63b0e8d285187129eb3a3e3b50a49a45c7a7a39ae8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:51Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.880618 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5c34a3d-82ad-4d03-abdd-0a20818ac2fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0261f41b82b492d324b5e50edd72cedd5e824de542a7b47104abd2ab757a7049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f43c7fe16acfd4d39301ee64665d4cd1d885397f981d70f73215da7f3aabe72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c871c1b454e4edd2f41076ecfd6a66592023d7161f094fbb4fd197fbf9c881f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:51Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.898797 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.898842 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.898855 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.898876 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.898891 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:51Z","lastTransitionTime":"2026-02-17T13:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.900422 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:51Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.924978 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"924c66b7-7c3c-4db7-a427-3f152efbac4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9beee007a179c3c2e04c2ce0de8ea6dbae9da3421ab24f23b718a23744c828a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fp5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:51Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.952327 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcbc7a69-58e7-4b7a-bb84-96276819ce88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76543c6dcedac4d9e47a49af549fb14d31c1b0561a025f81788de7d80bb39e33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dcwgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:51Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.965927 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d0966253f623158053125dffca956000824438bdc548eba6dafe9a162d5674a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d07bfb361f14d817913fa2cabd30649447790068545dda2f3a2f71213ea0e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:51Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.973968 4955 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Feb 17 13:04:51 crc kubenswrapper[4955]: I0217 13:04:51.974860 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2gsx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1face148-45d1-44b3-8775-209f9f8651fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82c5147255d021114a8639e9b60593f847896b613b5d862db588014f7485faa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnlcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2gsx6\": Patch \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/pods/node-ca-2gsx6/status\": read tcp 38.102.83.46:45546->38.102.83.46:6443: use of closed network connection" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.010132 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.010186 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.010202 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.010227 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.010246 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:52Z","lastTransitionTime":"2026-02-17T13:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.020473 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9ecd4380a6ae8ff62e27e3946815170c2f02b4c3921bda595467345289c9e0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:52Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.035335 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0f4056e4b5a554f05004c194485d25231b67024e4376a73c715fcb32004a3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:52Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.052250 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05aa39d-94e6-49d4-885e-2c1f986310f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T13:04:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 13:04:35.885968 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 13:04:35.888176 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3462920193/tls.crt::/tmp/serving-cert-3462920193/tls.key\\\\\\\"\\\\nI0217 13:04:42.370546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 13:04:42.384242 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 13:04:42.384284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 13:04:42.384317 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 13:04:42.384327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 13:04:42.394555 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 13:04:42.394614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394628 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 13:04:42.394646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 13:04:42.394654 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 13:04:42.394662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 13:04:42.394900 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 13:04:42.396494 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:52Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.066947 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:52Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.086375 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:52Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.102839 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bf36887-2a2b-4193-abdb-a15beaf46300\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11a8924eb6ce41d629feb2b42c7d4778a3ca9110a59c3ac25acbba4367272374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bec78f9e9f6416d6b600f6231f0d24652473b48b95bf0e0675cc85c9208d124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-29qxq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:52Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.114005 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.114084 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.114112 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.114143 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.114162 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:52Z","lastTransitionTime":"2026-02-17T13:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.119107 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9hx9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e57bdf8-e537-402a-9815-609fecdd68ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbc9014d80aee39922161014aac8fafa7d93be3e4092fe705c76da4a487a5e5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q22d4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9hx9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:52Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.134375 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-msn6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://277a12d62c90ef420c31d8dc71f0d7750dfd5f5ab7e90fec0dcd00e54b60245d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2tvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-msn6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:52Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.191340 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 03:08:36.262134996 +0000 UTC Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.216826 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.216876 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.216894 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.216923 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.216941 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:52Z","lastTransitionTime":"2026-02-17T13:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.240771 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9ecd4380a6ae8ff62e27e3946815170c2f02b4c3921bda595467345289c9e0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:52Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.255380 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0f4056e4b5a554f05004c194485d25231b67024e4376a73c715fcb32004a3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:52Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.272177 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05aa39d-94e6-49d4-885e-2c1f986310f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T13:04:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 13:04:35.885968 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 13:04:35.888176 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3462920193/tls.crt::/tmp/serving-cert-3462920193/tls.key\\\\\\\"\\\\nI0217 13:04:42.370546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 13:04:42.384242 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 13:04:42.384284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 13:04:42.384317 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 13:04:42.384327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 13:04:42.394555 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 13:04:42.394614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394628 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 13:04:42.394646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 13:04:42.394654 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 13:04:42.394662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 13:04:42.394900 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 13:04:42.396494 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:52Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.288611 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:52Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.301956 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:52Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.319304 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bf36887-2a2b-4193-abdb-a15beaf46300\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11a8924eb6ce41d629feb2b42c7d4778a3ca9110a59c3ac25acbba4367272374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bec78f9e9f6416d6b600f6231f0d24652473b48b95bf0e0675cc85c9208d124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-29qxq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:52Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.319764 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.319851 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.319883 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.319911 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.319931 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:52Z","lastTransitionTime":"2026-02-17T13:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.336538 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9hx9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e57bdf8-e537-402a-9815-609fecdd68ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbc9014d80aee39922161014aac8fafa7d93be3e4092fe705c76da4a487a5e5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q22d4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9hx9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:52Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.354962 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-msn6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://277a12d62c90ef420c31d8dc71f0d7750dfd5f5ab7e90fec0dcd00e54b60245d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2tvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-msn6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:52Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.388369 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b49c9af-148a-48ec-af37-f555f9186337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d6919c91918e8bacafd1442b06d14ffd9b547f0f37b2298e2d8d3acc5962d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e38466a067b20ca4323ab6983cb324bb63179f942d2c37c0daa8b03767e7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4830450cbada3d24c9747e9de5f48e53e118549f1dd67edf7fff8b2adc47128f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001dfb595f39d1da33b265474b6cca51c62f522ff061c9e7c52f93806034f4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f1fee7ac5f8d1442d9f63b0e8d285187129eb3a3e3b50a49a45c7a7a39ae8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:52Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.405567 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5c34a3d-82ad-4d03-abdd-0a20818ac2fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0261f41b82b492d324b5e50edd72cedd5e824de542a7b47104abd2ab757a7049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f43c7fe16acfd4d39301ee64665d4cd1d885397f981d70f73215da7f3aabe72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c871c1b454e4edd2f41076ecfd6a66592023d7161f094fbb4fd197fbf9c881f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:52Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.422603 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.422653 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.422668 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.422691 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.422713 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:52Z","lastTransitionTime":"2026-02-17T13:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.422988 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:52Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.441595 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"924c66b7-7c3c-4db7-a427-3f152efbac4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9beee007a179c3c2e04c2ce0de8ea6dbae9da3421ab24f23b718a23744c828a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fp5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:52Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.463798 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcbc7a69-58e7-4b7a-bb84-96276819ce88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76543c6dcedac4d9e47a49af549fb14d31c1b0561a025f81788de7d80bb39e33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dcwgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:52Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.486608 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d0966253f623158053125dffca956000824438bdc548eba6dafe9a162d5674a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d07bfb361f14d817913fa2cabd30649447790068545dda2f3a2f71213ea0e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:52Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.499669 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2gsx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1face148-45d1-44b3-8775-209f9f8651fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82c5147255d021114a8639e9b60593f847896b613b5d862db588014f7485faa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnlcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2gsx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:52Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.516767 4955 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.526043 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.526156 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.526220 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.526251 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.526311 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:52Z","lastTransitionTime":"2026-02-17T13:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.629180 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.629223 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.629237 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.629255 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.629266 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:52Z","lastTransitionTime":"2026-02-17T13:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.733312 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.733377 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.733478 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.733509 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.733531 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:52Z","lastTransitionTime":"2026-02-17T13:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.836656 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.836714 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.836726 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.836746 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.836761 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:52Z","lastTransitionTime":"2026-02-17T13:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.939735 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.939828 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.939847 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.939876 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:52 crc kubenswrapper[4955]: I0217 13:04:52.939894 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:52Z","lastTransitionTime":"2026-02-17T13:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.042272 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.042337 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.042345 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.042365 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.042374 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:53Z","lastTransitionTime":"2026-02-17T13:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.145161 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.145219 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.145231 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.145251 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.145265 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:53Z","lastTransitionTime":"2026-02-17T13:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.220196 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 14:46:35.185189918 +0000 UTC Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.222667 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.222830 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:04:53 crc kubenswrapper[4955]: E0217 13:04:53.222898 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:04:53 crc kubenswrapper[4955]: E0217 13:04:53.223074 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.222838 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:04:53 crc kubenswrapper[4955]: E0217 13:04:53.223246 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.248329 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.248391 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.248409 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.248435 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.248453 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:53Z","lastTransitionTime":"2026-02-17T13:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.351774 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.351891 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.351917 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.351945 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.351964 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:53Z","lastTransitionTime":"2026-02-17T13:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.455074 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.455141 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.455158 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.455185 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.455203 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:53Z","lastTransitionTime":"2026-02-17T13:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.521511 4955 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.558188 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.558244 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.558256 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.558280 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.558296 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:53Z","lastTransitionTime":"2026-02-17T13:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.662298 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.662345 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.662358 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.662378 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.662392 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:53Z","lastTransitionTime":"2026-02-17T13:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.765842 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.765918 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.765956 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.765978 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.765992 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:53Z","lastTransitionTime":"2026-02-17T13:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.869678 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.869754 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.869769 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.869828 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.869851 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:53Z","lastTransitionTime":"2026-02-17T13:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.973421 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.973495 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.973515 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.973546 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:53 crc kubenswrapper[4955]: I0217 13:04:53.973569 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:53Z","lastTransitionTime":"2026-02-17T13:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.076415 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.076489 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.076508 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.076537 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.076560 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:54Z","lastTransitionTime":"2026-02-17T13:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.179422 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.179489 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.179516 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.179547 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.179569 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:54Z","lastTransitionTime":"2026-02-17T13:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.220883 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 02:38:34.679604205 +0000 UTC Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.334701 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.334763 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.334775 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.334825 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.334839 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:54Z","lastTransitionTime":"2026-02-17T13:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.438096 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.438155 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.438175 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.438204 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.438226 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:54Z","lastTransitionTime":"2026-02-17T13:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.541373 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.541434 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.541444 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.541466 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.541480 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:54Z","lastTransitionTime":"2026-02-17T13:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.644831 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.644900 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.644915 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.644953 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.644969 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:54Z","lastTransitionTime":"2026-02-17T13:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.755996 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.756066 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.756081 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.756100 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.756113 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:54Z","lastTransitionTime":"2026-02-17T13:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.860722 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.860824 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.860843 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.860872 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.860891 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:54Z","lastTransitionTime":"2026-02-17T13:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.964536 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.964607 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.964627 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.964657 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:54 crc kubenswrapper[4955]: I0217 13:04:54.964680 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:54Z","lastTransitionTime":"2026-02-17T13:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.069163 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.069235 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.069258 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.069290 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.069310 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:55Z","lastTransitionTime":"2026-02-17T13:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.173099 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.173515 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.173626 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.173807 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.173905 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:55Z","lastTransitionTime":"2026-02-17T13:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.221352 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 16:54:56.870034925 +0000 UTC Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.223172 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.223256 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.223285 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:04:55 crc kubenswrapper[4955]: E0217 13:04:55.223387 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:04:55 crc kubenswrapper[4955]: E0217 13:04:55.223660 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:04:55 crc kubenswrapper[4955]: E0217 13:04:55.223933 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.277512 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.277569 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.277590 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.277614 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.277634 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:55Z","lastTransitionTime":"2026-02-17T13:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.382132 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.382193 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.382214 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.382243 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.382268 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:55Z","lastTransitionTime":"2026-02-17T13:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.485850 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.485925 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.485942 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.485987 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.486012 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:55Z","lastTransitionTime":"2026-02-17T13:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.535349 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dcwgg_dcbc7a69-58e7-4b7a-bb84-96276819ce88/ovnkube-controller/0.log" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.540858 4955 generic.go:334] "Generic (PLEG): container finished" podID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerID="76543c6dcedac4d9e47a49af549fb14d31c1b0561a025f81788de7d80bb39e33" exitCode=1 Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.540930 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" event={"ID":"dcbc7a69-58e7-4b7a-bb84-96276819ce88","Type":"ContainerDied","Data":"76543c6dcedac4d9e47a49af549fb14d31c1b0561a025f81788de7d80bb39e33"} Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.542316 4955 scope.go:117] "RemoveContainer" containerID="76543c6dcedac4d9e47a49af549fb14d31c1b0561a025f81788de7d80bb39e33" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.565073 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9ecd4380a6ae8ff62e27e3946815170c2f02b4c3921bda595467345289c9e0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:55Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.585619 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0f4056e4b5a554f05004c194485d25231b67024e4376a73c715fcb32004a3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:55Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.590200 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.590258 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.590281 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.590312 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.590331 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:55Z","lastTransitionTime":"2026-02-17T13:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.608565 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9hx9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e57bdf8-e537-402a-9815-609fecdd68ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbc9014d80aee39922161014aac8fafa7d93be3e4092fe705c76da4a487a5e5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q22d4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9hx9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:55Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.628544 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-msn6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://277a12d62c90ef420c31d8dc71f0d7750dfd5f5ab7e90fec0dcd00e54b60245d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2tvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-msn6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:55Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.650393 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05aa39d-94e6-49d4-885e-2c1f986310f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T13:04:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 13:04:35.885968 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 13:04:35.888176 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3462920193/tls.crt::/tmp/serving-cert-3462920193/tls.key\\\\\\\"\\\\nI0217 13:04:42.370546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 13:04:42.384242 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 13:04:42.384284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 13:04:42.384317 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 13:04:42.384327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 13:04:42.394555 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 13:04:42.394614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394628 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 13:04:42.394646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 13:04:42.394654 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 13:04:42.394662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 13:04:42.394900 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 13:04:42.396494 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:55Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.672530 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:55Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.689957 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:55Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.709476 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.709552 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.709571 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.709599 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.709619 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:55Z","lastTransitionTime":"2026-02-17T13:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.747727 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bf36887-2a2b-4193-abdb-a15beaf46300\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11a8924eb6ce41d629feb2b42c7d4778a3ca9110a59c3ac25acbba4367272374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bec78f9e9f6416d6b600f6231f0d24652473b48b95bf0e0675cc85c9208d124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-29qxq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:55Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.777113 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcbc7a69-58e7-4b7a-bb84-96276819ce88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76543c6dcedac4d9e47a49af549fb14d31c1b0561a025f81788de7d80bb39e33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76543c6dcedac4d9e47a49af549fb14d31c1b0561a025f81788de7d80bb39e33\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T13:04:54Z\\\",\\\"message\\\":\\\"54.248314 6215 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0217 13:04:54.248342 6215 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0217 13:04:54.248369 6215 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0217 13:04:54.248383 6215 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0217 13:04:54.248483 6215 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0217 13:04:54.248651 6215 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0217 13:04:54.249046 6215 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0217 13:04:54.249185 6215 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0217 13:04:54.249219 6215 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0217 13:04:54.249245 6215 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0217 13:04:54.249263 6215 factory.go:656] Stopping watch factory\\\\nI0217 13:04:54.249297 6215 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dcwgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:55Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.801892 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b49c9af-148a-48ec-af37-f555f9186337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d6919c91918e8bacafd1442b06d14ffd9b547f0f37b2298e2d8d3acc5962d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e38466a067b20ca4323ab6983cb324bb63179f942d2c37c0daa8b03767e7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4830450cbada3d24c9747e9de5f48e53e118549f1dd67edf7fff8b2adc47128f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001dfb595f39d1da33b265474b6cca51c62f522ff061c9e7c52f93806034f4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f1fee7ac5f8d1442d9f63b0e8d285187129eb3a3e3b50a49a45c7a7a39ae8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:55Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.812863 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.812904 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.812914 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.812930 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.812941 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:55Z","lastTransitionTime":"2026-02-17T13:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.816518 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5c34a3d-82ad-4d03-abdd-0a20818ac2fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0261f41b82b492d324b5e50edd72cedd5e824de542a7b47104abd2ab757a7049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f43c7fe16acfd4d39301ee64665d4cd1d885397f981d70f73215da7f3aabe72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c871c1b454e4edd2f41076ecfd6a66592023d7161f094fbb4fd197fbf9c881f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:55Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.828129 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:55Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.848662 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"924c66b7-7c3c-4db7-a427-3f152efbac4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9beee007a179c3c2e04c2ce0de8ea6dbae9da3421ab24f23b718a23744c828a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fp5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:55Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.865494 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d0966253f623158053125dffca956000824438bdc548eba6dafe9a162d5674a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d07bfb361f14d817913fa2cabd30649447790068545dda2f3a2f71213ea0e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:55Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.882376 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2gsx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1face148-45d1-44b3-8775-209f9f8651fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82c5147255d021114a8639e9b60593f847896b613b5d862db588014f7485faa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnlcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2gsx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:55Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.916250 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.916507 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.917008 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.917352 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:55 crc kubenswrapper[4955]: I0217 13:04:55.917594 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:55Z","lastTransitionTime":"2026-02-17T13:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.021561 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.022098 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.022245 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.022379 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.022547 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:56Z","lastTransitionTime":"2026-02-17T13:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.126989 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.127070 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.127090 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.127120 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.127143 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:56Z","lastTransitionTime":"2026-02-17T13:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.221874 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-30 21:19:04.26321441 +0000 UTC Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.229747 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.229839 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.229858 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.229886 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.229907 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:56Z","lastTransitionTime":"2026-02-17T13:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.333182 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.333253 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.333271 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.333301 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.333323 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:56Z","lastTransitionTime":"2026-02-17T13:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.437015 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.437081 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.437097 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.437125 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.437147 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:56Z","lastTransitionTime":"2026-02-17T13:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.540847 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.540888 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.540900 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.540918 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.540930 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:56Z","lastTransitionTime":"2026-02-17T13:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.546378 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dcwgg_dcbc7a69-58e7-4b7a-bb84-96276819ce88/ovnkube-controller/0.log" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.549574 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" event={"ID":"dcbc7a69-58e7-4b7a-bb84-96276819ce88","Type":"ContainerStarted","Data":"3a7f08954bce86cfbd2f8341b30ef5ab76ba1ea7889e74ed4f8405eeb9e17fbb"} Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.549699 4955 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.580881 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcbc7a69-58e7-4b7a-bb84-96276819ce88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7f08954bce86cfbd2f8341b30ef5ab76ba1ea7889e74ed4f8405eeb9e17fbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76543c6dcedac4d9e47a49af549fb14d31c1b0561a025f81788de7d80bb39e33\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T13:04:54Z\\\",\\\"message\\\":\\\"54.248314 6215 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0217 13:04:54.248342 6215 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0217 13:04:54.248369 6215 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0217 13:04:54.248383 6215 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0217 13:04:54.248483 6215 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0217 13:04:54.248651 6215 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0217 13:04:54.249046 6215 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0217 13:04:54.249185 6215 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0217 13:04:54.249219 6215 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0217 13:04:54.249245 6215 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0217 13:04:54.249263 6215 factory.go:656] Stopping watch factory\\\\nI0217 13:04:54.249297 6215 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dcwgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:56Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.614565 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b49c9af-148a-48ec-af37-f555f9186337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d6919c91918e8bacafd1442b06d14ffd9b547f0f37b2298e2d8d3acc5962d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e38466a067b20ca4323ab6983cb324bb63179f942d2c37c0daa8b03767e7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4830450cbada3d24c9747e9de5f48e53e118549f1dd67edf7fff8b2adc47128f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001dfb595f39d1da33b265474b6cca51c62f522ff061c9e7c52f93806034f4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f1fee7ac5f8d1442d9f63b0e8d285187129eb3a3e3b50a49a45c7a7a39ae8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:56Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.630791 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5c34a3d-82ad-4d03-abdd-0a20818ac2fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0261f41b82b492d324b5e50edd72cedd5e824de542a7b47104abd2ab757a7049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f43c7fe16acfd4d39301ee64665d4cd1d885397f981d70f73215da7f3aabe72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c871c1b454e4edd2f41076ecfd6a66592023d7161f094fbb4fd197fbf9c881f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:56Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.643846 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.643911 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.643925 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.644134 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.644149 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:56Z","lastTransitionTime":"2026-02-17T13:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.654725 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:56Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.674601 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"924c66b7-7c3c-4db7-a427-3f152efbac4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9beee007a179c3c2e04c2ce0de8ea6dbae9da3421ab24f23b718a23744c828a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fp5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:56Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.688360 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d0966253f623158053125dffca956000824438bdc548eba6dafe9a162d5674a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d07bfb361f14d817913fa2cabd30649447790068545dda2f3a2f71213ea0e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:56Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.699607 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2gsx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1face148-45d1-44b3-8775-209f9f8651fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82c5147255d021114a8639e9b60593f847896b613b5d862db588014f7485faa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnlcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2gsx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:56Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.712739 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9ecd4380a6ae8ff62e27e3946815170c2f02b4c3921bda595467345289c9e0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:56Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.724473 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0f4056e4b5a554f05004c194485d25231b67024e4376a73c715fcb32004a3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:56Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.739486 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9hx9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e57bdf8-e537-402a-9815-609fecdd68ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbc9014d80aee39922161014aac8fafa7d93be3e4092fe705c76da4a487a5e5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q22d4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9hx9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:56Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.747082 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.747172 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.747186 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.747230 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.747246 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:56Z","lastTransitionTime":"2026-02-17T13:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.756034 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-msn6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://277a12d62c90ef420c31d8dc71f0d7750dfd5f5ab7e90fec0dcd00e54b60245d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2tvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-msn6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:56Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.770346 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05aa39d-94e6-49d4-885e-2c1f986310f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T13:04:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 13:04:35.885968 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 13:04:35.888176 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3462920193/tls.crt::/tmp/serving-cert-3462920193/tls.key\\\\\\\"\\\\nI0217 13:04:42.370546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 13:04:42.384242 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 13:04:42.384284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 13:04:42.384317 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 13:04:42.384327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 13:04:42.394555 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 13:04:42.394614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394628 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 13:04:42.394646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 13:04:42.394654 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 13:04:42.394662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 13:04:42.394900 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 13:04:42.396494 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:56Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.784606 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:56Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.798714 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:56Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.812502 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bf36887-2a2b-4193-abdb-a15beaf46300\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11a8924eb6ce41d629feb2b42c7d4778a3ca9110a59c3ac25acbba4367272374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bec78f9e9f6416d6b600f6231f0d24652473b48b95bf0e0675cc85c9208d124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-29qxq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:56Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.850251 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.850316 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.850328 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.850353 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.850365 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:56Z","lastTransitionTime":"2026-02-17T13:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.953569 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.953652 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.953673 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.953705 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:56 crc kubenswrapper[4955]: I0217 13:04:56.953717 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:56Z","lastTransitionTime":"2026-02-17T13:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.061389 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.061741 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.061929 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.062135 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.062308 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:57Z","lastTransitionTime":"2026-02-17T13:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.167178 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.167380 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.167464 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.167538 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.167563 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:57Z","lastTransitionTime":"2026-02-17T13:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.222375 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-04 14:00:15.21783001 +0000 UTC Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.222472 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:04:57 crc kubenswrapper[4955]: E0217 13:04:57.223317 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.222688 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:04:57 crc kubenswrapper[4955]: E0217 13:04:57.223452 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.222590 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:04:57 crc kubenswrapper[4955]: E0217 13:04:57.223538 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.271707 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.271775 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.271839 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.271878 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.271906 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:57Z","lastTransitionTime":"2026-02-17T13:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.376170 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.376247 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.376265 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.376296 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.376318 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:57Z","lastTransitionTime":"2026-02-17T13:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.480018 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.480062 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.480070 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.480087 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.480097 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:57Z","lastTransitionTime":"2026-02-17T13:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.558822 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dcwgg_dcbc7a69-58e7-4b7a-bb84-96276819ce88/ovnkube-controller/1.log" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.560432 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dcwgg_dcbc7a69-58e7-4b7a-bb84-96276819ce88/ovnkube-controller/0.log" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.565613 4955 generic.go:334] "Generic (PLEG): container finished" podID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerID="3a7f08954bce86cfbd2f8341b30ef5ab76ba1ea7889e74ed4f8405eeb9e17fbb" exitCode=1 Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.565687 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" event={"ID":"dcbc7a69-58e7-4b7a-bb84-96276819ce88","Type":"ContainerDied","Data":"3a7f08954bce86cfbd2f8341b30ef5ab76ba1ea7889e74ed4f8405eeb9e17fbb"} Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.565759 4955 scope.go:117] "RemoveContainer" containerID="76543c6dcedac4d9e47a49af549fb14d31c1b0561a025f81788de7d80bb39e33" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.569831 4955 scope.go:117] "RemoveContainer" containerID="3a7f08954bce86cfbd2f8341b30ef5ab76ba1ea7889e74ed4f8405eeb9e17fbb" Feb 17 13:04:57 crc kubenswrapper[4955]: E0217 13:04:57.570202 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dcwgg_openshift-ovn-kubernetes(dcbc7a69-58e7-4b7a-bb84-96276819ce88)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.583191 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.583264 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.583285 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.583317 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.583338 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:57Z","lastTransitionTime":"2026-02-17T13:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.592635 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d0966253f623158053125dffca956000824438bdc548eba6dafe9a162d5674a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d07bfb361f14d817913fa2cabd30649447790068545dda2f3a2f71213ea0e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:57Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.608811 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2gsx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1face148-45d1-44b3-8775-209f9f8651fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82c5147255d021114a8639e9b60593f847896b613b5d862db588014f7485faa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnlcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2gsx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:57Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.635193 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9ecd4380a6ae8ff62e27e3946815170c2f02b4c3921bda595467345289c9e0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:57Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.652796 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0f4056e4b5a554f05004c194485d25231b67024e4376a73c715fcb32004a3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:57Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.669838 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-msn6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://277a12d62c90ef420c31d8dc71f0d7750dfd5f5ab7e90fec0dcd00e54b60245d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2tvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-msn6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:57Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.687581 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.687642 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.687661 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.687687 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.687710 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:57Z","lastTransitionTime":"2026-02-17T13:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.688352 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05aa39d-94e6-49d4-885e-2c1f986310f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T13:04:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 13:04:35.885968 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 13:04:35.888176 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3462920193/tls.crt::/tmp/serving-cert-3462920193/tls.key\\\\\\\"\\\\nI0217 13:04:42.370546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 13:04:42.384242 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 13:04:42.384284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 13:04:42.384317 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 13:04:42.384327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 13:04:42.394555 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 13:04:42.394614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394628 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 13:04:42.394646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 13:04:42.394654 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 13:04:42.394662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 13:04:42.394900 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 13:04:42.396494 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:57Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.707151 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:57Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.728712 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:57Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.746351 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bf36887-2a2b-4193-abdb-a15beaf46300\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11a8924eb6ce41d629feb2b42c7d4778a3ca9110a59c3ac25acbba4367272374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bec78f9e9f6416d6b600f6231f0d24652473b48b95bf0e0675cc85c9208d124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-29qxq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:57Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.760515 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9hx9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e57bdf8-e537-402a-9815-609fecdd68ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbc9014d80aee39922161014aac8fafa7d93be3e4092fe705c76da4a487a5e5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q22d4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9hx9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:57Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.787577 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b49c9af-148a-48ec-af37-f555f9186337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d6919c91918e8bacafd1442b06d14ffd9b547f0f37b2298e2d8d3acc5962d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e38466a067b20ca4323ab6983cb324bb63179f942d2c37c0daa8b03767e7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4830450cbada3d24c9747e9de5f48e53e118549f1dd67edf7fff8b2adc47128f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001dfb595f39d1da33b265474b6cca51c62f522ff061c9e7c52f93806034f4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f1fee7ac5f8d1442d9f63b0e8d285187129eb3a3e3b50a49a45c7a7a39ae8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:57Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.790991 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.791040 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.791061 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.791097 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.791113 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:57Z","lastTransitionTime":"2026-02-17T13:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.808246 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5c34a3d-82ad-4d03-abdd-0a20818ac2fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0261f41b82b492d324b5e50edd72cedd5e824de542a7b47104abd2ab757a7049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f43c7fe16acfd4d39301ee64665d4cd1d885397f981d70f73215da7f3aabe72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c871c1b454e4edd2f41076ecfd6a66592023d7161f094fbb4fd197fbf9c881f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:57Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.827526 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:57Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.828327 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lffhw"] Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.829674 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lffhw" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.833354 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.835520 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.847475 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"924c66b7-7c3c-4db7-a427-3f152efbac4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9beee007a179c3c2e04c2ce0de8ea6dbae9da3421ab24f23b718a23744c828a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fp5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:57Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.877917 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcbc7a69-58e7-4b7a-bb84-96276819ce88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7f08954bce86cfbd2f8341b30ef5ab76ba1ea7889e74ed4f8405eeb9e17fbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76543c6dcedac4d9e47a49af549fb14d31c1b0561a025f81788de7d80bb39e33\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T13:04:54Z\\\",\\\"message\\\":\\\"54.248314 6215 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0217 13:04:54.248342 6215 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0217 13:04:54.248369 6215 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0217 13:04:54.248383 6215 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0217 13:04:54.248483 6215 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0217 13:04:54.248651 6215 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0217 13:04:54.249046 6215 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0217 13:04:54.249185 6215 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0217 13:04:54.249219 6215 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0217 13:04:54.249245 6215 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0217 13:04:54.249263 6215 factory.go:656] Stopping watch factory\\\\nI0217 13:04:54.249297 6215 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a7f08954bce86cfbd2f8341b30ef5ab76ba1ea7889e74ed4f8405eeb9e17fbb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"message\\\":\\\"ervices.LB{Name:\\\\\\\"Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"89fe421e-04e8-4967-ac75-77a0e6f784ef\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8383, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8081, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF0217 13:04:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dcwgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:57Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.885166 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gw5w\" (UniqueName: \"kubernetes.io/projected/8155b25d-be0c-44ed-b3af-7bf25d55ec83-kube-api-access-2gw5w\") pod \"ovnkube-control-plane-749d76644c-lffhw\" (UID: \"8155b25d-be0c-44ed-b3af-7bf25d55ec83\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lffhw" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.885287 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8155b25d-be0c-44ed-b3af-7bf25d55ec83-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-lffhw\" (UID: \"8155b25d-be0c-44ed-b3af-7bf25d55ec83\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lffhw" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.885332 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8155b25d-be0c-44ed-b3af-7bf25d55ec83-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-lffhw\" (UID: \"8155b25d-be0c-44ed-b3af-7bf25d55ec83\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lffhw" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.885557 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8155b25d-be0c-44ed-b3af-7bf25d55ec83-env-overrides\") pod \"ovnkube-control-plane-749d76644c-lffhw\" (UID: \"8155b25d-be0c-44ed-b3af-7bf25d55ec83\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lffhw" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.894683 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.894745 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.894764 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.894823 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.894845 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:57Z","lastTransitionTime":"2026-02-17T13:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.904093 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b49c9af-148a-48ec-af37-f555f9186337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d6919c91918e8bacafd1442b06d14ffd9b547f0f37b2298e2d8d3acc5962d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e38466a067b20ca4323ab6983cb324bb63179f942d2c37c0daa8b03767e7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4830450cbada3d24c9747e9de5f48e53e118549f1dd67edf7fff8b2adc47128f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001dfb595f39d1da33b265474b6cca51c62f522ff061c9e7c52f93806034f4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f1fee7ac5f8d1442d9f63b0e8d285187129eb3a3e3b50a49a45c7a7a39ae8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:57Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.926644 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5c34a3d-82ad-4d03-abdd-0a20818ac2fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0261f41b82b492d324b5e50edd72cedd5e824de542a7b47104abd2ab757a7049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f43c7fe16acfd4d39301ee64665d4cd1d885397f981d70f73215da7f3aabe72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c871c1b454e4edd2f41076ecfd6a66592023d7161f094fbb4fd197fbf9c881f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:57Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.947401 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:57Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.985398 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"924c66b7-7c3c-4db7-a427-3f152efbac4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9beee007a179c3c2e04c2ce0de8ea6dbae9da3421ab24f23b718a23744c828a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fp5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:57Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.986500 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8155b25d-be0c-44ed-b3af-7bf25d55ec83-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-lffhw\" (UID: \"8155b25d-be0c-44ed-b3af-7bf25d55ec83\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lffhw" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.986560 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8155b25d-be0c-44ed-b3af-7bf25d55ec83-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-lffhw\" (UID: \"8155b25d-be0c-44ed-b3af-7bf25d55ec83\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lffhw" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.986610 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8155b25d-be0c-44ed-b3af-7bf25d55ec83-env-overrides\") pod \"ovnkube-control-plane-749d76644c-lffhw\" (UID: \"8155b25d-be0c-44ed-b3af-7bf25d55ec83\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lffhw" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.986688 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gw5w\" (UniqueName: \"kubernetes.io/projected/8155b25d-be0c-44ed-b3af-7bf25d55ec83-kube-api-access-2gw5w\") pod \"ovnkube-control-plane-749d76644c-lffhw\" (UID: \"8155b25d-be0c-44ed-b3af-7bf25d55ec83\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lffhw" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.988821 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8155b25d-be0c-44ed-b3af-7bf25d55ec83-env-overrides\") pod \"ovnkube-control-plane-749d76644c-lffhw\" (UID: \"8155b25d-be0c-44ed-b3af-7bf25d55ec83\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lffhw" Feb 17 13:04:57 crc kubenswrapper[4955]: I0217 13:04:57.989067 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8155b25d-be0c-44ed-b3af-7bf25d55ec83-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-lffhw\" (UID: \"8155b25d-be0c-44ed-b3af-7bf25d55ec83\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lffhw" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.011821 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8155b25d-be0c-44ed-b3af-7bf25d55ec83-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-lffhw\" (UID: \"8155b25d-be0c-44ed-b3af-7bf25d55ec83\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lffhw" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.013011 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.013063 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.013074 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.013096 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.013108 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:58Z","lastTransitionTime":"2026-02-17T13:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.015079 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gw5w\" (UniqueName: \"kubernetes.io/projected/8155b25d-be0c-44ed-b3af-7bf25d55ec83-kube-api-access-2gw5w\") pod \"ovnkube-control-plane-749d76644c-lffhw\" (UID: \"8155b25d-be0c-44ed-b3af-7bf25d55ec83\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lffhw" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.021961 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcbc7a69-58e7-4b7a-bb84-96276819ce88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7f08954bce86cfbd2f8341b30ef5ab76ba1ea7889e74ed4f8405eeb9e17fbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76543c6dcedac4d9e47a49af549fb14d31c1b0561a025f81788de7d80bb39e33\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T13:04:54Z\\\",\\\"message\\\":\\\"54.248314 6215 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0217 13:04:54.248342 6215 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0217 13:04:54.248369 6215 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0217 13:04:54.248383 6215 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0217 13:04:54.248483 6215 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0217 13:04:54.248651 6215 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0217 13:04:54.249046 6215 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0217 13:04:54.249185 6215 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0217 13:04:54.249219 6215 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0217 13:04:54.249245 6215 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0217 13:04:54.249263 6215 factory.go:656] Stopping watch factory\\\\nI0217 13:04:54.249297 6215 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a7f08954bce86cfbd2f8341b30ef5ab76ba1ea7889e74ed4f8405eeb9e17fbb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"message\\\":\\\"ervices.LB{Name:\\\\\\\"Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"89fe421e-04e8-4967-ac75-77a0e6f784ef\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8383, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8081, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF0217 13:04:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dcwgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.040460 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d0966253f623158053125dffca956000824438bdc548eba6dafe9a162d5674a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d07bfb361f14d817913fa2cabd30649447790068545dda2f3a2f71213ea0e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.061013 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2gsx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1face148-45d1-44b3-8775-209f9f8651fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82c5147255d021114a8639e9b60593f847896b613b5d862db588014f7485faa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnlcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2gsx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.081723 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lffhw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8155b25d-be0c-44ed-b3af-7bf25d55ec83\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gw5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gw5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lffhw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.100847 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9ecd4380a6ae8ff62e27e3946815170c2f02b4c3921bda595467345289c9e0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.116768 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.116879 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.116904 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.116945 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.116971 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:58Z","lastTransitionTime":"2026-02-17T13:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.118614 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0f4056e4b5a554f05004c194485d25231b67024e4376a73c715fcb32004a3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.139095 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05aa39d-94e6-49d4-885e-2c1f986310f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T13:04:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 13:04:35.885968 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 13:04:35.888176 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3462920193/tls.crt::/tmp/serving-cert-3462920193/tls.key\\\\\\\"\\\\nI0217 13:04:42.370546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 13:04:42.384242 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 13:04:42.384284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 13:04:42.384317 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 13:04:42.384327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 13:04:42.394555 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 13:04:42.394614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394628 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 13:04:42.394646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 13:04:42.394654 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 13:04:42.394662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 13:04:42.394900 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 13:04:42.396494 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.146286 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lffhw" Feb 17 13:04:58 crc kubenswrapper[4955]: W0217 13:04:58.166413 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8155b25d_be0c_44ed_b3af_7bf25d55ec83.slice/crio-554ab46047480780409cf067d0639d318a192ba6102a63485516a4099909d5a1 WatchSource:0}: Error finding container 554ab46047480780409cf067d0639d318a192ba6102a63485516a4099909d5a1: Status 404 returned error can't find the container with id 554ab46047480780409cf067d0639d318a192ba6102a63485516a4099909d5a1 Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.167043 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.184473 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.201080 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bf36887-2a2b-4193-abdb-a15beaf46300\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11a8924eb6ce41d629feb2b42c7d4778a3ca9110a59c3ac25acbba4367272374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bec78f9e9f6416d6b600f6231f0d24652473b48b95bf0e0675cc85c9208d124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-29qxq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.221541 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.221590 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.221609 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.221636 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.221656 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:58Z","lastTransitionTime":"2026-02-17T13:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.223096 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9hx9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e57bdf8-e537-402a-9815-609fecdd68ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbc9014d80aee39922161014aac8fafa7d93be3e4092fe705c76da4a487a5e5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q22d4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9hx9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.223164 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 16:06:14.098094311 +0000 UTC Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.238940 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-msn6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://277a12d62c90ef420c31d8dc71f0d7750dfd5f5ab7e90fec0dcd00e54b60245d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2tvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-msn6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.315416 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.325086 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.325137 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.325151 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.325170 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.325183 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:58Z","lastTransitionTime":"2026-02-17T13:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.337952 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9hx9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e57bdf8-e537-402a-9815-609fecdd68ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbc9014d80aee39922161014aac8fafa7d93be3e4092fe705c76da4a487a5e5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q22d4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9hx9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.353573 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-msn6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://277a12d62c90ef420c31d8dc71f0d7750dfd5f5ab7e90fec0dcd00e54b60245d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2tvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-msn6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.374680 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05aa39d-94e6-49d4-885e-2c1f986310f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T13:04:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 13:04:35.885968 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 13:04:35.888176 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3462920193/tls.crt::/tmp/serving-cert-3462920193/tls.key\\\\\\\"\\\\nI0217 13:04:42.370546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 13:04:42.384242 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 13:04:42.384284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 13:04:42.384317 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 13:04:42.384327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 13:04:42.394555 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 13:04:42.394614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394628 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 13:04:42.394646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 13:04:42.394654 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 13:04:42.394662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 13:04:42.394900 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 13:04:42.396494 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.397057 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.418611 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.427646 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.427703 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.427720 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.427742 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.427758 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:58Z","lastTransitionTime":"2026-02-17T13:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.436530 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bf36887-2a2b-4193-abdb-a15beaf46300\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11a8924eb6ce41d629feb2b42c7d4778a3ca9110a59c3ac25acbba4367272374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bec78f9e9f6416d6b600f6231f0d24652473b48b95bf0e0675cc85c9208d124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-29qxq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.463732 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcbc7a69-58e7-4b7a-bb84-96276819ce88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7f08954bce86cfbd2f8341b30ef5ab76ba1ea7889e74ed4f8405eeb9e17fbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76543c6dcedac4d9e47a49af549fb14d31c1b0561a025f81788de7d80bb39e33\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T13:04:54Z\\\",\\\"message\\\":\\\"54.248314 6215 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0217 13:04:54.248342 6215 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0217 13:04:54.248369 6215 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0217 13:04:54.248383 6215 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0217 13:04:54.248483 6215 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0217 13:04:54.248651 6215 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0217 13:04:54.249046 6215 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0217 13:04:54.249185 6215 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0217 13:04:54.249219 6215 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0217 13:04:54.249245 6215 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0217 13:04:54.249263 6215 factory.go:656] Stopping watch factory\\\\nI0217 13:04:54.249297 6215 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a7f08954bce86cfbd2f8341b30ef5ab76ba1ea7889e74ed4f8405eeb9e17fbb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"message\\\":\\\"ervices.LB{Name:\\\\\\\"Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"89fe421e-04e8-4967-ac75-77a0e6f784ef\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8383, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8081, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF0217 13:04:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dcwgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.499867 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b49c9af-148a-48ec-af37-f555f9186337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d6919c91918e8bacafd1442b06d14ffd9b547f0f37b2298e2d8d3acc5962d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e38466a067b20ca4323ab6983cb324bb63179f942d2c37c0daa8b03767e7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4830450cbada3d24c9747e9de5f48e53e118549f1dd67edf7fff8b2adc47128f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001dfb595f39d1da33b265474b6cca51c62f522ff061c9e7c52f93806034f4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f1fee7ac5f8d1442d9f63b0e8d285187129eb3a3e3b50a49a45c7a7a39ae8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.522906 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5c34a3d-82ad-4d03-abdd-0a20818ac2fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0261f41b82b492d324b5e50edd72cedd5e824de542a7b47104abd2ab757a7049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f43c7fe16acfd4d39301ee64665d4cd1d885397f981d70f73215da7f3aabe72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c871c1b454e4edd2f41076ecfd6a66592023d7161f094fbb4fd197fbf9c881f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.531113 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.531265 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.531310 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.531344 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.531366 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:58Z","lastTransitionTime":"2026-02-17T13:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.543088 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.563632 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"924c66b7-7c3c-4db7-a427-3f152efbac4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9beee007a179c3c2e04c2ce0de8ea6dbae9da3421ab24f23b718a23744c828a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fp5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.582134 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dcwgg_dcbc7a69-58e7-4b7a-bb84-96276819ce88/ovnkube-controller/1.log" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.583993 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d0966253f623158053125dffca956000824438bdc548eba6dafe9a162d5674a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d07bfb361f14d817913fa2cabd30649447790068545dda2f3a2f71213ea0e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.589742 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-2k5pz"] Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.591012 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:04:58 crc kubenswrapper[4955]: E0217 13:04:58.591154 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.600344 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2gsx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1face148-45d1-44b3-8775-209f9f8651fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82c5147255d021114a8639e9b60593f847896b613b5d862db588014f7485faa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnlcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2gsx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.604024 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lffhw" event={"ID":"8155b25d-be0c-44ed-b3af-7bf25d55ec83","Type":"ContainerStarted","Data":"ad540421c8104f89cf02f413145e08de7c3e65e47f78cb2be5b40f746d6376c1"} Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.604090 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lffhw" event={"ID":"8155b25d-be0c-44ed-b3af-7bf25d55ec83","Type":"ContainerStarted","Data":"554ab46047480780409cf067d0639d318a192ba6102a63485516a4099909d5a1"} Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.617088 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lffhw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8155b25d-be0c-44ed-b3af-7bf25d55ec83\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gw5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gw5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lffhw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.635402 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.635454 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.635466 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.635489 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.635502 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:58Z","lastTransitionTime":"2026-02-17T13:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.636045 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9ecd4380a6ae8ff62e27e3946815170c2f02b4c3921bda595467345289c9e0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.651024 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0f4056e4b5a554f05004c194485d25231b67024e4376a73c715fcb32004a3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.667771 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5c34a3d-82ad-4d03-abdd-0a20818ac2fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0261f41b82b492d324b5e50edd72cedd5e824de542a7b47104abd2ab757a7049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f43c7fe16acfd4d39301ee64665d4cd1d885397f981d70f73215da7f3aabe72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c871c1b454e4edd2f41076ecfd6a66592023d7161f094fbb4fd197fbf9c881f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.685831 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.697306 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/77480847-3271-4161-a833-e62b3bb4413e-metrics-certs\") pod \"network-metrics-daemon-2k5pz\" (UID: \"77480847-3271-4161-a833-e62b3bb4413e\") " pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.697374 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fffmp\" (UniqueName: \"kubernetes.io/projected/77480847-3271-4161-a833-e62b3bb4413e-kube-api-access-fffmp\") pod \"network-metrics-daemon-2k5pz\" (UID: \"77480847-3271-4161-a833-e62b3bb4413e\") " pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.704397 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"924c66b7-7c3c-4db7-a427-3f152efbac4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9beee007a179c3c2e04c2ce0de8ea6dbae9da3421ab24f23b718a23744c828a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fp5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.732726 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcbc7a69-58e7-4b7a-bb84-96276819ce88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7f08954bce86cfbd2f8341b30ef5ab76ba1ea7889e74ed4f8405eeb9e17fbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76543c6dcedac4d9e47a49af549fb14d31c1b0561a025f81788de7d80bb39e33\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T13:04:54Z\\\",\\\"message\\\":\\\"54.248314 6215 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0217 13:04:54.248342 6215 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0217 13:04:54.248369 6215 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0217 13:04:54.248383 6215 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0217 13:04:54.248483 6215 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0217 13:04:54.248651 6215 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0217 13:04:54.249046 6215 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0217 13:04:54.249185 6215 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0217 13:04:54.249219 6215 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0217 13:04:54.249245 6215 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0217 13:04:54.249263 6215 factory.go:656] Stopping watch factory\\\\nI0217 13:04:54.249297 6215 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a7f08954bce86cfbd2f8341b30ef5ab76ba1ea7889e74ed4f8405eeb9e17fbb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"message\\\":\\\"ervices.LB{Name:\\\\\\\"Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"89fe421e-04e8-4967-ac75-77a0e6f784ef\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8383, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8081, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF0217 13:04:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dcwgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.738948 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.738994 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.739007 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.739030 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.739045 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:58Z","lastTransitionTime":"2026-02-17T13:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.764185 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b49c9af-148a-48ec-af37-f555f9186337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d6919c91918e8bacafd1442b06d14ffd9b547f0f37b2298e2d8d3acc5962d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e38466a067b20ca4323ab6983cb324bb63179f942d2c37c0daa8b03767e7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4830450cbada3d24c9747e9de5f48e53e118549f1dd67edf7fff8b2adc47128f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001dfb595f39d1da33b265474b6cca51c62f522ff061c9e7c52f93806034f4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f1fee7ac5f8d1442d9f63b0e8d285187129eb3a3e3b50a49a45c7a7a39ae8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.789072 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d0966253f623158053125dffca956000824438bdc548eba6dafe9a162d5674a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d07bfb361f14d817913fa2cabd30649447790068545dda2f3a2f71213ea0e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.798964 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/77480847-3271-4161-a833-e62b3bb4413e-metrics-certs\") pod \"network-metrics-daemon-2k5pz\" (UID: \"77480847-3271-4161-a833-e62b3bb4413e\") " pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.799022 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fffmp\" (UniqueName: \"kubernetes.io/projected/77480847-3271-4161-a833-e62b3bb4413e-kube-api-access-fffmp\") pod \"network-metrics-daemon-2k5pz\" (UID: \"77480847-3271-4161-a833-e62b3bb4413e\") " pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:04:58 crc kubenswrapper[4955]: E0217 13:04:58.799216 4955 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 13:04:58 crc kubenswrapper[4955]: E0217 13:04:58.799340 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/77480847-3271-4161-a833-e62b3bb4413e-metrics-certs podName:77480847-3271-4161-a833-e62b3bb4413e nodeName:}" failed. No retries permitted until 2026-02-17 13:04:59.299312837 +0000 UTC m=+37.822042380 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/77480847-3271-4161-a833-e62b3bb4413e-metrics-certs") pod "network-metrics-daemon-2k5pz" (UID: "77480847-3271-4161-a833-e62b3bb4413e") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.806188 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2gsx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1face148-45d1-44b3-8775-209f9f8651fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82c5147255d021114a8639e9b60593f847896b613b5d862db588014f7485faa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnlcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2gsx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.816256 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fffmp\" (UniqueName: \"kubernetes.io/projected/77480847-3271-4161-a833-e62b3bb4413e-kube-api-access-fffmp\") pod \"network-metrics-daemon-2k5pz\" (UID: \"77480847-3271-4161-a833-e62b3bb4413e\") " pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.822639 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lffhw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8155b25d-be0c-44ed-b3af-7bf25d55ec83\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gw5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gw5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lffhw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.836923 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2k5pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77480847-3271-4161-a833-e62b3bb4413e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fffmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fffmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2k5pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.842351 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.842416 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.842433 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.842461 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.842477 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:58Z","lastTransitionTime":"2026-02-17T13:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.854941 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9ecd4380a6ae8ff62e27e3946815170c2f02b4c3921bda595467345289c9e0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.870494 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0f4056e4b5a554f05004c194485d25231b67024e4376a73c715fcb32004a3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.883334 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.900083 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.914821 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bf36887-2a2b-4193-abdb-a15beaf46300\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11a8924eb6ce41d629feb2b42c7d4778a3ca9110a59c3ac25acbba4367272374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bec78f9e9f6416d6b600f6231f0d24652473b48b95bf0e0675cc85c9208d124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-29qxq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.933628 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9hx9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e57bdf8-e537-402a-9815-609fecdd68ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbc9014d80aee39922161014aac8fafa7d93be3e4092fe705c76da4a487a5e5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q22d4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9hx9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.945377 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.945439 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.945454 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.945482 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.945499 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:58Z","lastTransitionTime":"2026-02-17T13:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.949119 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-msn6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://277a12d62c90ef420c31d8dc71f0d7750dfd5f5ab7e90fec0dcd00e54b60245d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2tvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-msn6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:58 crc kubenswrapper[4955]: I0217 13:04:58.966125 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05aa39d-94e6-49d4-885e-2c1f986310f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T13:04:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 13:04:35.885968 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 13:04:35.888176 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3462920193/tls.crt::/tmp/serving-cert-3462920193/tls.key\\\\\\\"\\\\nI0217 13:04:42.370546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 13:04:42.384242 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 13:04:42.384284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 13:04:42.384317 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 13:04:42.384327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 13:04:42.394555 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 13:04:42.394614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394628 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 13:04:42.394646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 13:04:42.394654 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 13:04:42.394662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 13:04:42.394900 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 13:04:42.396494 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:58Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.001379 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.001650 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:04:59 crc kubenswrapper[4955]: E0217 13:04:59.001846 4955 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 17 13:04:59 crc kubenswrapper[4955]: E0217 13:04:59.001900 4955 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 17 13:04:59 crc kubenswrapper[4955]: E0217 13:04:59.001928 4955 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 13:04:59 crc kubenswrapper[4955]: E0217 13:04:59.002019 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-17 13:05:15.001989171 +0000 UTC m=+53.524718754 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 13:04:59 crc kubenswrapper[4955]: E0217 13:04:59.002069 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:05:15.002050983 +0000 UTC m=+53.524780556 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.002210 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.002313 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.002406 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:04:59 crc kubenswrapper[4955]: E0217 13:04:59.002397 4955 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 17 13:04:59 crc kubenswrapper[4955]: E0217 13:04:59.002690 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-17 13:05:15.00266859 +0000 UTC m=+53.525398133 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 17 13:04:59 crc kubenswrapper[4955]: E0217 13:04:59.002551 4955 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 17 13:04:59 crc kubenswrapper[4955]: E0217 13:04:59.002849 4955 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 17 13:04:59 crc kubenswrapper[4955]: E0217 13:04:59.002877 4955 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 13:04:59 crc kubenswrapper[4955]: E0217 13:04:59.002564 4955 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 17 13:04:59 crc kubenswrapper[4955]: E0217 13:04:59.002944 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-17 13:05:15.002919827 +0000 UTC m=+53.525649540 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 13:04:59 crc kubenswrapper[4955]: E0217 13:04:59.003098 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-17 13:05:15.003065971 +0000 UTC m=+53.525795554 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.006697 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.006750 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.006768 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.006902 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.006938 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:59Z","lastTransitionTime":"2026-02-17T13:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:59 crc kubenswrapper[4955]: E0217 13:04:59.031574 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7605a498-360b-4ce4-b7a1-dfbc354d798d\\\",\\\"systemUUID\\\":\\\"6d4602d9-c718-4393-8f91-3c381f800c9e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:59Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.036733 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.036827 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.036849 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.036879 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.036899 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:59Z","lastTransitionTime":"2026-02-17T13:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:59 crc kubenswrapper[4955]: E0217 13:04:59.057924 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7605a498-360b-4ce4-b7a1-dfbc354d798d\\\",\\\"systemUUID\\\":\\\"6d4602d9-c718-4393-8f91-3c381f800c9e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:59Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.063405 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.063438 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.063448 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.063466 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.063476 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:59Z","lastTransitionTime":"2026-02-17T13:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:59 crc kubenswrapper[4955]: E0217 13:04:59.085629 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7605a498-360b-4ce4-b7a1-dfbc354d798d\\\",\\\"systemUUID\\\":\\\"6d4602d9-c718-4393-8f91-3c381f800c9e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:59Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.091525 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.091660 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.091744 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.091846 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.091917 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:59Z","lastTransitionTime":"2026-02-17T13:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:59 crc kubenswrapper[4955]: E0217 13:04:59.110429 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7605a498-360b-4ce4-b7a1-dfbc354d798d\\\",\\\"systemUUID\\\":\\\"6d4602d9-c718-4393-8f91-3c381f800c9e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:59Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.115981 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.116042 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.116060 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.116086 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.116103 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:59Z","lastTransitionTime":"2026-02-17T13:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:59 crc kubenswrapper[4955]: E0217 13:04:59.146102 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7605a498-360b-4ce4-b7a1-dfbc354d798d\\\",\\\"systemUUID\\\":\\\"6d4602d9-c718-4393-8f91-3c381f800c9e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:59Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:59 crc kubenswrapper[4955]: E0217 13:04:59.146269 4955 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.149243 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.149318 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.149340 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.149371 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.149392 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:59Z","lastTransitionTime":"2026-02-17T13:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.222199 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.222317 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:04:59 crc kubenswrapper[4955]: E0217 13:04:59.222439 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.222317 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:04:59 crc kubenswrapper[4955]: E0217 13:04:59.222550 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:04:59 crc kubenswrapper[4955]: E0217 13:04:59.222634 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.224236 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 10:28:51.910963301 +0000 UTC Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.253178 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.253482 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.253544 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.253608 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.253665 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:59Z","lastTransitionTime":"2026-02-17T13:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.307200 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/77480847-3271-4161-a833-e62b3bb4413e-metrics-certs\") pod \"network-metrics-daemon-2k5pz\" (UID: \"77480847-3271-4161-a833-e62b3bb4413e\") " pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:04:59 crc kubenswrapper[4955]: E0217 13:04:59.307461 4955 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 13:04:59 crc kubenswrapper[4955]: E0217 13:04:59.307645 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/77480847-3271-4161-a833-e62b3bb4413e-metrics-certs podName:77480847-3271-4161-a833-e62b3bb4413e nodeName:}" failed. No retries permitted until 2026-02-17 13:05:00.307596651 +0000 UTC m=+38.830326334 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/77480847-3271-4161-a833-e62b3bb4413e-metrics-certs") pod "network-metrics-daemon-2k5pz" (UID: "77480847-3271-4161-a833-e62b3bb4413e") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.357256 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.357327 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.357347 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.357378 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.357398 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:59Z","lastTransitionTime":"2026-02-17T13:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.429231 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.430671 4955 scope.go:117] "RemoveContainer" containerID="3a7f08954bce86cfbd2f8341b30ef5ab76ba1ea7889e74ed4f8405eeb9e17fbb" Feb 17 13:04:59 crc kubenswrapper[4955]: E0217 13:04:59.431016 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dcwgg_openshift-ovn-kubernetes(dcbc7a69-58e7-4b7a-bb84-96276819ce88)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.453631 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9hx9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e57bdf8-e537-402a-9815-609fecdd68ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbc9014d80aee39922161014aac8fafa7d93be3e4092fe705c76da4a487a5e5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q22d4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9hx9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:59Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.460460 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.460556 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.460576 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.460607 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.460633 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:59Z","lastTransitionTime":"2026-02-17T13:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.472030 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-msn6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://277a12d62c90ef420c31d8dc71f0d7750dfd5f5ab7e90fec0dcd00e54b60245d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2tvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-msn6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:59Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.495763 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05aa39d-94e6-49d4-885e-2c1f986310f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T13:04:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 13:04:35.885968 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 13:04:35.888176 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3462920193/tls.crt::/tmp/serving-cert-3462920193/tls.key\\\\\\\"\\\\nI0217 13:04:42.370546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 13:04:42.384242 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 13:04:42.384284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 13:04:42.384317 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 13:04:42.384327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 13:04:42.394555 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 13:04:42.394614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394628 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 13:04:42.394646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 13:04:42.394654 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 13:04:42.394662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 13:04:42.394900 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 13:04:42.396494 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:59Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.512722 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:59Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.530436 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:59Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.544016 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bf36887-2a2b-4193-abdb-a15beaf46300\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11a8924eb6ce41d629feb2b42c7d4778a3ca9110a59c3ac25acbba4367272374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bec78f9e9f6416d6b600f6231f0d24652473b48b95bf0e0675cc85c9208d124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-29qxq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:59Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.564643 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.564692 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.564709 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.564734 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.564752 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:59Z","lastTransitionTime":"2026-02-17T13:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.565697 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcbc7a69-58e7-4b7a-bb84-96276819ce88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7f08954bce86cfbd2f8341b30ef5ab76ba1ea7889e74ed4f8405eeb9e17fbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a7f08954bce86cfbd2f8341b30ef5ab76ba1ea7889e74ed4f8405eeb9e17fbb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"message\\\":\\\"ervices.LB{Name:\\\\\\\"Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"89fe421e-04e8-4967-ac75-77a0e6f784ef\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8383, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8081, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF0217 13:04:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dcwgg_openshift-ovn-kubernetes(dcbc7a69-58e7-4b7a-bb84-96276819ce88)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dcwgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:59Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.600948 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b49c9af-148a-48ec-af37-f555f9186337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d6919c91918e8bacafd1442b06d14ffd9b547f0f37b2298e2d8d3acc5962d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e38466a067b20ca4323ab6983cb324bb63179f942d2c37c0daa8b03767e7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4830450cbada3d24c9747e9de5f48e53e118549f1dd67edf7fff8b2adc47128f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001dfb595f39d1da33b265474b6cca51c62f522ff061c9e7c52f93806034f4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f1fee7ac5f8d1442d9f63b0e8d285187129eb3a3e3b50a49a45c7a7a39ae8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:59Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.609034 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lffhw" event={"ID":"8155b25d-be0c-44ed-b3af-7bf25d55ec83","Type":"ContainerStarted","Data":"c52f6c4f214dd72b2df8ba07e3448bb4a2bedc698f47c16d027a774e2723c201"} Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.621077 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5c34a3d-82ad-4d03-abdd-0a20818ac2fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0261f41b82b492d324b5e50edd72cedd5e824de542a7b47104abd2ab757a7049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f43c7fe16acfd4d39301ee64665d4cd1d885397f981d70f73215da7f3aabe72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c871c1b454e4edd2f41076ecfd6a66592023d7161f094fbb4fd197fbf9c881f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:59Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.640359 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:59Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.664005 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"924c66b7-7c3c-4db7-a427-3f152efbac4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9beee007a179c3c2e04c2ce0de8ea6dbae9da3421ab24f23b718a23744c828a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fp5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:59Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.668928 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.668996 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.669013 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.669037 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.669053 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:59Z","lastTransitionTime":"2026-02-17T13:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.681425 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2k5pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77480847-3271-4161-a833-e62b3bb4413e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fffmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fffmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2k5pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:59Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.699317 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d0966253f623158053125dffca956000824438bdc548eba6dafe9a162d5674a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d07bfb361f14d817913fa2cabd30649447790068545dda2f3a2f71213ea0e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:59Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.713475 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2gsx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1face148-45d1-44b3-8775-209f9f8651fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82c5147255d021114a8639e9b60593f847896b613b5d862db588014f7485faa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnlcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2gsx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:59Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.728899 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lffhw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8155b25d-be0c-44ed-b3af-7bf25d55ec83\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gw5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gw5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lffhw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:59Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.749042 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9ecd4380a6ae8ff62e27e3946815170c2f02b4c3921bda595467345289c9e0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:59Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.772340 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0f4056e4b5a554f05004c194485d25231b67024e4376a73c715fcb32004a3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:59Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.773098 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.773192 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.773221 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.773259 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.773288 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:59Z","lastTransitionTime":"2026-02-17T13:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.796748 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b49c9af-148a-48ec-af37-f555f9186337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d6919c91918e8bacafd1442b06d14ffd9b547f0f37b2298e2d8d3acc5962d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e38466a067b20ca4323ab6983cb324bb63179f942d2c37c0daa8b03767e7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4830450cbada3d24c9747e9de5f48e53e118549f1dd67edf7fff8b2adc47128f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001dfb595f39d1da33b265474b6cca51c62f522ff061c9e7c52f93806034f4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f1fee7ac5f8d1442d9f63b0e8d285187129eb3a3e3b50a49a45c7a7a39ae8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:59Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.817508 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5c34a3d-82ad-4d03-abdd-0a20818ac2fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0261f41b82b492d324b5e50edd72cedd5e824de542a7b47104abd2ab757a7049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f43c7fe16acfd4d39301ee64665d4cd1d885397f981d70f73215da7f3aabe72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c871c1b454e4edd2f41076ecfd6a66592023d7161f094fbb4fd197fbf9c881f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:59Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.835623 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:59Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.859163 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"924c66b7-7c3c-4db7-a427-3f152efbac4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9beee007a179c3c2e04c2ce0de8ea6dbae9da3421ab24f23b718a23744c828a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fp5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:59Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.877320 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.877407 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.877435 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.877473 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.877500 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:59Z","lastTransitionTime":"2026-02-17T13:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.892602 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcbc7a69-58e7-4b7a-bb84-96276819ce88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7f08954bce86cfbd2f8341b30ef5ab76ba1ea7889e74ed4f8405eeb9e17fbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a7f08954bce86cfbd2f8341b30ef5ab76ba1ea7889e74ed4f8405eeb9e17fbb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"message\\\":\\\"ervices.LB{Name:\\\\\\\"Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"89fe421e-04e8-4967-ac75-77a0e6f784ef\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8383, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8081, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF0217 13:04:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dcwgg_openshift-ovn-kubernetes(dcbc7a69-58e7-4b7a-bb84-96276819ce88)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dcwgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:59Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.915357 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d0966253f623158053125dffca956000824438bdc548eba6dafe9a162d5674a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d07bfb361f14d817913fa2cabd30649447790068545dda2f3a2f71213ea0e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:59Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.936284 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2gsx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1face148-45d1-44b3-8775-209f9f8651fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82c5147255d021114a8639e9b60593f847896b613b5d862db588014f7485faa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnlcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2gsx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:59Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.949716 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lffhw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8155b25d-be0c-44ed-b3af-7bf25d55ec83\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad540421c8104f89cf02f413145e08de7c3e65e47f78cb2be5b40f746d6376c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gw5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52f6c4f214dd72b2df8ba07e3448bb4a2bedc698f47c16d027a774e2723c201\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gw5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lffhw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:59Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.965709 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2k5pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77480847-3271-4161-a833-e62b3bb4413e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fffmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fffmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2k5pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:59Z is after 2025-08-24T17:21:41Z" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.981309 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.981365 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.981380 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.981400 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.981413 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:04:59Z","lastTransitionTime":"2026-02-17T13:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:04:59 crc kubenswrapper[4955]: I0217 13:04:59.989522 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9ecd4380a6ae8ff62e27e3946815170c2f02b4c3921bda595467345289c9e0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:04:59Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.010543 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0f4056e4b5a554f05004c194485d25231b67024e4376a73c715fcb32004a3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:00Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.030663 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05aa39d-94e6-49d4-885e-2c1f986310f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T13:04:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 13:04:35.885968 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 13:04:35.888176 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3462920193/tls.crt::/tmp/serving-cert-3462920193/tls.key\\\\\\\"\\\\nI0217 13:04:42.370546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 13:04:42.384242 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 13:04:42.384284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 13:04:42.384317 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 13:04:42.384327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 13:04:42.394555 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 13:04:42.394614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394628 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 13:04:42.394646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 13:04:42.394654 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 13:04:42.394662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 13:04:42.394900 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 13:04:42.396494 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:00Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.050014 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:00Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.072874 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:00Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.086027 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.086142 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.086156 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.086204 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.086220 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:00Z","lastTransitionTime":"2026-02-17T13:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.089511 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bf36887-2a2b-4193-abdb-a15beaf46300\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11a8924eb6ce41d629feb2b42c7d4778a3ca9110a59c3ac25acbba4367272374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bec78f9e9f6416d6b600f6231f0d24652473b48b95bf0e0675cc85c9208d124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-29qxq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:00Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.108086 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9hx9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e57bdf8-e537-402a-9815-609fecdd68ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbc9014d80aee39922161014aac8fafa7d93be3e4092fe705c76da4a487a5e5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q22d4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9hx9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:00Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.123864 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-msn6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://277a12d62c90ef420c31d8dc71f0d7750dfd5f5ab7e90fec0dcd00e54b60245d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2tvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-msn6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:00Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.189992 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.190045 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.190062 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.190091 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.190110 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:00Z","lastTransitionTime":"2026-02-17T13:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.223084 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:05:00 crc kubenswrapper[4955]: E0217 13:05:00.223259 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.224993 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 21:41:59.173292884 +0000 UTC Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.293353 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.293399 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.293410 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.293426 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.293438 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:00Z","lastTransitionTime":"2026-02-17T13:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.320625 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/77480847-3271-4161-a833-e62b3bb4413e-metrics-certs\") pod \"network-metrics-daemon-2k5pz\" (UID: \"77480847-3271-4161-a833-e62b3bb4413e\") " pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:05:00 crc kubenswrapper[4955]: E0217 13:05:00.320870 4955 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 13:05:00 crc kubenswrapper[4955]: E0217 13:05:00.320951 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/77480847-3271-4161-a833-e62b3bb4413e-metrics-certs podName:77480847-3271-4161-a833-e62b3bb4413e nodeName:}" failed. No retries permitted until 2026-02-17 13:05:02.320932931 +0000 UTC m=+40.843662474 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/77480847-3271-4161-a833-e62b3bb4413e-metrics-certs") pod "network-metrics-daemon-2k5pz" (UID: "77480847-3271-4161-a833-e62b3bb4413e") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.396914 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.396977 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.397026 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.397051 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.397064 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:00Z","lastTransitionTime":"2026-02-17T13:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.500461 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.500530 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.500548 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.500577 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.500597 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:00Z","lastTransitionTime":"2026-02-17T13:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.604093 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.604175 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.604193 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.604222 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.604245 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:00Z","lastTransitionTime":"2026-02-17T13:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.707068 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.707144 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.707162 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.707188 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.707206 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:00Z","lastTransitionTime":"2026-02-17T13:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.809995 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.810049 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.810068 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.810095 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.810116 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:00Z","lastTransitionTime":"2026-02-17T13:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.915141 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.915242 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.915270 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.915309 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:00 crc kubenswrapper[4955]: I0217 13:05:00.915334 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:00Z","lastTransitionTime":"2026-02-17T13:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.020423 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.020495 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.020516 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.020550 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.020573 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:01Z","lastTransitionTime":"2026-02-17T13:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.124232 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.124288 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.124307 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.124343 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.124360 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:01Z","lastTransitionTime":"2026-02-17T13:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.222138 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.222197 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.222325 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:05:01 crc kubenswrapper[4955]: E0217 13:05:01.222431 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:05:01 crc kubenswrapper[4955]: E0217 13:05:01.222602 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:05:01 crc kubenswrapper[4955]: E0217 13:05:01.222775 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.225375 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 20:20:17.832726043 +0000 UTC Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.227902 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.227969 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.227992 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.228022 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.228049 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:01Z","lastTransitionTime":"2026-02-17T13:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.332146 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.332216 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.332238 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.332271 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.332292 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:01Z","lastTransitionTime":"2026-02-17T13:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.436026 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.436099 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.436119 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.436149 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.436172 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:01Z","lastTransitionTime":"2026-02-17T13:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.539733 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.539810 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.539830 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.539857 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.539880 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:01Z","lastTransitionTime":"2026-02-17T13:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.642670 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.642747 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.642771 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.642836 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.642860 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:01Z","lastTransitionTime":"2026-02-17T13:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.745751 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.745855 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.745884 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.745910 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.745930 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:01Z","lastTransitionTime":"2026-02-17T13:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.849153 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.849221 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.849240 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.849268 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.849294 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:01Z","lastTransitionTime":"2026-02-17T13:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.951410 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.951448 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.951459 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.951473 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:01 crc kubenswrapper[4955]: I0217 13:05:01.951484 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:01Z","lastTransitionTime":"2026-02-17T13:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.055099 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.055176 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.055201 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.055233 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.055349 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:02Z","lastTransitionTime":"2026-02-17T13:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.158741 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.158836 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.158856 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.158892 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.158912 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:02Z","lastTransitionTime":"2026-02-17T13:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.222691 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:05:02 crc kubenswrapper[4955]: E0217 13:05:02.223044 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.225478 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 17:07:46.137368258 +0000 UTC Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.255052 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9ecd4380a6ae8ff62e27e3946815170c2f02b4c3921bda595467345289c9e0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:02Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.262509 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.262548 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.262558 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.262578 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.262592 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:02Z","lastTransitionTime":"2026-02-17T13:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.276172 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0f4056e4b5a554f05004c194485d25231b67024e4376a73c715fcb32004a3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:02Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.293981 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-msn6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://277a12d62c90ef420c31d8dc71f0d7750dfd5f5ab7e90fec0dcd00e54b60245d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2tvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-msn6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:02Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.317353 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05aa39d-94e6-49d4-885e-2c1f986310f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T13:04:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 13:04:35.885968 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 13:04:35.888176 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3462920193/tls.crt::/tmp/serving-cert-3462920193/tls.key\\\\\\\"\\\\nI0217 13:04:42.370546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 13:04:42.384242 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 13:04:42.384284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 13:04:42.384317 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 13:04:42.384327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 13:04:42.394555 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 13:04:42.394614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394628 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 13:04:42.394646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 13:04:42.394654 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 13:04:42.394662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 13:04:42.394900 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 13:04:42.396494 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:02Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.346623 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/77480847-3271-4161-a833-e62b3bb4413e-metrics-certs\") pod \"network-metrics-daemon-2k5pz\" (UID: \"77480847-3271-4161-a833-e62b3bb4413e\") " pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:05:02 crc kubenswrapper[4955]: E0217 13:05:02.346974 4955 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 13:05:02 crc kubenswrapper[4955]: E0217 13:05:02.347164 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/77480847-3271-4161-a833-e62b3bb4413e-metrics-certs podName:77480847-3271-4161-a833-e62b3bb4413e nodeName:}" failed. No retries permitted until 2026-02-17 13:05:06.347124856 +0000 UTC m=+44.869854429 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/77480847-3271-4161-a833-e62b3bb4413e-metrics-certs") pod "network-metrics-daemon-2k5pz" (UID: "77480847-3271-4161-a833-e62b3bb4413e") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.347334 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:02Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.365720 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.365842 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.365868 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.365904 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.365929 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:02Z","lastTransitionTime":"2026-02-17T13:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.380663 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:02Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.420887 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bf36887-2a2b-4193-abdb-a15beaf46300\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11a8924eb6ce41d629feb2b42c7d4778a3ca9110a59c3ac25acbba4367272374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bec78f9e9f6416d6b600f6231f0d24652473b48b95bf0e0675cc85c9208d124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-29qxq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:02Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.441314 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9hx9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e57bdf8-e537-402a-9815-609fecdd68ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbc9014d80aee39922161014aac8fafa7d93be3e4092fe705c76da4a487a5e5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q22d4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9hx9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:02Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.462801 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b49c9af-148a-48ec-af37-f555f9186337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d6919c91918e8bacafd1442b06d14ffd9b547f0f37b2298e2d8d3acc5962d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e38466a067b20ca4323ab6983cb324bb63179f942d2c37c0daa8b03767e7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4830450cbada3d24c9747e9de5f48e53e118549f1dd67edf7fff8b2adc47128f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001dfb595f39d1da33b265474b6cca51c62f522ff061c9e7c52f93806034f4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f1fee7ac5f8d1442d9f63b0e8d285187129eb3a3e3b50a49a45c7a7a39ae8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:02Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.468838 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.468871 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.468883 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.468903 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.468921 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:02Z","lastTransitionTime":"2026-02-17T13:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.482631 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5c34a3d-82ad-4d03-abdd-0a20818ac2fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0261f41b82b492d324b5e50edd72cedd5e824de542a7b47104abd2ab757a7049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f43c7fe16acfd4d39301ee64665d4cd1d885397f981d70f73215da7f3aabe72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c871c1b454e4edd2f41076ecfd6a66592023d7161f094fbb4fd197fbf9c881f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:02Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.497214 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:02Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.520899 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"924c66b7-7c3c-4db7-a427-3f152efbac4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9beee007a179c3c2e04c2ce0de8ea6dbae9da3421ab24f23b718a23744c828a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fp5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:02Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.554282 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcbc7a69-58e7-4b7a-bb84-96276819ce88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7f08954bce86cfbd2f8341b30ef5ab76ba1ea7889e74ed4f8405eeb9e17fbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a7f08954bce86cfbd2f8341b30ef5ab76ba1ea7889e74ed4f8405eeb9e17fbb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"message\\\":\\\"ervices.LB{Name:\\\\\\\"Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"89fe421e-04e8-4967-ac75-77a0e6f784ef\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8383, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8081, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF0217 13:04:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dcwgg_openshift-ovn-kubernetes(dcbc7a69-58e7-4b7a-bb84-96276819ce88)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dcwgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:02Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.572533 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.572592 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.572610 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.572636 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.572662 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:02Z","lastTransitionTime":"2026-02-17T13:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.582851 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d0966253f623158053125dffca956000824438bdc548eba6dafe9a162d5674a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d07bfb361f14d817913fa2cabd30649447790068545dda2f3a2f71213ea0e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:02Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.600574 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2gsx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1face148-45d1-44b3-8775-209f9f8651fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82c5147255d021114a8639e9b60593f847896b613b5d862db588014f7485faa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnlcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2gsx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:02Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.621268 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lffhw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8155b25d-be0c-44ed-b3af-7bf25d55ec83\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad540421c8104f89cf02f413145e08de7c3e65e47f78cb2be5b40f746d6376c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gw5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52f6c4f214dd72b2df8ba07e3448bb4a2bedc698f47c16d027a774e2723c201\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gw5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lffhw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:02Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.642271 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2k5pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77480847-3271-4161-a833-e62b3bb4413e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fffmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fffmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2k5pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:02Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.676973 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.677030 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.677050 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.677077 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.677097 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:02Z","lastTransitionTime":"2026-02-17T13:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.780305 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.780377 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.780396 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.780426 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.780447 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:02Z","lastTransitionTime":"2026-02-17T13:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.883629 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.883699 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.883716 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.883743 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.883763 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:02Z","lastTransitionTime":"2026-02-17T13:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.987005 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.987087 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.987105 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.987133 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:02 crc kubenswrapper[4955]: I0217 13:05:02.987152 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:02Z","lastTransitionTime":"2026-02-17T13:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.090108 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.090167 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.090185 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.090215 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.090235 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:03Z","lastTransitionTime":"2026-02-17T13:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.192831 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.192893 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.192920 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.192950 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.192967 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:03Z","lastTransitionTime":"2026-02-17T13:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.222471 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:05:03 crc kubenswrapper[4955]: E0217 13:05:03.222650 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.223003 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.223043 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:05:03 crc kubenswrapper[4955]: E0217 13:05:03.224096 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:05:03 crc kubenswrapper[4955]: E0217 13:05:03.224257 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.226880 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 12:25:02.816215692 +0000 UTC Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.296088 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.296138 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.296154 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.296182 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.296203 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:03Z","lastTransitionTime":"2026-02-17T13:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.399881 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.399942 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.399958 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.399985 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.400003 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:03Z","lastTransitionTime":"2026-02-17T13:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.503815 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.503889 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.503912 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.503943 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.503967 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:03Z","lastTransitionTime":"2026-02-17T13:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.607304 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.607358 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.607375 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.607401 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.607420 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:03Z","lastTransitionTime":"2026-02-17T13:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.710747 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.710850 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.710942 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.710976 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.710999 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:03Z","lastTransitionTime":"2026-02-17T13:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.815357 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.815703 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.815816 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.815853 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.815876 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:03Z","lastTransitionTime":"2026-02-17T13:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.919770 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.919853 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.919870 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.919898 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:03 crc kubenswrapper[4955]: I0217 13:05:03.919915 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:03Z","lastTransitionTime":"2026-02-17T13:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.023990 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.024065 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.024094 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.024125 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.024149 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:04Z","lastTransitionTime":"2026-02-17T13:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.128810 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.128875 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.128893 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.128922 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.128941 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:04Z","lastTransitionTime":"2026-02-17T13:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.222406 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:05:04 crc kubenswrapper[4955]: E0217 13:05:04.222818 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.227033 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 15:12:38.544089696 +0000 UTC Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.232634 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.232700 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.232718 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.232750 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.232771 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:04Z","lastTransitionTime":"2026-02-17T13:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.336589 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.336657 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.336681 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.336713 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.336736 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:04Z","lastTransitionTime":"2026-02-17T13:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.439606 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.439699 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.439716 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.439743 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.439762 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:04Z","lastTransitionTime":"2026-02-17T13:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.543138 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.543520 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.543715 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.544001 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.544297 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:04Z","lastTransitionTime":"2026-02-17T13:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.647292 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.647356 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.647373 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.647399 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.647418 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:04Z","lastTransitionTime":"2026-02-17T13:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.751116 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.751207 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.751230 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.751263 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.751286 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:04Z","lastTransitionTime":"2026-02-17T13:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.854860 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.854930 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.854949 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.854976 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.854994 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:04Z","lastTransitionTime":"2026-02-17T13:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.958996 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.959076 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.959094 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.959122 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:04 crc kubenswrapper[4955]: I0217 13:05:04.959142 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:04Z","lastTransitionTime":"2026-02-17T13:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.062714 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.062840 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.062868 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.062906 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.062933 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:05Z","lastTransitionTime":"2026-02-17T13:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.166933 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.167008 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.167027 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.167074 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.167100 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:05Z","lastTransitionTime":"2026-02-17T13:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.222492 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.222546 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.222701 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:05:05 crc kubenswrapper[4955]: E0217 13:05:05.222858 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:05:05 crc kubenswrapper[4955]: E0217 13:05:05.223032 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:05:05 crc kubenswrapper[4955]: E0217 13:05:05.223229 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.227906 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 00:11:52.667838695 +0000 UTC Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.270719 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.270817 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.270842 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.270875 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.270898 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:05Z","lastTransitionTime":"2026-02-17T13:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.374230 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.374301 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.374357 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.374386 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.374427 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:05Z","lastTransitionTime":"2026-02-17T13:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.477729 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.477820 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.477840 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.477869 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.477887 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:05Z","lastTransitionTime":"2026-02-17T13:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.581464 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.581526 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.581546 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.581572 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.581592 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:05Z","lastTransitionTime":"2026-02-17T13:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.684553 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.684618 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.684637 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.684666 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.684684 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:05Z","lastTransitionTime":"2026-02-17T13:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.803877 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.803938 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.803959 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.803986 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.804005 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:05Z","lastTransitionTime":"2026-02-17T13:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.907409 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.907460 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.907470 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.907487 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:05 crc kubenswrapper[4955]: I0217 13:05:05.907499 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:05Z","lastTransitionTime":"2026-02-17T13:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.010905 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.010988 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.011013 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.011045 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.011071 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:06Z","lastTransitionTime":"2026-02-17T13:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.115287 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.115381 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.115408 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.115444 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.115467 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:06Z","lastTransitionTime":"2026-02-17T13:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.218916 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.218985 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.219007 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.219035 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.219058 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:06Z","lastTransitionTime":"2026-02-17T13:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.222404 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:05:06 crc kubenswrapper[4955]: E0217 13:05:06.222649 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.228110 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 18:59:01.715612744 +0000 UTC Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.323101 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.323532 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.323669 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.323906 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.324073 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:06Z","lastTransitionTime":"2026-02-17T13:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.397269 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/77480847-3271-4161-a833-e62b3bb4413e-metrics-certs\") pod \"network-metrics-daemon-2k5pz\" (UID: \"77480847-3271-4161-a833-e62b3bb4413e\") " pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:05:06 crc kubenswrapper[4955]: E0217 13:05:06.397503 4955 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 13:05:06 crc kubenswrapper[4955]: E0217 13:05:06.397768 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/77480847-3271-4161-a833-e62b3bb4413e-metrics-certs podName:77480847-3271-4161-a833-e62b3bb4413e nodeName:}" failed. No retries permitted until 2026-02-17 13:05:14.397739125 +0000 UTC m=+52.920468678 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/77480847-3271-4161-a833-e62b3bb4413e-metrics-certs") pod "network-metrics-daemon-2k5pz" (UID: "77480847-3271-4161-a833-e62b3bb4413e") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.427821 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.427898 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.427970 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.428002 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.428022 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:06Z","lastTransitionTime":"2026-02-17T13:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.531708 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.531773 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.531829 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.531862 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.531887 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:06Z","lastTransitionTime":"2026-02-17T13:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.635656 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.635727 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.635751 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.635820 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.635847 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:06Z","lastTransitionTime":"2026-02-17T13:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.740105 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.740199 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.740218 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.740247 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.740269 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:06Z","lastTransitionTime":"2026-02-17T13:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.844251 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.844341 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.844373 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.844408 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.844430 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:06Z","lastTransitionTime":"2026-02-17T13:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.948020 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.948091 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.948110 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.948140 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:06 crc kubenswrapper[4955]: I0217 13:05:06.948165 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:06Z","lastTransitionTime":"2026-02-17T13:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.050748 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.050879 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.050897 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.050926 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.050945 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:07Z","lastTransitionTime":"2026-02-17T13:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.154575 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.154619 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.154631 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.154653 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.154667 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:07Z","lastTransitionTime":"2026-02-17T13:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.222616 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:05:07 crc kubenswrapper[4955]: E0217 13:05:07.222827 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.223104 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.223168 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:05:07 crc kubenswrapper[4955]: E0217 13:05:07.223284 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:05:07 crc kubenswrapper[4955]: E0217 13:05:07.223354 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.229291 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 12:59:41.442517653 +0000 UTC Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.258158 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.258221 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.258233 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.258254 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.258268 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:07Z","lastTransitionTime":"2026-02-17T13:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.361508 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.361581 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.361600 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.361628 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.361648 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:07Z","lastTransitionTime":"2026-02-17T13:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.465105 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.465176 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.465191 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.465217 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.465233 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:07Z","lastTransitionTime":"2026-02-17T13:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.568500 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.568559 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.568576 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.568600 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.568617 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:07Z","lastTransitionTime":"2026-02-17T13:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.672387 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.672444 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.672457 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.672483 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.672498 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:07Z","lastTransitionTime":"2026-02-17T13:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.775401 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.775450 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.775462 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.775480 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.775493 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:07Z","lastTransitionTime":"2026-02-17T13:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.885105 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.885177 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.885197 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.885223 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.885240 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:07Z","lastTransitionTime":"2026-02-17T13:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.988469 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.988541 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.988565 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.988592 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:07 crc kubenswrapper[4955]: I0217 13:05:07.988611 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:07Z","lastTransitionTime":"2026-02-17T13:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.093072 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.093126 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.093135 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.093153 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.093167 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:08Z","lastTransitionTime":"2026-02-17T13:05:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.196153 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.196198 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.196208 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.196225 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.196235 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:08Z","lastTransitionTime":"2026-02-17T13:05:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.222160 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:05:08 crc kubenswrapper[4955]: E0217 13:05:08.222328 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.230467 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 01:19:50.432096414 +0000 UTC Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.298884 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.298983 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.299003 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.299031 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.299049 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:08Z","lastTransitionTime":"2026-02-17T13:05:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.402902 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.402984 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.403009 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.403038 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.403059 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:08Z","lastTransitionTime":"2026-02-17T13:05:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.506350 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.506410 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.506425 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.506448 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.506467 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:08Z","lastTransitionTime":"2026-02-17T13:05:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.610010 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.610074 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.610099 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.610128 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.610148 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:08Z","lastTransitionTime":"2026-02-17T13:05:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.714167 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.714237 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.714265 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.714293 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.714328 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:08Z","lastTransitionTime":"2026-02-17T13:05:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.817876 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.817946 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.817970 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.818003 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.818031 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:08Z","lastTransitionTime":"2026-02-17T13:05:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.922364 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.922438 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.922628 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.922659 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:08 crc kubenswrapper[4955]: I0217 13:05:08.922681 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:08Z","lastTransitionTime":"2026-02-17T13:05:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.026845 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.026915 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.026937 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.026971 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.026994 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:09Z","lastTransitionTime":"2026-02-17T13:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.130192 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.130246 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.130264 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.130289 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.130308 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:09Z","lastTransitionTime":"2026-02-17T13:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.197012 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.210276 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.212240 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d0966253f623158053125dffca956000824438bdc548eba6dafe9a162d5674a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d07bfb361f14d817913fa2cabd30649447790068545dda2f3a2f71213ea0e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:09Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.222078 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.222108 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.222114 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:05:09 crc kubenswrapper[4955]: E0217 13:05:09.222225 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:05:09 crc kubenswrapper[4955]: E0217 13:05:09.222303 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:05:09 crc kubenswrapper[4955]: E0217 13:05:09.222380 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.226153 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2gsx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1face148-45d1-44b3-8775-209f9f8651fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82c5147255d021114a8639e9b60593f847896b613b5d862db588014f7485faa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnlcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2gsx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:09Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.231292 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 14:59:30.806759443 +0000 UTC Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.233280 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.233338 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.233356 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.233379 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.233396 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:09Z","lastTransitionTime":"2026-02-17T13:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.239196 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lffhw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8155b25d-be0c-44ed-b3af-7bf25d55ec83\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad540421c8104f89cf02f413145e08de7c3e65e47f78cb2be5b40f746d6376c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gw5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52f6c4f214dd72b2df8ba07e3448bb4a2bedc698f47c16d027a774e2723c201\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gw5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lffhw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:09Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.250662 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2k5pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77480847-3271-4161-a833-e62b3bb4413e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fffmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fffmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2k5pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:09Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.269143 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9ecd4380a6ae8ff62e27e3946815170c2f02b4c3921bda595467345289c9e0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:09Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.287322 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0f4056e4b5a554f05004c194485d25231b67024e4376a73c715fcb32004a3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:09Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.300980 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-msn6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://277a12d62c90ef420c31d8dc71f0d7750dfd5f5ab7e90fec0dcd00e54b60245d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2tvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-msn6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:09Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.317904 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05aa39d-94e6-49d4-885e-2c1f986310f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T13:04:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 13:04:35.885968 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 13:04:35.888176 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3462920193/tls.crt::/tmp/serving-cert-3462920193/tls.key\\\\\\\"\\\\nI0217 13:04:42.370546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 13:04:42.384242 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 13:04:42.384284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 13:04:42.384317 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 13:04:42.384327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 13:04:42.394555 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 13:04:42.394614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394628 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 13:04:42.394646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 13:04:42.394654 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 13:04:42.394662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 13:04:42.394900 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 13:04:42.396494 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:09Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.336166 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:09Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.337203 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.337285 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.337300 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.337323 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.337362 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:09Z","lastTransitionTime":"2026-02-17T13:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.351858 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:09Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.368525 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bf36887-2a2b-4193-abdb-a15beaf46300\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11a8924eb6ce41d629feb2b42c7d4778a3ca9110a59c3ac25acbba4367272374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bec78f9e9f6416d6b600f6231f0d24652473b48b95bf0e0675cc85c9208d124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-29qxq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:09Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.381884 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9hx9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e57bdf8-e537-402a-9815-609fecdd68ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbc9014d80aee39922161014aac8fafa7d93be3e4092fe705c76da4a487a5e5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q22d4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9hx9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:09Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.405670 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b49c9af-148a-48ec-af37-f555f9186337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d6919c91918e8bacafd1442b06d14ffd9b547f0f37b2298e2d8d3acc5962d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e38466a067b20ca4323ab6983cb324bb63179f942d2c37c0daa8b03767e7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4830450cbada3d24c9747e9de5f48e53e118549f1dd67edf7fff8b2adc47128f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001dfb595f39d1da33b265474b6cca51c62f522ff061c9e7c52f93806034f4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f1fee7ac5f8d1442d9f63b0e8d285187129eb3a3e3b50a49a45c7a7a39ae8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:09Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.424927 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5c34a3d-82ad-4d03-abdd-0a20818ac2fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0261f41b82b492d324b5e50edd72cedd5e824de542a7b47104abd2ab757a7049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f43c7fe16acfd4d39301ee64665d4cd1d885397f981d70f73215da7f3aabe72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c871c1b454e4edd2f41076ecfd6a66592023d7161f094fbb4fd197fbf9c881f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:09Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.440324 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.440389 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.440408 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.440432 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.440448 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:09Z","lastTransitionTime":"2026-02-17T13:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.445331 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:09Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.463681 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"924c66b7-7c3c-4db7-a427-3f152efbac4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9beee007a179c3c2e04c2ce0de8ea6dbae9da3421ab24f23b718a23744c828a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fp5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:09Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.485295 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcbc7a69-58e7-4b7a-bb84-96276819ce88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7f08954bce86cfbd2f8341b30ef5ab76ba1ea7889e74ed4f8405eeb9e17fbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a7f08954bce86cfbd2f8341b30ef5ab76ba1ea7889e74ed4f8405eeb9e17fbb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"message\\\":\\\"ervices.LB{Name:\\\\\\\"Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"89fe421e-04e8-4967-ac75-77a0e6f784ef\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8383, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8081, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF0217 13:04:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dcwgg_openshift-ovn-kubernetes(dcbc7a69-58e7-4b7a-bb84-96276819ce88)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dcwgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:09Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.530533 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.530594 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.530613 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.530639 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.530657 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:09Z","lastTransitionTime":"2026-02-17T13:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:09 crc kubenswrapper[4955]: E0217 13:05:09.550222 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7605a498-360b-4ce4-b7a1-dfbc354d798d\\\",\\\"systemUUID\\\":\\\"6d4602d9-c718-4393-8f91-3c381f800c9e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:09Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.555206 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.555263 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.555279 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.555302 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.555322 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:09Z","lastTransitionTime":"2026-02-17T13:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:09 crc kubenswrapper[4955]: E0217 13:05:09.573339 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7605a498-360b-4ce4-b7a1-dfbc354d798d\\\",\\\"systemUUID\\\":\\\"6d4602d9-c718-4393-8f91-3c381f800c9e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:09Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.577638 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.577687 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.577705 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.577728 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.577746 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:09Z","lastTransitionTime":"2026-02-17T13:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:09 crc kubenswrapper[4955]: E0217 13:05:09.596866 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7605a498-360b-4ce4-b7a1-dfbc354d798d\\\",\\\"systemUUID\\\":\\\"6d4602d9-c718-4393-8f91-3c381f800c9e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:09Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.602037 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.602089 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.602105 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.602131 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.602148 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:09Z","lastTransitionTime":"2026-02-17T13:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:09 crc kubenswrapper[4955]: E0217 13:05:09.616510 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7605a498-360b-4ce4-b7a1-dfbc354d798d\\\",\\\"systemUUID\\\":\\\"6d4602d9-c718-4393-8f91-3c381f800c9e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:09Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.622506 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.622550 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.622565 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.622586 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.622599 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:09Z","lastTransitionTime":"2026-02-17T13:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:09 crc kubenswrapper[4955]: E0217 13:05:09.637233 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7605a498-360b-4ce4-b7a1-dfbc354d798d\\\",\\\"systemUUID\\\":\\\"6d4602d9-c718-4393-8f91-3c381f800c9e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:09Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:09 crc kubenswrapper[4955]: E0217 13:05:09.637463 4955 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.640352 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.640420 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.640445 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.640475 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.640495 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:09Z","lastTransitionTime":"2026-02-17T13:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.745498 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.745572 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.745587 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.745610 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.745632 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:09Z","lastTransitionTime":"2026-02-17T13:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.848424 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.848496 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.848519 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.848551 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.848578 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:09Z","lastTransitionTime":"2026-02-17T13:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.952102 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.952173 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.952192 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.952225 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:09 crc kubenswrapper[4955]: I0217 13:05:09.952248 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:09Z","lastTransitionTime":"2026-02-17T13:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.056166 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.056238 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.056260 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.056293 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.056315 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:10Z","lastTransitionTime":"2026-02-17T13:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.160498 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.160579 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.160596 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.160625 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.160646 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:10Z","lastTransitionTime":"2026-02-17T13:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.222246 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:05:10 crc kubenswrapper[4955]: E0217 13:05:10.222491 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.231401 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 06:21:26.898375441 +0000 UTC Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.263241 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.263301 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.263341 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.263369 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.263388 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:10Z","lastTransitionTime":"2026-02-17T13:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.366959 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.367033 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.367046 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.367069 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.367083 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:10Z","lastTransitionTime":"2026-02-17T13:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.469763 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.469836 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.469849 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.469867 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.469881 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:10Z","lastTransitionTime":"2026-02-17T13:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.572887 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.572956 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.572978 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.573013 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.573039 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:10Z","lastTransitionTime":"2026-02-17T13:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.676419 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.676496 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.676521 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.676556 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.676577 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:10Z","lastTransitionTime":"2026-02-17T13:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.779152 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.779211 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.779224 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.779247 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.779264 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:10Z","lastTransitionTime":"2026-02-17T13:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.882241 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.882286 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.882300 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.882319 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.882330 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:10Z","lastTransitionTime":"2026-02-17T13:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.985206 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.985266 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.985277 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.985297 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:10 crc kubenswrapper[4955]: I0217 13:05:10.985311 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:10Z","lastTransitionTime":"2026-02-17T13:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.088333 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.088388 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.088399 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.088420 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.088431 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:11Z","lastTransitionTime":"2026-02-17T13:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.191814 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.191874 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.191886 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.191908 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.191922 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:11Z","lastTransitionTime":"2026-02-17T13:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.222230 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.222318 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.222279 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:05:11 crc kubenswrapper[4955]: E0217 13:05:11.222473 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:05:11 crc kubenswrapper[4955]: E0217 13:05:11.222647 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:05:11 crc kubenswrapper[4955]: E0217 13:05:11.222871 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.232442 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 18:54:01.286524486 +0000 UTC Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.295936 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.295994 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.296007 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.296027 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.296041 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:11Z","lastTransitionTime":"2026-02-17T13:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.399374 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.399413 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.399423 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.399439 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.399449 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:11Z","lastTransitionTime":"2026-02-17T13:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.503150 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.503208 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.503219 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.503240 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.503255 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:11Z","lastTransitionTime":"2026-02-17T13:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.608745 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.608818 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.608828 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.608852 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.608865 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:11Z","lastTransitionTime":"2026-02-17T13:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.712224 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.712458 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.712489 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.712591 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.712622 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:11Z","lastTransitionTime":"2026-02-17T13:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.816179 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.816509 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.816523 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.816545 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.816560 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:11Z","lastTransitionTime":"2026-02-17T13:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.919312 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.919363 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.919375 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.919393 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:11 crc kubenswrapper[4955]: I0217 13:05:11.919405 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:11Z","lastTransitionTime":"2026-02-17T13:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.022706 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.022755 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.022768 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.022818 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.022832 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:12Z","lastTransitionTime":"2026-02-17T13:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.126067 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.126118 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.126130 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.126151 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.126163 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:12Z","lastTransitionTime":"2026-02-17T13:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.222273 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:05:12 crc kubenswrapper[4955]: E0217 13:05:12.222457 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.229645 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.229677 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.229685 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.229703 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.229714 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:12Z","lastTransitionTime":"2026-02-17T13:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.232906 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-30 12:07:12.552933463 +0000 UTC Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.238262 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2gsx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1face148-45d1-44b3-8775-209f9f8651fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82c5147255d021114a8639e9b60593f847896b613b5d862db588014f7485faa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnlcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2gsx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:12Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.254685 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lffhw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8155b25d-be0c-44ed-b3af-7bf25d55ec83\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad540421c8104f89cf02f413145e08de7c3e65e47f78cb2be5b40f746d6376c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gw5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52f6c4f214dd72b2df8ba07e3448bb4a2bedc698f47c16d027a774e2723c201\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gw5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lffhw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:12Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.271841 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2k5pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77480847-3271-4161-a833-e62b3bb4413e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fffmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fffmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2k5pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:12Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.287922 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aea1a9-3570-4339-a1a4-db2dc358328f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3285726bf736fdd9d99c9974bef45c180410c4e3b795acbbf6c4ac576be3ee9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c8e77568b5ec88d39510e2cdf7b9a3438a9c7cd0ad83cb350c9d7c927f0dbc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bbbc1ab7b069d1abd31929ec33bd494a7f403437f0be70da380ea0a5ff2e711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25f110bf0e9d951f4d48a491daae897a8250cf77fe464f1914cb2f1cfcde1d72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25f110bf0e9d951f4d48a491daae897a8250cf77fe464f1914cb2f1cfcde1d72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:12Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.308454 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d0966253f623158053125dffca956000824438bdc548eba6dafe9a162d5674a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d07bfb361f14d817913fa2cabd30649447790068545dda2f3a2f71213ea0e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:12Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.325744 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0f4056e4b5a554f05004c194485d25231b67024e4376a73c715fcb32004a3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:12Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.332115 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.332153 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.332164 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.332182 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.332195 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:12Z","lastTransitionTime":"2026-02-17T13:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.342550 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9ecd4380a6ae8ff62e27e3946815170c2f02b4c3921bda595467345289c9e0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:12Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.359305 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:12Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.373588 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bf36887-2a2b-4193-abdb-a15beaf46300\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11a8924eb6ce41d629feb2b42c7d4778a3ca9110a59c3ac25acbba4367272374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bec78f9e9f6416d6b600f6231f0d24652473b48b95bf0e0675cc85c9208d124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-29qxq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:12Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.390721 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9hx9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e57bdf8-e537-402a-9815-609fecdd68ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbc9014d80aee39922161014aac8fafa7d93be3e4092fe705c76da4a487a5e5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q22d4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9hx9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:12Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.405172 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-msn6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://277a12d62c90ef420c31d8dc71f0d7750dfd5f5ab7e90fec0dcd00e54b60245d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2tvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-msn6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:12Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.426058 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05aa39d-94e6-49d4-885e-2c1f986310f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T13:04:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 13:04:35.885968 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 13:04:35.888176 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3462920193/tls.crt::/tmp/serving-cert-3462920193/tls.key\\\\\\\"\\\\nI0217 13:04:42.370546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 13:04:42.384242 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 13:04:42.384284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 13:04:42.384317 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 13:04:42.384327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 13:04:42.394555 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 13:04:42.394614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394628 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 13:04:42.394646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 13:04:42.394654 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 13:04:42.394662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 13:04:42.394900 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 13:04:42.396494 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:12Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.434948 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.434979 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.434988 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.435003 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.435013 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:12Z","lastTransitionTime":"2026-02-17T13:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.442069 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:12Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.458129 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:12Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.477398 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"924c66b7-7c3c-4db7-a427-3f152efbac4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9beee007a179c3c2e04c2ce0de8ea6dbae9da3421ab24f23b718a23744c828a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fp5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:12Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.501084 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcbc7a69-58e7-4b7a-bb84-96276819ce88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7f08954bce86cfbd2f8341b30ef5ab76ba1ea7889e74ed4f8405eeb9e17fbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a7f08954bce86cfbd2f8341b30ef5ab76ba1ea7889e74ed4f8405eeb9e17fbb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"message\\\":\\\"ervices.LB{Name:\\\\\\\"Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"89fe421e-04e8-4967-ac75-77a0e6f784ef\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8383, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8081, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF0217 13:04:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dcwgg_openshift-ovn-kubernetes(dcbc7a69-58e7-4b7a-bb84-96276819ce88)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dcwgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:12Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.525166 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b49c9af-148a-48ec-af37-f555f9186337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d6919c91918e8bacafd1442b06d14ffd9b547f0f37b2298e2d8d3acc5962d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e38466a067b20ca4323ab6983cb324bb63179f942d2c37c0daa8b03767e7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4830450cbada3d24c9747e9de5f48e53e118549f1dd67edf7fff8b2adc47128f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001dfb595f39d1da33b265474b6cca51c62f522ff061c9e7c52f93806034f4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f1fee7ac5f8d1442d9f63b0e8d285187129eb3a3e3b50a49a45c7a7a39ae8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:12Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.538074 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.538131 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.538146 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.538168 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.538182 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:12Z","lastTransitionTime":"2026-02-17T13:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.541650 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5c34a3d-82ad-4d03-abdd-0a20818ac2fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0261f41b82b492d324b5e50edd72cedd5e824de542a7b47104abd2ab757a7049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f43c7fe16acfd4d39301ee64665d4cd1d885397f981d70f73215da7f3aabe72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c871c1b454e4edd2f41076ecfd6a66592023d7161f094fbb4fd197fbf9c881f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:12Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.640878 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.640919 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.640929 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.640946 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.640960 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:12Z","lastTransitionTime":"2026-02-17T13:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.744475 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.744533 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.744738 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.744759 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.744773 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:12Z","lastTransitionTime":"2026-02-17T13:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.848389 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.848469 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.848673 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.848699 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.848716 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:12Z","lastTransitionTime":"2026-02-17T13:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.951544 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.951592 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.951606 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.951640 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:12 crc kubenswrapper[4955]: I0217 13:05:12.951657 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:12Z","lastTransitionTime":"2026-02-17T13:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.054560 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.054628 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.054645 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.054670 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.054688 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:13Z","lastTransitionTime":"2026-02-17T13:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.158131 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.158205 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.158223 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.158249 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.158271 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:13Z","lastTransitionTime":"2026-02-17T13:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.222900 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.222920 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.222933 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:05:13 crc kubenswrapper[4955]: E0217 13:05:13.223080 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:05:13 crc kubenswrapper[4955]: E0217 13:05:13.223315 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:05:13 crc kubenswrapper[4955]: E0217 13:05:13.223406 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.233219 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 22:00:53.432066922 +0000 UTC Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.261320 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.261364 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.261376 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.261395 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.261411 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:13Z","lastTransitionTime":"2026-02-17T13:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.364659 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.364709 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.364718 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.364737 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.364748 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:13Z","lastTransitionTime":"2026-02-17T13:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.468622 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.468670 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.468683 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.468702 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.468718 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:13Z","lastTransitionTime":"2026-02-17T13:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.572060 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.572141 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.572164 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.572199 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.572223 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:13Z","lastTransitionTime":"2026-02-17T13:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.674845 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.674978 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.674991 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.675011 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.675021 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:13Z","lastTransitionTime":"2026-02-17T13:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.777677 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.777749 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.777767 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.777823 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.777844 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:13Z","lastTransitionTime":"2026-02-17T13:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.880974 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.881026 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.881039 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.881062 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.881076 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:13Z","lastTransitionTime":"2026-02-17T13:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.984542 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.984662 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.984681 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.984715 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:13 crc kubenswrapper[4955]: I0217 13:05:13.984738 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:13Z","lastTransitionTime":"2026-02-17T13:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.087484 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.087532 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.087542 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.087560 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.087571 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:14Z","lastTransitionTime":"2026-02-17T13:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.190027 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.190073 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.190086 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.190104 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.190116 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:14Z","lastTransitionTime":"2026-02-17T13:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.223003 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:05:14 crc kubenswrapper[4955]: E0217 13:05:14.223239 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.234074 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 17:15:54.002847668 +0000 UTC Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.293275 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.293354 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.293373 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.293398 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.293446 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:14Z","lastTransitionTime":"2026-02-17T13:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.396166 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.396234 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.396253 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.396280 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.396300 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:14Z","lastTransitionTime":"2026-02-17T13:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.401638 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/77480847-3271-4161-a833-e62b3bb4413e-metrics-certs\") pod \"network-metrics-daemon-2k5pz\" (UID: \"77480847-3271-4161-a833-e62b3bb4413e\") " pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:05:14 crc kubenswrapper[4955]: E0217 13:05:14.401883 4955 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 13:05:14 crc kubenswrapper[4955]: E0217 13:05:14.401967 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/77480847-3271-4161-a833-e62b3bb4413e-metrics-certs podName:77480847-3271-4161-a833-e62b3bb4413e nodeName:}" failed. No retries permitted until 2026-02-17 13:05:30.401940723 +0000 UTC m=+68.924670266 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/77480847-3271-4161-a833-e62b3bb4413e-metrics-certs") pod "network-metrics-daemon-2k5pz" (UID: "77480847-3271-4161-a833-e62b3bb4413e") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.499230 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.499271 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.499282 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.499301 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.499314 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:14Z","lastTransitionTime":"2026-02-17T13:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.601842 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.601896 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.601908 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.601928 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.601938 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:14Z","lastTransitionTime":"2026-02-17T13:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.705864 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.705951 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.705999 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.706097 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.706150 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:14Z","lastTransitionTime":"2026-02-17T13:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.809047 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.809134 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.809153 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.809181 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.809202 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:14Z","lastTransitionTime":"2026-02-17T13:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.913257 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.913299 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.913311 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.913330 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:14 crc kubenswrapper[4955]: I0217 13:05:14.913343 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:14Z","lastTransitionTime":"2026-02-17T13:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.009083 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.009230 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:05:15 crc kubenswrapper[4955]: E0217 13:05:15.009390 4955 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 17 13:05:15 crc kubenswrapper[4955]: E0217 13:05:15.009476 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:05:47.009338287 +0000 UTC m=+85.532067840 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:05:15 crc kubenswrapper[4955]: E0217 13:05:15.009764 4955 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 17 13:05:15 crc kubenswrapper[4955]: E0217 13:05:15.009772 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-17 13:05:47.009707087 +0000 UTC m=+85.532436630 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 17 13:05:15 crc kubenswrapper[4955]: E0217 13:05:15.009819 4955 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.009586 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:05:15 crc kubenswrapper[4955]: E0217 13:05:15.009867 4955 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.009928 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:05:15 crc kubenswrapper[4955]: E0217 13:05:15.009951 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-17 13:05:47.009937534 +0000 UTC m=+85.532667077 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 13:05:15 crc kubenswrapper[4955]: E0217 13:05:15.010052 4955 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 17 13:05:15 crc kubenswrapper[4955]: E0217 13:05:15.010115 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-17 13:05:47.010100468 +0000 UTC m=+85.532830101 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.010210 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:05:15 crc kubenswrapper[4955]: E0217 13:05:15.010320 4955 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 17 13:05:15 crc kubenswrapper[4955]: E0217 13:05:15.010336 4955 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 17 13:05:15 crc kubenswrapper[4955]: E0217 13:05:15.010351 4955 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 13:05:15 crc kubenswrapper[4955]: E0217 13:05:15.010393 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-17 13:05:47.010383426 +0000 UTC m=+85.533113049 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.017068 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.017108 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.017127 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.017155 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.017173 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:15Z","lastTransitionTime":"2026-02-17T13:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.119636 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.120024 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.120159 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.120290 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.120414 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:15Z","lastTransitionTime":"2026-02-17T13:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.222039 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.222128 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.222235 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:05:15 crc kubenswrapper[4955]: E0217 13:05:15.222274 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:05:15 crc kubenswrapper[4955]: E0217 13:05:15.222522 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:05:15 crc kubenswrapper[4955]: E0217 13:05:15.222756 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.223797 4955 scope.go:117] "RemoveContainer" containerID="3a7f08954bce86cfbd2f8341b30ef5ab76ba1ea7889e74ed4f8405eeb9e17fbb" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.224587 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.224624 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.224637 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.224657 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.224669 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:15Z","lastTransitionTime":"2026-02-17T13:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.234679 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 05:56:43.2633336 +0000 UTC Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.327369 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.327810 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.327833 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.327857 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.327873 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:15Z","lastTransitionTime":"2026-02-17T13:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.431073 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.431132 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.431148 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.431174 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.431247 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:15Z","lastTransitionTime":"2026-02-17T13:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.534474 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.534533 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.534545 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.534567 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.534579 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:15Z","lastTransitionTime":"2026-02-17T13:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.667762 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.667836 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.667847 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.667867 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.667883 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:15Z","lastTransitionTime":"2026-02-17T13:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.679256 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dcwgg_dcbc7a69-58e7-4b7a-bb84-96276819ce88/ovnkube-controller/1.log" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.682908 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" event={"ID":"dcbc7a69-58e7-4b7a-bb84-96276819ce88","Type":"ContainerStarted","Data":"29376a23843b866c283dc983b136aead4139d8ce2a97ca45633e17ad6028ced3"} Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.683488 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.699090 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9ecd4380a6ae8ff62e27e3946815170c2f02b4c3921bda595467345289c9e0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:15Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.715714 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0f4056e4b5a554f05004c194485d25231b67024e4376a73c715fcb32004a3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:15Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.735514 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05aa39d-94e6-49d4-885e-2c1f986310f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T13:04:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 13:04:35.885968 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 13:04:35.888176 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3462920193/tls.crt::/tmp/serving-cert-3462920193/tls.key\\\\\\\"\\\\nI0217 13:04:42.370546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 13:04:42.384242 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 13:04:42.384284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 13:04:42.384317 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 13:04:42.384327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 13:04:42.394555 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 13:04:42.394614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394628 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 13:04:42.394646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 13:04:42.394654 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 13:04:42.394662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 13:04:42.394900 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 13:04:42.396494 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:15Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.750685 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:15Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.764197 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:15Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.770267 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.770307 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.770318 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.770337 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.770348 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:15Z","lastTransitionTime":"2026-02-17T13:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.777469 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bf36887-2a2b-4193-abdb-a15beaf46300\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11a8924eb6ce41d629feb2b42c7d4778a3ca9110a59c3ac25acbba4367272374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bec78f9e9f6416d6b600f6231f0d24652473b48b95bf0e0675cc85c9208d124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-29qxq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:15Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.789857 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9hx9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e57bdf8-e537-402a-9815-609fecdd68ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbc9014d80aee39922161014aac8fafa7d93be3e4092fe705c76da4a487a5e5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q22d4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9hx9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:15Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.799975 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-msn6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://277a12d62c90ef420c31d8dc71f0d7750dfd5f5ab7e90fec0dcd00e54b60245d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2tvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-msn6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:15Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.820165 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b49c9af-148a-48ec-af37-f555f9186337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d6919c91918e8bacafd1442b06d14ffd9b547f0f37b2298e2d8d3acc5962d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e38466a067b20ca4323ab6983cb324bb63179f942d2c37c0daa8b03767e7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4830450cbada3d24c9747e9de5f48e53e118549f1dd67edf7fff8b2adc47128f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001dfb595f39d1da33b265474b6cca51c62f522ff061c9e7c52f93806034f4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f1fee7ac5f8d1442d9f63b0e8d285187129eb3a3e3b50a49a45c7a7a39ae8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:15Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.835219 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5c34a3d-82ad-4d03-abdd-0a20818ac2fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0261f41b82b492d324b5e50edd72cedd5e824de542a7b47104abd2ab757a7049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f43c7fe16acfd4d39301ee64665d4cd1d885397f981d70f73215da7f3aabe72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c871c1b454e4edd2f41076ecfd6a66592023d7161f094fbb4fd197fbf9c881f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:15Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.851825 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:15Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.867563 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"924c66b7-7c3c-4db7-a427-3f152efbac4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9beee007a179c3c2e04c2ce0de8ea6dbae9da3421ab24f23b718a23744c828a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fp5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:15Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.872734 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.872798 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.872810 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.872829 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.872843 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:15Z","lastTransitionTime":"2026-02-17T13:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.886980 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcbc7a69-58e7-4b7a-bb84-96276819ce88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29376a23843b866c283dc983b136aead4139d8ce2a97ca45633e17ad6028ced3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a7f08954bce86cfbd2f8341b30ef5ab76ba1ea7889e74ed4f8405eeb9e17fbb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"message\\\":\\\"ervices.LB{Name:\\\\\\\"Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"89fe421e-04e8-4967-ac75-77a0e6f784ef\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8383, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8081, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF0217 13:04:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dcwgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:15Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.900937 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aea1a9-3570-4339-a1a4-db2dc358328f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3285726bf736fdd9d99c9974bef45c180410c4e3b795acbbf6c4ac576be3ee9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c8e77568b5ec88d39510e2cdf7b9a3438a9c7cd0ad83cb350c9d7c927f0dbc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bbbc1ab7b069d1abd31929ec33bd494a7f403437f0be70da380ea0a5ff2e711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25f110bf0e9d951f4d48a491daae897a8250cf77fe464f1914cb2f1cfcde1d72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25f110bf0e9d951f4d48a491daae897a8250cf77fe464f1914cb2f1cfcde1d72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:15Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.914231 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d0966253f623158053125dffca956000824438bdc548eba6dafe9a162d5674a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d07bfb361f14d817913fa2cabd30649447790068545dda2f3a2f71213ea0e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:15Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.932651 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2gsx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1face148-45d1-44b3-8775-209f9f8651fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82c5147255d021114a8639e9b60593f847896b613b5d862db588014f7485faa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnlcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2gsx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:15Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.945747 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lffhw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8155b25d-be0c-44ed-b3af-7bf25d55ec83\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad540421c8104f89cf02f413145e08de7c3e65e47f78cb2be5b40f746d6376c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gw5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52f6c4f214dd72b2df8ba07e3448bb4a2bedc698f47c16d027a774e2723c201\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gw5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lffhw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:15Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.960099 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2k5pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77480847-3271-4161-a833-e62b3bb4413e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fffmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fffmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2k5pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:15Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.975655 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.975745 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.975763 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.975802 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:15 crc kubenswrapper[4955]: I0217 13:05:15.975818 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:15Z","lastTransitionTime":"2026-02-17T13:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.079290 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.079351 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.079370 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.079396 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.079414 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:16Z","lastTransitionTime":"2026-02-17T13:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.182583 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.182648 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.182665 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.182693 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.182713 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:16Z","lastTransitionTime":"2026-02-17T13:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.221918 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:05:16 crc kubenswrapper[4955]: E0217 13:05:16.222115 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.235732 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 13:47:37.410376382 +0000 UTC Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.286370 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.286455 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.286474 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.286507 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.286527 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:16Z","lastTransitionTime":"2026-02-17T13:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.389133 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.389212 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.389239 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.389276 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.389297 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:16Z","lastTransitionTime":"2026-02-17T13:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.492573 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.492648 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.492674 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.492711 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.492737 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:16Z","lastTransitionTime":"2026-02-17T13:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.596125 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.596194 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.596211 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.596235 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.596251 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:16Z","lastTransitionTime":"2026-02-17T13:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.689368 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dcwgg_dcbc7a69-58e7-4b7a-bb84-96276819ce88/ovnkube-controller/2.log" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.689964 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dcwgg_dcbc7a69-58e7-4b7a-bb84-96276819ce88/ovnkube-controller/1.log" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.693981 4955 generic.go:334] "Generic (PLEG): container finished" podID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerID="29376a23843b866c283dc983b136aead4139d8ce2a97ca45633e17ad6028ced3" exitCode=1 Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.694064 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" event={"ID":"dcbc7a69-58e7-4b7a-bb84-96276819ce88","Type":"ContainerDied","Data":"29376a23843b866c283dc983b136aead4139d8ce2a97ca45633e17ad6028ced3"} Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.694183 4955 scope.go:117] "RemoveContainer" containerID="3a7f08954bce86cfbd2f8341b30ef5ab76ba1ea7889e74ed4f8405eeb9e17fbb" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.694751 4955 scope.go:117] "RemoveContainer" containerID="29376a23843b866c283dc983b136aead4139d8ce2a97ca45633e17ad6028ced3" Feb 17 13:05:16 crc kubenswrapper[4955]: E0217 13:05:16.695073 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dcwgg_openshift-ovn-kubernetes(dcbc7a69-58e7-4b7a-bb84-96276819ce88)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.698891 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.698936 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.698946 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.698963 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.698975 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:16Z","lastTransitionTime":"2026-02-17T13:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.711369 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9ecd4380a6ae8ff62e27e3946815170c2f02b4c3921bda595467345289c9e0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:16Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.735610 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0f4056e4b5a554f05004c194485d25231b67024e4376a73c715fcb32004a3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:16Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.760592 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:16Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.775767 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:16Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.792580 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bf36887-2a2b-4193-abdb-a15beaf46300\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11a8924eb6ce41d629feb2b42c7d4778a3ca9110a59c3ac25acbba4367272374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bec78f9e9f6416d6b600f6231f0d24652473b48b95bf0e0675cc85c9208d124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-29qxq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:16Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.802066 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.802102 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.802112 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.802129 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.802140 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:16Z","lastTransitionTime":"2026-02-17T13:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.807199 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9hx9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e57bdf8-e537-402a-9815-609fecdd68ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbc9014d80aee39922161014aac8fafa7d93be3e4092fe705c76da4a487a5e5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q22d4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9hx9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:16Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.819500 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-msn6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://277a12d62c90ef420c31d8dc71f0d7750dfd5f5ab7e90fec0dcd00e54b60245d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2tvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-msn6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:16Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.836460 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05aa39d-94e6-49d4-885e-2c1f986310f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T13:04:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 13:04:35.885968 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 13:04:35.888176 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3462920193/tls.crt::/tmp/serving-cert-3462920193/tls.key\\\\\\\"\\\\nI0217 13:04:42.370546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 13:04:42.384242 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 13:04:42.384284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 13:04:42.384317 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 13:04:42.384327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 13:04:42.394555 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 13:04:42.394614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394628 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 13:04:42.394646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 13:04:42.394654 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 13:04:42.394662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 13:04:42.394900 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 13:04:42.396494 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:16Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.854528 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5c34a3d-82ad-4d03-abdd-0a20818ac2fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0261f41b82b492d324b5e50edd72cedd5e824de542a7b47104abd2ab757a7049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f43c7fe16acfd4d39301ee64665d4cd1d885397f981d70f73215da7f3aabe72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c871c1b454e4edd2f41076ecfd6a66592023d7161f094fbb4fd197fbf9c881f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:16Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.866855 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:16Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.882801 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"924c66b7-7c3c-4db7-a427-3f152efbac4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9beee007a179c3c2e04c2ce0de8ea6dbae9da3421ab24f23b718a23744c828a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fp5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:16Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.900363 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcbc7a69-58e7-4b7a-bb84-96276819ce88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29376a23843b866c283dc983b136aead4139d8ce2a97ca45633e17ad6028ced3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a7f08954bce86cfbd2f8341b30ef5ab76ba1ea7889e74ed4f8405eeb9e17fbb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"message\\\":\\\"ervices.LB{Name:\\\\\\\"Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"89fe421e-04e8-4967-ac75-77a0e6f784ef\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8383, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8081, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF0217 13:04:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29376a23843b866c283dc983b136aead4139d8ce2a97ca45633e17ad6028ced3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T13:05:16Z\\\",\\\"message\\\":\\\"licy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0217 13:05:16.060354 6583 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0217 13:05:16.060455 6583 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0217 13:05:16.060913 6583 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0217 13:05:16.061605 6583 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0217 13:05:16.061682 6583 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0217 13:05:16.061709 6583 factory.go:656] Stopping watch factory\\\\nI0217 13:05:16.061733 6583 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0217 13:05:16.061766 6583 handler.go:208] Removed *v1.Node event handler 2\\\\nI0217 13:05:16.088613 6583 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0217 13:05:16.088662 6583 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0217 13:05:16.088736 6583 ovnkube.go:599] Stopped ovnkube\\\\nI0217 13:05:16.088764 6583 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0217 13:05:16.088897 6583 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dcwgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:16Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.904834 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.904875 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.904886 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.904907 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.904921 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:16Z","lastTransitionTime":"2026-02-17T13:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.919224 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b49c9af-148a-48ec-af37-f555f9186337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d6919c91918e8bacafd1442b06d14ffd9b547f0f37b2298e2d8d3acc5962d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e38466a067b20ca4323ab6983cb324bb63179f942d2c37c0daa8b03767e7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4830450cbada3d24c9747e9de5f48e53e118549f1dd67edf7fff8b2adc47128f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001dfb595f39d1da33b265474b6cca51c62f522ff061c9e7c52f93806034f4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f1fee7ac5f8d1442d9f63b0e8d285187129eb3a3e3b50a49a45c7a7a39ae8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:16Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.932951 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d0966253f623158053125dffca956000824438bdc548eba6dafe9a162d5674a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d07bfb361f14d817913fa2cabd30649447790068545dda2f3a2f71213ea0e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:16Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.945041 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2gsx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1face148-45d1-44b3-8775-209f9f8651fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82c5147255d021114a8639e9b60593f847896b613b5d862db588014f7485faa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnlcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2gsx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:16Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.957861 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lffhw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8155b25d-be0c-44ed-b3af-7bf25d55ec83\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad540421c8104f89cf02f413145e08de7c3e65e47f78cb2be5b40f746d6376c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gw5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52f6c4f214dd72b2df8ba07e3448bb4a2bedc698f47c16d027a774e2723c201\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gw5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lffhw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:16Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.968729 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2k5pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77480847-3271-4161-a833-e62b3bb4413e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fffmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fffmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2k5pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:16Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:16 crc kubenswrapper[4955]: I0217 13:05:16.982672 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aea1a9-3570-4339-a1a4-db2dc358328f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3285726bf736fdd9d99c9974bef45c180410c4e3b795acbbf6c4ac576be3ee9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c8e77568b5ec88d39510e2cdf7b9a3438a9c7cd0ad83cb350c9d7c927f0dbc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bbbc1ab7b069d1abd31929ec33bd494a7f403437f0be70da380ea0a5ff2e711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25f110bf0e9d951f4d48a491daae897a8250cf77fe464f1914cb2f1cfcde1d72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25f110bf0e9d951f4d48a491daae897a8250cf77fe464f1914cb2f1cfcde1d72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:16Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.008294 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.008352 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.008366 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.008387 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.008406 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:17Z","lastTransitionTime":"2026-02-17T13:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.117418 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.117465 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.117476 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.117496 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.117508 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:17Z","lastTransitionTime":"2026-02-17T13:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.220375 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.220424 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.220437 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.220462 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.220478 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:17Z","lastTransitionTime":"2026-02-17T13:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.222619 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.222619 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:05:17 crc kubenswrapper[4955]: E0217 13:05:17.222749 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.222790 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:05:17 crc kubenswrapper[4955]: E0217 13:05:17.222851 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:05:17 crc kubenswrapper[4955]: E0217 13:05:17.223049 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.236853 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 09:51:21.243518888 +0000 UTC Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.323618 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.323661 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.323672 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.323694 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.323708 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:17Z","lastTransitionTime":"2026-02-17T13:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.426201 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.426250 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.426265 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.426285 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.426297 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:17Z","lastTransitionTime":"2026-02-17T13:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.529836 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.529920 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.529946 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.529968 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.529982 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:17Z","lastTransitionTime":"2026-02-17T13:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.632616 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.632676 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.632686 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.632702 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.632713 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:17Z","lastTransitionTime":"2026-02-17T13:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.700885 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dcwgg_dcbc7a69-58e7-4b7a-bb84-96276819ce88/ovnkube-controller/2.log" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.705254 4955 scope.go:117] "RemoveContainer" containerID="29376a23843b866c283dc983b136aead4139d8ce2a97ca45633e17ad6028ced3" Feb 17 13:05:17 crc kubenswrapper[4955]: E0217 13:05:17.705527 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dcwgg_openshift-ovn-kubernetes(dcbc7a69-58e7-4b7a-bb84-96276819ce88)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.721758 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:17Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.738187 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.738239 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.738258 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.738289 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.738312 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:17Z","lastTransitionTime":"2026-02-17T13:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.740173 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:17Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.758705 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bf36887-2a2b-4193-abdb-a15beaf46300\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11a8924eb6ce41d629feb2b42c7d4778a3ca9110a59c3ac25acbba4367272374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bec78f9e9f6416d6b600f6231f0d24652473b48b95bf0e0675cc85c9208d124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-29qxq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:17Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.780235 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9hx9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e57bdf8-e537-402a-9815-609fecdd68ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbc9014d80aee39922161014aac8fafa7d93be3e4092fe705c76da4a487a5e5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q22d4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9hx9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:17Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.798762 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-msn6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://277a12d62c90ef420c31d8dc71f0d7750dfd5f5ab7e90fec0dcd00e54b60245d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2tvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-msn6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:17Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.821904 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05aa39d-94e6-49d4-885e-2c1f986310f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T13:04:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 13:04:35.885968 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 13:04:35.888176 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3462920193/tls.crt::/tmp/serving-cert-3462920193/tls.key\\\\\\\"\\\\nI0217 13:04:42.370546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 13:04:42.384242 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 13:04:42.384284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 13:04:42.384317 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 13:04:42.384327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 13:04:42.394555 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 13:04:42.394614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394628 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 13:04:42.394646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 13:04:42.394654 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 13:04:42.394662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 13:04:42.394900 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 13:04:42.396494 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:17Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.841862 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.841918 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.841931 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.841957 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.841973 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:17Z","lastTransitionTime":"2026-02-17T13:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.841923 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5c34a3d-82ad-4d03-abdd-0a20818ac2fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0261f41b82b492d324b5e50edd72cedd5e824de542a7b47104abd2ab757a7049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f43c7fe16acfd4d39301ee64665d4cd1d885397f981d70f73215da7f3aabe72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c871c1b454e4edd2f41076ecfd6a66592023d7161f094fbb4fd197fbf9c881f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:17Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.859769 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:17Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.882886 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"924c66b7-7c3c-4db7-a427-3f152efbac4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9beee007a179c3c2e04c2ce0de8ea6dbae9da3421ab24f23b718a23744c828a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fp5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:17Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.915058 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcbc7a69-58e7-4b7a-bb84-96276819ce88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29376a23843b866c283dc983b136aead4139d8ce2a97ca45633e17ad6028ced3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29376a23843b866c283dc983b136aead4139d8ce2a97ca45633e17ad6028ced3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T13:05:16Z\\\",\\\"message\\\":\\\"licy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0217 13:05:16.060354 6583 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0217 13:05:16.060455 6583 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0217 13:05:16.060913 6583 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0217 13:05:16.061605 6583 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0217 13:05:16.061682 6583 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0217 13:05:16.061709 6583 factory.go:656] Stopping watch factory\\\\nI0217 13:05:16.061733 6583 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0217 13:05:16.061766 6583 handler.go:208] Removed *v1.Node event handler 2\\\\nI0217 13:05:16.088613 6583 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0217 13:05:16.088662 6583 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0217 13:05:16.088736 6583 ovnkube.go:599] Stopped ovnkube\\\\nI0217 13:05:16.088764 6583 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0217 13:05:16.088897 6583 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:05:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dcwgg_openshift-ovn-kubernetes(dcbc7a69-58e7-4b7a-bb84-96276819ce88)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dcwgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:17Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.945284 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.945353 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.945370 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.945399 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.945419 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:17Z","lastTransitionTime":"2026-02-17T13:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.952089 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b49c9af-148a-48ec-af37-f555f9186337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d6919c91918e8bacafd1442b06d14ffd9b547f0f37b2298e2d8d3acc5962d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e38466a067b20ca4323ab6983cb324bb63179f942d2c37c0daa8b03767e7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4830450cbada3d24c9747e9de5f48e53e118549f1dd67edf7fff8b2adc47128f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001dfb595f39d1da33b265474b6cca51c62f522ff061c9e7c52f93806034f4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f1fee7ac5f8d1442d9f63b0e8d285187129eb3a3e3b50a49a45c7a7a39ae8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:17Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.976661 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d0966253f623158053125dffca956000824438bdc548eba6dafe9a162d5674a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d07bfb361f14d817913fa2cabd30649447790068545dda2f3a2f71213ea0e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:17Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:17 crc kubenswrapper[4955]: I0217 13:05:17.991887 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2gsx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1face148-45d1-44b3-8775-209f9f8651fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82c5147255d021114a8639e9b60593f847896b613b5d862db588014f7485faa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnlcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2gsx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:17Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.011677 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lffhw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8155b25d-be0c-44ed-b3af-7bf25d55ec83\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad540421c8104f89cf02f413145e08de7c3e65e47f78cb2be5b40f746d6376c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gw5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52f6c4f214dd72b2df8ba07e3448bb4a2bedc698f47c16d027a774e2723c201\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gw5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lffhw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:18Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.026901 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2k5pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77480847-3271-4161-a833-e62b3bb4413e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fffmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fffmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2k5pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:18Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.042753 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aea1a9-3570-4339-a1a4-db2dc358328f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3285726bf736fdd9d99c9974bef45c180410c4e3b795acbbf6c4ac576be3ee9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c8e77568b5ec88d39510e2cdf7b9a3438a9c7cd0ad83cb350c9d7c927f0dbc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bbbc1ab7b069d1abd31929ec33bd494a7f403437f0be70da380ea0a5ff2e711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25f110bf0e9d951f4d48a491daae897a8250cf77fe464f1914cb2f1cfcde1d72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25f110bf0e9d951f4d48a491daae897a8250cf77fe464f1914cb2f1cfcde1d72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:18Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.059863 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.059919 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.059932 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.059955 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.059968 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:18Z","lastTransitionTime":"2026-02-17T13:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.065381 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9ecd4380a6ae8ff62e27e3946815170c2f02b4c3921bda595467345289c9e0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:18Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.084083 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0f4056e4b5a554f05004c194485d25231b67024e4376a73c715fcb32004a3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:18Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.163679 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.163752 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.163777 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.163831 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.163852 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:18Z","lastTransitionTime":"2026-02-17T13:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.222163 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:05:18 crc kubenswrapper[4955]: E0217 13:05:18.222376 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.237651 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 15:11:35.27499193 +0000 UTC Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.267250 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.267329 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.267348 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.267375 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.267396 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:18Z","lastTransitionTime":"2026-02-17T13:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.370578 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.370642 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.370660 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.370686 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.370703 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:18Z","lastTransitionTime":"2026-02-17T13:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.473462 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.473507 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.473515 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.473534 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.473545 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:18Z","lastTransitionTime":"2026-02-17T13:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.576181 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.576230 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.576243 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.576262 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.576276 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:18Z","lastTransitionTime":"2026-02-17T13:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.679065 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.679119 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.679130 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.679148 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.679158 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:18Z","lastTransitionTime":"2026-02-17T13:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.782743 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.782819 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.782831 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.782853 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.782869 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:18Z","lastTransitionTime":"2026-02-17T13:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.885989 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.886037 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.886047 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.886062 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.886074 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:18Z","lastTransitionTime":"2026-02-17T13:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.989093 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.989160 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.989180 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.989206 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:18 crc kubenswrapper[4955]: I0217 13:05:18.989225 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:18Z","lastTransitionTime":"2026-02-17T13:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.092421 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.092481 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.092495 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.092518 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.092538 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:19Z","lastTransitionTime":"2026-02-17T13:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.196534 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.196581 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.196595 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.196618 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.196634 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:19Z","lastTransitionTime":"2026-02-17T13:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.222962 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.223063 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:05:19 crc kubenswrapper[4955]: E0217 13:05:19.223203 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.223279 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:05:19 crc kubenswrapper[4955]: E0217 13:05:19.223429 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:05:19 crc kubenswrapper[4955]: E0217 13:05:19.223657 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.238467 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 02:11:24.641384457 +0000 UTC Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.299433 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.299502 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.299522 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.299556 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.299598 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:19Z","lastTransitionTime":"2026-02-17T13:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.402670 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.402725 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.402734 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.402751 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.402762 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:19Z","lastTransitionTime":"2026-02-17T13:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.506224 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.506286 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.506308 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.506333 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.506347 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:19Z","lastTransitionTime":"2026-02-17T13:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.609659 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.609872 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.609904 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.609936 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.609955 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:19Z","lastTransitionTime":"2026-02-17T13:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.713362 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.713413 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.713424 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.713443 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.713462 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:19Z","lastTransitionTime":"2026-02-17T13:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.807724 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.808173 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.808250 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.808329 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.808397 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:19Z","lastTransitionTime":"2026-02-17T13:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:19 crc kubenswrapper[4955]: E0217 13:05:19.823246 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7605a498-360b-4ce4-b7a1-dfbc354d798d\\\",\\\"systemUUID\\\":\\\"6d4602d9-c718-4393-8f91-3c381f800c9e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:19Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.828336 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.828626 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.828879 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.829024 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.829170 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:19Z","lastTransitionTime":"2026-02-17T13:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:19 crc kubenswrapper[4955]: E0217 13:05:19.851315 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7605a498-360b-4ce4-b7a1-dfbc354d798d\\\",\\\"systemUUID\\\":\\\"6d4602d9-c718-4393-8f91-3c381f800c9e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:19Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.855480 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.855542 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.855561 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.855586 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.855605 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:19Z","lastTransitionTime":"2026-02-17T13:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:19 crc kubenswrapper[4955]: E0217 13:05:19.872542 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7605a498-360b-4ce4-b7a1-dfbc354d798d\\\",\\\"systemUUID\\\":\\\"6d4602d9-c718-4393-8f91-3c381f800c9e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:19Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.878508 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.878576 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.878597 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.878626 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.878646 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:19Z","lastTransitionTime":"2026-02-17T13:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:19 crc kubenswrapper[4955]: E0217 13:05:19.894482 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7605a498-360b-4ce4-b7a1-dfbc354d798d\\\",\\\"systemUUID\\\":\\\"6d4602d9-c718-4393-8f91-3c381f800c9e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:19Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.899889 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.899950 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.899963 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.899987 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.900004 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:19Z","lastTransitionTime":"2026-02-17T13:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:19 crc kubenswrapper[4955]: E0217 13:05:19.919842 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7605a498-360b-4ce4-b7a1-dfbc354d798d\\\",\\\"systemUUID\\\":\\\"6d4602d9-c718-4393-8f91-3c381f800c9e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:19Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:19 crc kubenswrapper[4955]: E0217 13:05:19.920061 4955 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.922109 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.922159 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.922178 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.922202 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:19 crc kubenswrapper[4955]: I0217 13:05:19.922222 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:19Z","lastTransitionTime":"2026-02-17T13:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.026311 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.026389 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.026408 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.026437 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.026459 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:20Z","lastTransitionTime":"2026-02-17T13:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.130395 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.130455 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.130468 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.130491 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.130504 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:20Z","lastTransitionTime":"2026-02-17T13:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.222952 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:05:20 crc kubenswrapper[4955]: E0217 13:05:20.223153 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.235665 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.235748 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.235766 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.235818 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.235843 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:20Z","lastTransitionTime":"2026-02-17T13:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.238897 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 07:21:09.630460611 +0000 UTC Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.340627 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.340691 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.340706 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.340724 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.340739 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:20Z","lastTransitionTime":"2026-02-17T13:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.444541 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.444620 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.444637 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.444671 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.444696 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:20Z","lastTransitionTime":"2026-02-17T13:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.549137 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.549213 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.549235 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.549263 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.549285 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:20Z","lastTransitionTime":"2026-02-17T13:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.652263 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.652356 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.652381 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.652419 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.652443 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:20Z","lastTransitionTime":"2026-02-17T13:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.755480 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.755544 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.755566 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.755593 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.755611 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:20Z","lastTransitionTime":"2026-02-17T13:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.858772 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.858919 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.858936 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.858961 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.858982 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:20Z","lastTransitionTime":"2026-02-17T13:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.962880 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.962934 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.962947 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.962967 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:20 crc kubenswrapper[4955]: I0217 13:05:20.962981 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:20Z","lastTransitionTime":"2026-02-17T13:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.066145 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.066212 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.066226 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.066251 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.066267 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:21Z","lastTransitionTime":"2026-02-17T13:05:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.169209 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.169282 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.169308 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.169327 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.169359 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:21Z","lastTransitionTime":"2026-02-17T13:05:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.222290 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.222424 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:05:21 crc kubenswrapper[4955]: E0217 13:05:21.222486 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.222501 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:05:21 crc kubenswrapper[4955]: E0217 13:05:21.222676 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:05:21 crc kubenswrapper[4955]: E0217 13:05:21.222860 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.239617 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 19:07:41.927674271 +0000 UTC Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.272418 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.272472 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.272488 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.272507 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.272519 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:21Z","lastTransitionTime":"2026-02-17T13:05:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.375674 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.375720 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.375731 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.375752 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.375767 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:21Z","lastTransitionTime":"2026-02-17T13:05:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.478689 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.478736 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.478747 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.478764 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.478775 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:21Z","lastTransitionTime":"2026-02-17T13:05:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.582497 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.582572 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.582591 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.582620 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.582638 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:21Z","lastTransitionTime":"2026-02-17T13:05:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.685827 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.685916 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.685936 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.685964 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.685982 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:21Z","lastTransitionTime":"2026-02-17T13:05:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.788773 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.788871 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.788888 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.788916 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.788934 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:21Z","lastTransitionTime":"2026-02-17T13:05:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.892758 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.892985 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.893013 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.893046 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.893072 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:21Z","lastTransitionTime":"2026-02-17T13:05:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.996066 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.996145 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.996163 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.996186 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:21 crc kubenswrapper[4955]: I0217 13:05:21.996202 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:21Z","lastTransitionTime":"2026-02-17T13:05:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.099624 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.099697 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.099716 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.099746 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.099766 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:22Z","lastTransitionTime":"2026-02-17T13:05:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.202910 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.202974 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.202988 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.203012 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.203027 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:22Z","lastTransitionTime":"2026-02-17T13:05:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.221980 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:05:22 crc kubenswrapper[4955]: E0217 13:05:22.223378 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.240652 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 04:58:57.776732441 +0000 UTC Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.243814 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9ecd4380a6ae8ff62e27e3946815170c2f02b4c3921bda595467345289c9e0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:22Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.256765 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0f4056e4b5a554f05004c194485d25231b67024e4376a73c715fcb32004a3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:22Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.269698 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05aa39d-94e6-49d4-885e-2c1f986310f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T13:04:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 13:04:35.885968 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 13:04:35.888176 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3462920193/tls.crt::/tmp/serving-cert-3462920193/tls.key\\\\\\\"\\\\nI0217 13:04:42.370546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 13:04:42.384242 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 13:04:42.384284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 13:04:42.384317 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 13:04:42.384327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 13:04:42.394555 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 13:04:42.394614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394628 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 13:04:42.394646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 13:04:42.394654 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 13:04:42.394662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 13:04:42.394900 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 13:04:42.396494 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:22Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.282005 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:22Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.293637 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:22Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.305892 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bf36887-2a2b-4193-abdb-a15beaf46300\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11a8924eb6ce41d629feb2b42c7d4778a3ca9110a59c3ac25acbba4367272374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bec78f9e9f6416d6b600f6231f0d24652473b48b95bf0e0675cc85c9208d124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-29qxq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:22Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.306976 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.307013 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.307022 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.307038 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.307048 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:22Z","lastTransitionTime":"2026-02-17T13:05:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.320269 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9hx9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e57bdf8-e537-402a-9815-609fecdd68ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbc9014d80aee39922161014aac8fafa7d93be3e4092fe705c76da4a487a5e5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q22d4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9hx9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:22Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.332254 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-msn6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://277a12d62c90ef420c31d8dc71f0d7750dfd5f5ab7e90fec0dcd00e54b60245d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2tvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-msn6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:22Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.360947 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b49c9af-148a-48ec-af37-f555f9186337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d6919c91918e8bacafd1442b06d14ffd9b547f0f37b2298e2d8d3acc5962d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e38466a067b20ca4323ab6983cb324bb63179f942d2c37c0daa8b03767e7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4830450cbada3d24c9747e9de5f48e53e118549f1dd67edf7fff8b2adc47128f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001dfb595f39d1da33b265474b6cca51c62f522ff061c9e7c52f93806034f4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f1fee7ac5f8d1442d9f63b0e8d285187129eb3a3e3b50a49a45c7a7a39ae8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:22Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.379666 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5c34a3d-82ad-4d03-abdd-0a20818ac2fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0261f41b82b492d324b5e50edd72cedd5e824de542a7b47104abd2ab757a7049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f43c7fe16acfd4d39301ee64665d4cd1d885397f981d70f73215da7f3aabe72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c871c1b454e4edd2f41076ecfd6a66592023d7161f094fbb4fd197fbf9c881f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:22Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.395003 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:22Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.410118 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.410153 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.410165 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.410183 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.410196 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:22Z","lastTransitionTime":"2026-02-17T13:05:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.414171 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"924c66b7-7c3c-4db7-a427-3f152efbac4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9beee007a179c3c2e04c2ce0de8ea6dbae9da3421ab24f23b718a23744c828a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fp5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:22Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.432933 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcbc7a69-58e7-4b7a-bb84-96276819ce88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29376a23843b866c283dc983b136aead4139d8ce2a97ca45633e17ad6028ced3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29376a23843b866c283dc983b136aead4139d8ce2a97ca45633e17ad6028ced3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T13:05:16Z\\\",\\\"message\\\":\\\"licy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0217 13:05:16.060354 6583 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0217 13:05:16.060455 6583 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0217 13:05:16.060913 6583 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0217 13:05:16.061605 6583 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0217 13:05:16.061682 6583 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0217 13:05:16.061709 6583 factory.go:656] Stopping watch factory\\\\nI0217 13:05:16.061733 6583 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0217 13:05:16.061766 6583 handler.go:208] Removed *v1.Node event handler 2\\\\nI0217 13:05:16.088613 6583 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0217 13:05:16.088662 6583 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0217 13:05:16.088736 6583 ovnkube.go:599] Stopped ovnkube\\\\nI0217 13:05:16.088764 6583 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0217 13:05:16.088897 6583 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:05:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dcwgg_openshift-ovn-kubernetes(dcbc7a69-58e7-4b7a-bb84-96276819ce88)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dcwgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:22Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.450668 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aea1a9-3570-4339-a1a4-db2dc358328f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3285726bf736fdd9d99c9974bef45c180410c4e3b795acbbf6c4ac576be3ee9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c8e77568b5ec88d39510e2cdf7b9a3438a9c7cd0ad83cb350c9d7c927f0dbc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bbbc1ab7b069d1abd31929ec33bd494a7f403437f0be70da380ea0a5ff2e711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25f110bf0e9d951f4d48a491daae897a8250cf77fe464f1914cb2f1cfcde1d72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25f110bf0e9d951f4d48a491daae897a8250cf77fe464f1914cb2f1cfcde1d72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:22Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.464954 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d0966253f623158053125dffca956000824438bdc548eba6dafe9a162d5674a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d07bfb361f14d817913fa2cabd30649447790068545dda2f3a2f71213ea0e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:22Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.477764 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2gsx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1face148-45d1-44b3-8775-209f9f8651fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82c5147255d021114a8639e9b60593f847896b613b5d862db588014f7485faa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnlcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2gsx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:22Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.489364 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lffhw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8155b25d-be0c-44ed-b3af-7bf25d55ec83\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad540421c8104f89cf02f413145e08de7c3e65e47f78cb2be5b40f746d6376c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gw5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52f6c4f214dd72b2df8ba07e3448bb4a2bedc698f47c16d027a774e2723c201\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gw5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lffhw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:22Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.502473 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2k5pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77480847-3271-4161-a833-e62b3bb4413e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fffmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fffmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2k5pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:22Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.513680 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.513736 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.513749 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.513770 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.513800 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:22Z","lastTransitionTime":"2026-02-17T13:05:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.616982 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.617379 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.617487 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.617582 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.617670 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:22Z","lastTransitionTime":"2026-02-17T13:05:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.720730 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.720778 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.720802 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.720860 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.720884 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:22Z","lastTransitionTime":"2026-02-17T13:05:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.823738 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.824091 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.824188 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.824280 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.824355 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:22Z","lastTransitionTime":"2026-02-17T13:05:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.927888 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.927941 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.927957 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.928002 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:22 crc kubenswrapper[4955]: I0217 13:05:22.928017 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:22Z","lastTransitionTime":"2026-02-17T13:05:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.031527 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.031599 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.031618 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.031645 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.031665 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:23Z","lastTransitionTime":"2026-02-17T13:05:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.135120 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.135192 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.135213 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.135236 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.135252 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:23Z","lastTransitionTime":"2026-02-17T13:05:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.222479 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.222519 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.222630 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:05:23 crc kubenswrapper[4955]: E0217 13:05:23.223050 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:05:23 crc kubenswrapper[4955]: E0217 13:05:23.223347 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:05:23 crc kubenswrapper[4955]: E0217 13:05:23.223419 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.237539 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.237582 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.237594 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.237611 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.237624 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:23Z","lastTransitionTime":"2026-02-17T13:05:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.242158 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 10:13:38.891004651 +0000 UTC Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.340180 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.340241 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.340256 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.340280 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.340295 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:23Z","lastTransitionTime":"2026-02-17T13:05:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.443920 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.443986 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.443999 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.444023 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.444041 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:23Z","lastTransitionTime":"2026-02-17T13:05:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.547296 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.547365 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.547382 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.547406 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.547426 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:23Z","lastTransitionTime":"2026-02-17T13:05:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.650684 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.650768 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.650837 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.650864 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.650884 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:23Z","lastTransitionTime":"2026-02-17T13:05:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.754565 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.754610 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.754620 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.754639 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.754650 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:23Z","lastTransitionTime":"2026-02-17T13:05:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.857901 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.857963 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.857977 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.857998 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.858013 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:23Z","lastTransitionTime":"2026-02-17T13:05:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.961956 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.962027 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.962070 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.962101 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:23 crc kubenswrapper[4955]: I0217 13:05:23.962124 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:23Z","lastTransitionTime":"2026-02-17T13:05:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.065762 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.065840 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.065863 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.065889 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.065907 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:24Z","lastTransitionTime":"2026-02-17T13:05:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.170204 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.170284 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.170301 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.170330 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.170349 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:24Z","lastTransitionTime":"2026-02-17T13:05:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.222887 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:05:24 crc kubenswrapper[4955]: E0217 13:05:24.223133 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.242360 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 15:01:34.526972429 +0000 UTC Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.274086 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.274169 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.274197 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.274231 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.274255 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:24Z","lastTransitionTime":"2026-02-17T13:05:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.377491 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.377566 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.377583 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.377612 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.377633 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:24Z","lastTransitionTime":"2026-02-17T13:05:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.489231 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.489302 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.489318 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.489343 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.489362 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:24Z","lastTransitionTime":"2026-02-17T13:05:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.592836 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.592981 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.593002 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.593068 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.593093 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:24Z","lastTransitionTime":"2026-02-17T13:05:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.696221 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.696292 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.696310 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.696336 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.696356 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:24Z","lastTransitionTime":"2026-02-17T13:05:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.800483 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.800544 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.800560 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.800589 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.800610 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:24Z","lastTransitionTime":"2026-02-17T13:05:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.904436 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.904495 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.904514 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.904539 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:24 crc kubenswrapper[4955]: I0217 13:05:24.904557 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:24Z","lastTransitionTime":"2026-02-17T13:05:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.008336 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.008422 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.008437 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.008487 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.008510 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:25Z","lastTransitionTime":"2026-02-17T13:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.111376 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.111435 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.111446 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.111465 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.111480 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:25Z","lastTransitionTime":"2026-02-17T13:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.214394 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.214529 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.214593 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.214635 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.214664 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:25Z","lastTransitionTime":"2026-02-17T13:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.222808 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.222933 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.223068 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:05:25 crc kubenswrapper[4955]: E0217 13:05:25.223186 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:05:25 crc kubenswrapper[4955]: E0217 13:05:25.223309 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:05:25 crc kubenswrapper[4955]: E0217 13:05:25.223394 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.243480 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 11:02:22.754081136 +0000 UTC Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.318101 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.318166 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.318179 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.318201 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.318216 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:25Z","lastTransitionTime":"2026-02-17T13:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.420609 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.420653 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.420663 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.420683 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.420698 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:25Z","lastTransitionTime":"2026-02-17T13:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.523533 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.523599 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.523620 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.523649 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.523669 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:25Z","lastTransitionTime":"2026-02-17T13:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.626917 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.626981 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.626999 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.627025 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.627047 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:25Z","lastTransitionTime":"2026-02-17T13:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.730060 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.730128 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.730147 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.730173 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.730195 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:25Z","lastTransitionTime":"2026-02-17T13:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.834349 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.834402 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.834415 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.834436 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.834450 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:25Z","lastTransitionTime":"2026-02-17T13:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.937352 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.937395 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.937409 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.937428 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:25 crc kubenswrapper[4955]: I0217 13:05:25.937441 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:25Z","lastTransitionTime":"2026-02-17T13:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.040429 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.040491 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.040506 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.040528 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.040544 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:26Z","lastTransitionTime":"2026-02-17T13:05:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.143958 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.144031 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.144046 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.144067 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.144083 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:26Z","lastTransitionTime":"2026-02-17T13:05:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.222852 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:05:26 crc kubenswrapper[4955]: E0217 13:05:26.223055 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.244430 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 07:37:07.1455324 +0000 UTC Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.246604 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.246665 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.246679 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.246699 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.246713 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:26Z","lastTransitionTime":"2026-02-17T13:05:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.349809 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.349863 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.349876 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.349898 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.349915 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:26Z","lastTransitionTime":"2026-02-17T13:05:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.452990 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.453052 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.453065 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.453103 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.453119 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:26Z","lastTransitionTime":"2026-02-17T13:05:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.556829 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.556900 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.556915 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.556938 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.556953 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:26Z","lastTransitionTime":"2026-02-17T13:05:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.659837 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.659932 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.659959 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.659996 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.660022 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:26Z","lastTransitionTime":"2026-02-17T13:05:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.763824 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.763889 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.763906 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.763933 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.763952 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:26Z","lastTransitionTime":"2026-02-17T13:05:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.866639 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.866689 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.866700 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.866721 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.866734 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:26Z","lastTransitionTime":"2026-02-17T13:05:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.970142 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.970194 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.970204 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.970221 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:26 crc kubenswrapper[4955]: I0217 13:05:26.970232 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:26Z","lastTransitionTime":"2026-02-17T13:05:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.073261 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.073322 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.073340 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.073366 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.073388 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:27Z","lastTransitionTime":"2026-02-17T13:05:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.175889 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.175948 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.175963 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.175983 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.175995 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:27Z","lastTransitionTime":"2026-02-17T13:05:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.222720 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.222825 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.222739 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:05:27 crc kubenswrapper[4955]: E0217 13:05:27.223000 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:05:27 crc kubenswrapper[4955]: E0217 13:05:27.223162 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:05:27 crc kubenswrapper[4955]: E0217 13:05:27.223320 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.245460 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 20:49:08.448545997 +0000 UTC Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.279391 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.279429 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.279445 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.279468 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.279483 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:27Z","lastTransitionTime":"2026-02-17T13:05:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.381959 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.382016 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.382035 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.382061 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.382080 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:27Z","lastTransitionTime":"2026-02-17T13:05:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.484485 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.484527 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.484538 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.484555 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.484567 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:27Z","lastTransitionTime":"2026-02-17T13:05:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.586991 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.587037 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.587048 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.587067 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.587078 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:27Z","lastTransitionTime":"2026-02-17T13:05:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.690074 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.690141 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.690155 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.690171 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.690184 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:27Z","lastTransitionTime":"2026-02-17T13:05:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.792087 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.792139 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.792159 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.792183 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.792197 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:27Z","lastTransitionTime":"2026-02-17T13:05:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.894330 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.894368 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.894378 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.894393 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.894403 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:27Z","lastTransitionTime":"2026-02-17T13:05:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.998337 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.998382 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.998394 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.998412 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:27 crc kubenswrapper[4955]: I0217 13:05:27.998425 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:27Z","lastTransitionTime":"2026-02-17T13:05:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.100559 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.100598 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.100606 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.100621 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.100631 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:28Z","lastTransitionTime":"2026-02-17T13:05:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.203827 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.203878 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.203894 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.203914 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.203931 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:28Z","lastTransitionTime":"2026-02-17T13:05:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.222438 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:05:28 crc kubenswrapper[4955]: E0217 13:05:28.222624 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.245953 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 10:56:52.335717209 +0000 UTC Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.306156 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.306196 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.306210 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.306250 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.306266 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:28Z","lastTransitionTime":"2026-02-17T13:05:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.408420 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.408459 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.408471 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.408488 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.408501 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:28Z","lastTransitionTime":"2026-02-17T13:05:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.511120 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.511186 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.511202 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.511227 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.511258 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:28Z","lastTransitionTime":"2026-02-17T13:05:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.616066 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.616123 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.616140 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.616165 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.616183 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:28Z","lastTransitionTime":"2026-02-17T13:05:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.718960 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.719007 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.719021 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.719044 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.719063 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:28Z","lastTransitionTime":"2026-02-17T13:05:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.821353 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.821390 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.821401 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.821416 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.821429 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:28Z","lastTransitionTime":"2026-02-17T13:05:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.924355 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.924443 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.924463 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.924530 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:28 crc kubenswrapper[4955]: I0217 13:05:28.924550 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:28Z","lastTransitionTime":"2026-02-17T13:05:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.026636 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.026672 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.026684 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.026703 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.026717 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:29Z","lastTransitionTime":"2026-02-17T13:05:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.129944 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.129994 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.130005 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.130024 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.130041 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:29Z","lastTransitionTime":"2026-02-17T13:05:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.222502 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.222511 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.222515 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:05:29 crc kubenswrapper[4955]: E0217 13:05:29.222846 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:05:29 crc kubenswrapper[4955]: E0217 13:05:29.222933 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:05:29 crc kubenswrapper[4955]: E0217 13:05:29.223084 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.232302 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.232351 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.232362 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.232378 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.232388 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:29Z","lastTransitionTime":"2026-02-17T13:05:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.246750 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 13:59:16.807894057 +0000 UTC Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.339186 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.339229 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.339240 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.339256 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.339266 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:29Z","lastTransitionTime":"2026-02-17T13:05:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.442499 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.442547 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.442558 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.442577 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.442588 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:29Z","lastTransitionTime":"2026-02-17T13:05:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.545490 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.545539 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.545549 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.545568 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.545581 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:29Z","lastTransitionTime":"2026-02-17T13:05:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.648172 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.648210 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.648220 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.648237 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.648248 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:29Z","lastTransitionTime":"2026-02-17T13:05:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.750887 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.750957 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.750977 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.751003 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.751022 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:29Z","lastTransitionTime":"2026-02-17T13:05:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.853462 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.853518 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.853531 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.853549 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.853560 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:29Z","lastTransitionTime":"2026-02-17T13:05:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.956333 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.956370 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.956379 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.956395 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:29 crc kubenswrapper[4955]: I0217 13:05:29.956409 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:29Z","lastTransitionTime":"2026-02-17T13:05:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.059303 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.059353 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.059367 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.059388 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.059403 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:30Z","lastTransitionTime":"2026-02-17T13:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.162002 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.162042 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.162057 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.162075 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.162086 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:30Z","lastTransitionTime":"2026-02-17T13:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.188555 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.188599 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.188649 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.188672 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.188687 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:30Z","lastTransitionTime":"2026-02-17T13:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:30 crc kubenswrapper[4955]: E0217 13:05:30.207921 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7605a498-360b-4ce4-b7a1-dfbc354d798d\\\",\\\"systemUUID\\\":\\\"6d4602d9-c718-4393-8f91-3c381f800c9e\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:30Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.213515 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.213558 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.213577 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.213597 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.213614 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:30Z","lastTransitionTime":"2026-02-17T13:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.222682 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:05:30 crc kubenswrapper[4955]: E0217 13:05:30.223103 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:05:30 crc kubenswrapper[4955]: E0217 13:05:30.228968 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7605a498-360b-4ce4-b7a1-dfbc354d798d\\\",\\\"systemUUID\\\":\\\"6d4602d9-c718-4393-8f91-3c381f800c9e\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:30Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.234352 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.234394 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.234407 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.234426 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.234439 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:30Z","lastTransitionTime":"2026-02-17T13:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.247695 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 21:36:36.680945223 +0000 UTC Feb 17 13:05:30 crc kubenswrapper[4955]: E0217 13:05:30.248016 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7605a498-360b-4ce4-b7a1-dfbc354d798d\\\",\\\"systemUUID\\\":\\\"6d4602d9-c718-4393-8f91-3c381f800c9e\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:30Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.251906 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.252122 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.252320 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.252520 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.252708 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:30Z","lastTransitionTime":"2026-02-17T13:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:30 crc kubenswrapper[4955]: E0217 13:05:30.268839 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7605a498-360b-4ce4-b7a1-dfbc354d798d\\\",\\\"systemUUID\\\":\\\"6d4602d9-c718-4393-8f91-3c381f800c9e\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:30Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.272572 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.272606 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.272622 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.272640 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.272652 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:30Z","lastTransitionTime":"2026-02-17T13:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:30 crc kubenswrapper[4955]: E0217 13:05:30.286847 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7605a498-360b-4ce4-b7a1-dfbc354d798d\\\",\\\"systemUUID\\\":\\\"6d4602d9-c718-4393-8f91-3c381f800c9e\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:30Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:30 crc kubenswrapper[4955]: E0217 13:05:30.287641 4955 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.289494 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.289704 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.289931 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.290109 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.290264 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:30Z","lastTransitionTime":"2026-02-17T13:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.392997 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.393035 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.393043 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.393056 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.393065 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:30Z","lastTransitionTime":"2026-02-17T13:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.495398 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.495434 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.495447 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.495464 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.495476 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:30Z","lastTransitionTime":"2026-02-17T13:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.499691 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/77480847-3271-4161-a833-e62b3bb4413e-metrics-certs\") pod \"network-metrics-daemon-2k5pz\" (UID: \"77480847-3271-4161-a833-e62b3bb4413e\") " pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:05:30 crc kubenswrapper[4955]: E0217 13:05:30.499890 4955 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 13:05:30 crc kubenswrapper[4955]: E0217 13:05:30.499972 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/77480847-3271-4161-a833-e62b3bb4413e-metrics-certs podName:77480847-3271-4161-a833-e62b3bb4413e nodeName:}" failed. No retries permitted until 2026-02-17 13:06:02.499937504 +0000 UTC m=+101.022667047 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/77480847-3271-4161-a833-e62b3bb4413e-metrics-certs") pod "network-metrics-daemon-2k5pz" (UID: "77480847-3271-4161-a833-e62b3bb4413e") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.597325 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.597361 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.597375 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.597391 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.597404 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:30Z","lastTransitionTime":"2026-02-17T13:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.700667 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.700725 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.700736 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.700754 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.700766 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:30Z","lastTransitionTime":"2026-02-17T13:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.804304 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.804377 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.804394 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.804417 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.804433 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:30Z","lastTransitionTime":"2026-02-17T13:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.907477 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.907522 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.907532 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.907549 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:30 crc kubenswrapper[4955]: I0217 13:05:30.907559 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:30Z","lastTransitionTime":"2026-02-17T13:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.010628 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.010692 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.010713 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.010744 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.010762 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:31Z","lastTransitionTime":"2026-02-17T13:05:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.113959 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.114012 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.114025 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.114044 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.114058 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:31Z","lastTransitionTime":"2026-02-17T13:05:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.217453 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.217886 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.218146 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.218435 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.218670 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:31Z","lastTransitionTime":"2026-02-17T13:05:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.222775 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:05:31 crc kubenswrapper[4955]: E0217 13:05:31.223012 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.223033 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.223314 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:05:31 crc kubenswrapper[4955]: E0217 13:05:31.223819 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:05:31 crc kubenswrapper[4955]: E0217 13:05:31.224168 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.225487 4955 scope.go:117] "RemoveContainer" containerID="29376a23843b866c283dc983b136aead4139d8ce2a97ca45633e17ad6028ced3" Feb 17 13:05:31 crc kubenswrapper[4955]: E0217 13:05:31.225776 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dcwgg_openshift-ovn-kubernetes(dcbc7a69-58e7-4b7a-bb84-96276819ce88)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.248437 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 01:35:06.264401702 +0000 UTC Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.322337 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.322633 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.322887 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.323161 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.323352 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:31Z","lastTransitionTime":"2026-02-17T13:05:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.426215 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.426563 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.426660 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.426749 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.426854 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:31Z","lastTransitionTime":"2026-02-17T13:05:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.530174 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.530510 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.530706 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.530884 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.531031 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:31Z","lastTransitionTime":"2026-02-17T13:05:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.634049 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.634106 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.634120 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.634153 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.634177 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:31Z","lastTransitionTime":"2026-02-17T13:05:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.737326 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.737374 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.737387 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.737407 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.737420 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:31Z","lastTransitionTime":"2026-02-17T13:05:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.840750 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.840809 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.840821 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.840838 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.840849 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:31Z","lastTransitionTime":"2026-02-17T13:05:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.943199 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.943257 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.943269 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.943296 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:31 crc kubenswrapper[4955]: I0217 13:05:31.943312 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:31Z","lastTransitionTime":"2026-02-17T13:05:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.046437 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.046479 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.046492 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.046509 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.046522 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:32Z","lastTransitionTime":"2026-02-17T13:05:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.149535 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.149589 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.149605 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.149629 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.149642 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:32Z","lastTransitionTime":"2026-02-17T13:05:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.222273 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:05:32 crc kubenswrapper[4955]: E0217 13:05:32.222471 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.240628 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:32Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.249518 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-04 03:16:41.154669079 +0000 UTC Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.252826 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.252964 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.253040 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.253119 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.253180 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:32Z","lastTransitionTime":"2026-02-17T13:05:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.255949 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:32Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.269905 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bf36887-2a2b-4193-abdb-a15beaf46300\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11a8924eb6ce41d629feb2b42c7d4778a3ca9110a59c3ac25acbba4367272374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bec78f9e9f6416d6b600f6231f0d24652473b48b95bf0e0675cc85c9208d124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-29qxq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:32Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.286353 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9hx9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e57bdf8-e537-402a-9815-609fecdd68ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbc9014d80aee39922161014aac8fafa7d93be3e4092fe705c76da4a487a5e5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q22d4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9hx9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:32Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.299754 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-msn6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://277a12d62c90ef420c31d8dc71f0d7750dfd5f5ab7e90fec0dcd00e54b60245d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2tvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-msn6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:32Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.314847 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05aa39d-94e6-49d4-885e-2c1f986310f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T13:04:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 13:04:35.885968 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 13:04:35.888176 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3462920193/tls.crt::/tmp/serving-cert-3462920193/tls.key\\\\\\\"\\\\nI0217 13:04:42.370546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 13:04:42.384242 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 13:04:42.384284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 13:04:42.384317 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 13:04:42.384327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 13:04:42.394555 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 13:04:42.394614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394628 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 13:04:42.394646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 13:04:42.394654 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 13:04:42.394662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 13:04:42.394900 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 13:04:42.396494 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:32Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.329226 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5c34a3d-82ad-4d03-abdd-0a20818ac2fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0261f41b82b492d324b5e50edd72cedd5e824de542a7b47104abd2ab757a7049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f43c7fe16acfd4d39301ee64665d4cd1d885397f981d70f73215da7f3aabe72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c871c1b454e4edd2f41076ecfd6a66592023d7161f094fbb4fd197fbf9c881f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:32Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.341348 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:32Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.355603 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.355902 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.356004 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.356107 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.356203 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:32Z","lastTransitionTime":"2026-02-17T13:05:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.357934 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"924c66b7-7c3c-4db7-a427-3f152efbac4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9beee007a179c3c2e04c2ce0de8ea6dbae9da3421ab24f23b718a23744c828a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fp5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:32Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.378380 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcbc7a69-58e7-4b7a-bb84-96276819ce88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29376a23843b866c283dc983b136aead4139d8ce2a97ca45633e17ad6028ced3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29376a23843b866c283dc983b136aead4139d8ce2a97ca45633e17ad6028ced3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T13:05:16Z\\\",\\\"message\\\":\\\"licy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0217 13:05:16.060354 6583 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0217 13:05:16.060455 6583 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0217 13:05:16.060913 6583 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0217 13:05:16.061605 6583 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0217 13:05:16.061682 6583 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0217 13:05:16.061709 6583 factory.go:656] Stopping watch factory\\\\nI0217 13:05:16.061733 6583 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0217 13:05:16.061766 6583 handler.go:208] Removed *v1.Node event handler 2\\\\nI0217 13:05:16.088613 6583 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0217 13:05:16.088662 6583 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0217 13:05:16.088736 6583 ovnkube.go:599] Stopped ovnkube\\\\nI0217 13:05:16.088764 6583 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0217 13:05:16.088897 6583 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:05:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dcwgg_openshift-ovn-kubernetes(dcbc7a69-58e7-4b7a-bb84-96276819ce88)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dcwgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:32Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.402712 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b49c9af-148a-48ec-af37-f555f9186337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d6919c91918e8bacafd1442b06d14ffd9b547f0f37b2298e2d8d3acc5962d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e38466a067b20ca4323ab6983cb324bb63179f942d2c37c0daa8b03767e7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4830450cbada3d24c9747e9de5f48e53e118549f1dd67edf7fff8b2adc47128f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001dfb595f39d1da33b265474b6cca51c62f522ff061c9e7c52f93806034f4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f1fee7ac5f8d1442d9f63b0e8d285187129eb3a3e3b50a49a45c7a7a39ae8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:32Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.417532 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d0966253f623158053125dffca956000824438bdc548eba6dafe9a162d5674a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d07bfb361f14d817913fa2cabd30649447790068545dda2f3a2f71213ea0e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:32Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.427509 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2gsx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1face148-45d1-44b3-8775-209f9f8651fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82c5147255d021114a8639e9b60593f847896b613b5d862db588014f7485faa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnlcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2gsx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:32Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.439885 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lffhw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8155b25d-be0c-44ed-b3af-7bf25d55ec83\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad540421c8104f89cf02f413145e08de7c3e65e47f78cb2be5b40f746d6376c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gw5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52f6c4f214dd72b2df8ba07e3448bb4a2bedc698f47c16d027a774e2723c201\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gw5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lffhw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:32Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.459078 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.459070 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2k5pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77480847-3271-4161-a833-e62b3bb4413e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fffmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fffmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2k5pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:32Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.459163 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.459576 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.459641 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.459661 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:32Z","lastTransitionTime":"2026-02-17T13:05:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.473851 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aea1a9-3570-4339-a1a4-db2dc358328f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3285726bf736fdd9d99c9974bef45c180410c4e3b795acbbf6c4ac576be3ee9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c8e77568b5ec88d39510e2cdf7b9a3438a9c7cd0ad83cb350c9d7c927f0dbc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bbbc1ab7b069d1abd31929ec33bd494a7f403437f0be70da380ea0a5ff2e711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25f110bf0e9d951f4d48a491daae897a8250cf77fe464f1914cb2f1cfcde1d72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25f110bf0e9d951f4d48a491daae897a8250cf77fe464f1914cb2f1cfcde1d72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:32Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.488693 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9ecd4380a6ae8ff62e27e3946815170c2f02b4c3921bda595467345289c9e0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:32Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.506281 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0f4056e4b5a554f05004c194485d25231b67024e4376a73c715fcb32004a3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:32Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.563377 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.563442 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.563454 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.563477 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.563491 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:32Z","lastTransitionTime":"2026-02-17T13:05:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.665893 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.665944 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.665954 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.665973 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.665985 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:32Z","lastTransitionTime":"2026-02-17T13:05:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.765283 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9hx9f_1e57bdf8-e537-402a-9815-609fecdd68ea/kube-multus/0.log" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.765345 4955 generic.go:334] "Generic (PLEG): container finished" podID="1e57bdf8-e537-402a-9815-609fecdd68ea" containerID="dbc9014d80aee39922161014aac8fafa7d93be3e4092fe705c76da4a487a5e5d" exitCode=1 Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.765384 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-9hx9f" event={"ID":"1e57bdf8-e537-402a-9815-609fecdd68ea","Type":"ContainerDied","Data":"dbc9014d80aee39922161014aac8fafa7d93be3e4092fe705c76da4a487a5e5d"} Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.766942 4955 scope.go:117] "RemoveContainer" containerID="dbc9014d80aee39922161014aac8fafa7d93be3e4092fe705c76da4a487a5e5d" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.768762 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.768817 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.768829 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.768844 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.768855 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:32Z","lastTransitionTime":"2026-02-17T13:05:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.779181 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:32Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.795832 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"924c66b7-7c3c-4db7-a427-3f152efbac4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9beee007a179c3c2e04c2ce0de8ea6dbae9da3421ab24f23b718a23744c828a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fp5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:32Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.817006 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcbc7a69-58e7-4b7a-bb84-96276819ce88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29376a23843b866c283dc983b136aead4139d8ce2a97ca45633e17ad6028ced3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29376a23843b866c283dc983b136aead4139d8ce2a97ca45633e17ad6028ced3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T13:05:16Z\\\",\\\"message\\\":\\\"licy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0217 13:05:16.060354 6583 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0217 13:05:16.060455 6583 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0217 13:05:16.060913 6583 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0217 13:05:16.061605 6583 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0217 13:05:16.061682 6583 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0217 13:05:16.061709 6583 factory.go:656] Stopping watch factory\\\\nI0217 13:05:16.061733 6583 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0217 13:05:16.061766 6583 handler.go:208] Removed *v1.Node event handler 2\\\\nI0217 13:05:16.088613 6583 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0217 13:05:16.088662 6583 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0217 13:05:16.088736 6583 ovnkube.go:599] Stopped ovnkube\\\\nI0217 13:05:16.088764 6583 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0217 13:05:16.088897 6583 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:05:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dcwgg_openshift-ovn-kubernetes(dcbc7a69-58e7-4b7a-bb84-96276819ce88)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dcwgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:32Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.839469 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b49c9af-148a-48ec-af37-f555f9186337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d6919c91918e8bacafd1442b06d14ffd9b547f0f37b2298e2d8d3acc5962d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e38466a067b20ca4323ab6983cb324bb63179f942d2c37c0daa8b03767e7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4830450cbada3d24c9747e9de5f48e53e118549f1dd67edf7fff8b2adc47128f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001dfb595f39d1da33b265474b6cca51c62f522ff061c9e7c52f93806034f4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f1fee7ac5f8d1442d9f63b0e8d285187129eb3a3e3b50a49a45c7a7a39ae8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:32Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.862901 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5c34a3d-82ad-4d03-abdd-0a20818ac2fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0261f41b82b492d324b5e50edd72cedd5e824de542a7b47104abd2ab757a7049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f43c7fe16acfd4d39301ee64665d4cd1d885397f981d70f73215da7f3aabe72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c871c1b454e4edd2f41076ecfd6a66592023d7161f094fbb4fd197fbf9c881f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:32Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.871919 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.871950 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.871960 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.871977 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.871990 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:32Z","lastTransitionTime":"2026-02-17T13:05:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.876333 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2gsx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1face148-45d1-44b3-8775-209f9f8651fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82c5147255d021114a8639e9b60593f847896b613b5d862db588014f7485faa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnlcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2gsx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:32Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.890186 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lffhw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8155b25d-be0c-44ed-b3af-7bf25d55ec83\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad540421c8104f89cf02f413145e08de7c3e65e47f78cb2be5b40f746d6376c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gw5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52f6c4f214dd72b2df8ba07e3448bb4a2bedc698f47c16d027a774e2723c201\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gw5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lffhw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:32Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.901356 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2k5pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77480847-3271-4161-a833-e62b3bb4413e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fffmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fffmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2k5pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:32Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.912393 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aea1a9-3570-4339-a1a4-db2dc358328f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3285726bf736fdd9d99c9974bef45c180410c4e3b795acbbf6c4ac576be3ee9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c8e77568b5ec88d39510e2cdf7b9a3438a9c7cd0ad83cb350c9d7c927f0dbc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bbbc1ab7b069d1abd31929ec33bd494a7f403437f0be70da380ea0a5ff2e711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25f110bf0e9d951f4d48a491daae897a8250cf77fe464f1914cb2f1cfcde1d72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25f110bf0e9d951f4d48a491daae897a8250cf77fe464f1914cb2f1cfcde1d72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:32Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.924661 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d0966253f623158053125dffca956000824438bdc548eba6dafe9a162d5674a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d07bfb361f14d817913fa2cabd30649447790068545dda2f3a2f71213ea0e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:32Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.934712 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0f4056e4b5a554f05004c194485d25231b67024e4376a73c715fcb32004a3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:32Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.945849 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9ecd4380a6ae8ff62e27e3946815170c2f02b4c3921bda595467345289c9e0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:32Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.955785 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:32Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.965236 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bf36887-2a2b-4193-abdb-a15beaf46300\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11a8924eb6ce41d629feb2b42c7d4778a3ca9110a59c3ac25acbba4367272374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bec78f9e9f6416d6b600f6231f0d24652473b48b95bf0e0675cc85c9208d124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-29qxq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:32Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.974877 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.974921 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.974931 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.974949 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.974969 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:32Z","lastTransitionTime":"2026-02-17T13:05:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.978277 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9hx9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e57bdf8-e537-402a-9815-609fecdd68ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbc9014d80aee39922161014aac8fafa7d93be3e4092fe705c76da4a487a5e5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbc9014d80aee39922161014aac8fafa7d93be3e4092fe705c76da4a487a5e5d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T13:05:31Z\\\",\\\"message\\\":\\\"2026-02-17T13:04:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c2764cb0-2a4e-4449-bac5-794fed2f307a\\\\n2026-02-17T13:04:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c2764cb0-2a4e-4449-bac5-794fed2f307a to /host/opt/cni/bin/\\\\n2026-02-17T13:04:46Z [verbose] multus-daemon started\\\\n2026-02-17T13:04:46Z [verbose] Readiness Indicator file check\\\\n2026-02-17T13:05:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q22d4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9hx9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:32Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:32 crc kubenswrapper[4955]: I0217 13:05:32.988414 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-msn6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://277a12d62c90ef420c31d8dc71f0d7750dfd5f5ab7e90fec0dcd00e54b60245d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2tvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-msn6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:32Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.004099 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05aa39d-94e6-49d4-885e-2c1f986310f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T13:04:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 13:04:35.885968 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 13:04:35.888176 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3462920193/tls.crt::/tmp/serving-cert-3462920193/tls.key\\\\\\\"\\\\nI0217 13:04:42.370546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 13:04:42.384242 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 13:04:42.384284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 13:04:42.384317 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 13:04:42.384327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 13:04:42.394555 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 13:04:42.394614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394628 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 13:04:42.394646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 13:04:42.394654 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 13:04:42.394662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 13:04:42.394900 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 13:04:42.396494 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:33Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.016840 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:33Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.077490 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.077531 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.077543 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.077560 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.077574 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:33Z","lastTransitionTime":"2026-02-17T13:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.180398 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.180451 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.180462 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.180481 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.180494 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:33Z","lastTransitionTime":"2026-02-17T13:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.222446 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.222481 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.222496 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:05:33 crc kubenswrapper[4955]: E0217 13:05:33.222613 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:05:33 crc kubenswrapper[4955]: E0217 13:05:33.222894 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:05:33 crc kubenswrapper[4955]: E0217 13:05:33.223041 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.250958 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 05:43:52.696745094 +0000 UTC Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.282564 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.282600 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.282612 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.282632 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.282645 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:33Z","lastTransitionTime":"2026-02-17T13:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.385095 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.385140 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.385153 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.385172 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.385185 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:33Z","lastTransitionTime":"2026-02-17T13:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.488441 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.488496 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.488512 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.488537 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.488556 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:33Z","lastTransitionTime":"2026-02-17T13:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.592018 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.592090 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.592104 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.592126 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.592139 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:33Z","lastTransitionTime":"2026-02-17T13:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.695337 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.695392 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.695405 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.695426 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.695440 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:33Z","lastTransitionTime":"2026-02-17T13:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.772075 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9hx9f_1e57bdf8-e537-402a-9815-609fecdd68ea/kube-multus/0.log" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.772135 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-9hx9f" event={"ID":"1e57bdf8-e537-402a-9815-609fecdd68ea","Type":"ContainerStarted","Data":"c81948fc2806ea8718909fd269c029efc6a26675e45359e971272354eccfa74d"} Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.791272 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05aa39d-94e6-49d4-885e-2c1f986310f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T13:04:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 13:04:35.885968 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 13:04:35.888176 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3462920193/tls.crt::/tmp/serving-cert-3462920193/tls.key\\\\\\\"\\\\nI0217 13:04:42.370546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 13:04:42.384242 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 13:04:42.384284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 13:04:42.384317 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 13:04:42.384327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 13:04:42.394555 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 13:04:42.394614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394628 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 13:04:42.394646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 13:04:42.394654 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 13:04:42.394662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 13:04:42.394900 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 13:04:42.396494 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:33Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.797257 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.797314 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.797325 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.797345 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.797356 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:33Z","lastTransitionTime":"2026-02-17T13:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.807388 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:33Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.819518 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:33Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.832646 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bf36887-2a2b-4193-abdb-a15beaf46300\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11a8924eb6ce41d629feb2b42c7d4778a3ca9110a59c3ac25acbba4367272374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bec78f9e9f6416d6b600f6231f0d24652473b48b95bf0e0675cc85c9208d124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-29qxq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:33Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.849633 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9hx9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e57bdf8-e537-402a-9815-609fecdd68ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c81948fc2806ea8718909fd269c029efc6a26675e45359e971272354eccfa74d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbc9014d80aee39922161014aac8fafa7d93be3e4092fe705c76da4a487a5e5d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T13:05:31Z\\\",\\\"message\\\":\\\"2026-02-17T13:04:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c2764cb0-2a4e-4449-bac5-794fed2f307a\\\\n2026-02-17T13:04:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c2764cb0-2a4e-4449-bac5-794fed2f307a to /host/opt/cni/bin/\\\\n2026-02-17T13:04:46Z [verbose] multus-daemon started\\\\n2026-02-17T13:04:46Z [verbose] Readiness Indicator file check\\\\n2026-02-17T13:05:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:05:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q22d4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9hx9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:33Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.861025 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-msn6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://277a12d62c90ef420c31d8dc71f0d7750dfd5f5ab7e90fec0dcd00e54b60245d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2tvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-msn6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:33Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.880706 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b49c9af-148a-48ec-af37-f555f9186337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d6919c91918e8bacafd1442b06d14ffd9b547f0f37b2298e2d8d3acc5962d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e38466a067b20ca4323ab6983cb324bb63179f942d2c37c0daa8b03767e7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4830450cbada3d24c9747e9de5f48e53e118549f1dd67edf7fff8b2adc47128f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001dfb595f39d1da33b265474b6cca51c62f522ff061c9e7c52f93806034f4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f1fee7ac5f8d1442d9f63b0e8d285187129eb3a3e3b50a49a45c7a7a39ae8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:33Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.895845 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5c34a3d-82ad-4d03-abdd-0a20818ac2fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0261f41b82b492d324b5e50edd72cedd5e824de542a7b47104abd2ab757a7049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f43c7fe16acfd4d39301ee64665d4cd1d885397f981d70f73215da7f3aabe72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c871c1b454e4edd2f41076ecfd6a66592023d7161f094fbb4fd197fbf9c881f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:33Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.899841 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.899873 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.899881 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.899895 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.899904 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:33Z","lastTransitionTime":"2026-02-17T13:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.912061 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:33Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.926314 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"924c66b7-7c3c-4db7-a427-3f152efbac4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9beee007a179c3c2e04c2ce0de8ea6dbae9da3421ab24f23b718a23744c828a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fp5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:33Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.944302 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcbc7a69-58e7-4b7a-bb84-96276819ce88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29376a23843b866c283dc983b136aead4139d8ce2a97ca45633e17ad6028ced3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29376a23843b866c283dc983b136aead4139d8ce2a97ca45633e17ad6028ced3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T13:05:16Z\\\",\\\"message\\\":\\\"licy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0217 13:05:16.060354 6583 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0217 13:05:16.060455 6583 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0217 13:05:16.060913 6583 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0217 13:05:16.061605 6583 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0217 13:05:16.061682 6583 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0217 13:05:16.061709 6583 factory.go:656] Stopping watch factory\\\\nI0217 13:05:16.061733 6583 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0217 13:05:16.061766 6583 handler.go:208] Removed *v1.Node event handler 2\\\\nI0217 13:05:16.088613 6583 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0217 13:05:16.088662 6583 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0217 13:05:16.088736 6583 ovnkube.go:599] Stopped ovnkube\\\\nI0217 13:05:16.088764 6583 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0217 13:05:16.088897 6583 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:05:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dcwgg_openshift-ovn-kubernetes(dcbc7a69-58e7-4b7a-bb84-96276819ce88)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dcwgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:33Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.958074 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aea1a9-3570-4339-a1a4-db2dc358328f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3285726bf736fdd9d99c9974bef45c180410c4e3b795acbbf6c4ac576be3ee9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c8e77568b5ec88d39510e2cdf7b9a3438a9c7cd0ad83cb350c9d7c927f0dbc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bbbc1ab7b069d1abd31929ec33bd494a7f403437f0be70da380ea0a5ff2e711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25f110bf0e9d951f4d48a491daae897a8250cf77fe464f1914cb2f1cfcde1d72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25f110bf0e9d951f4d48a491daae897a8250cf77fe464f1914cb2f1cfcde1d72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:33Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.971439 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d0966253f623158053125dffca956000824438bdc548eba6dafe9a162d5674a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d07bfb361f14d817913fa2cabd30649447790068545dda2f3a2f71213ea0e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:33Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.982037 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2gsx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1face148-45d1-44b3-8775-209f9f8651fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82c5147255d021114a8639e9b60593f847896b613b5d862db588014f7485faa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnlcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2gsx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:33Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:33 crc kubenswrapper[4955]: I0217 13:05:33.992930 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lffhw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8155b25d-be0c-44ed-b3af-7bf25d55ec83\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad540421c8104f89cf02f413145e08de7c3e65e47f78cb2be5b40f746d6376c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gw5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52f6c4f214dd72b2df8ba07e3448bb4a2bedc698f47c16d027a774e2723c201\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gw5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lffhw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:33Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.003637 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.003679 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.003508 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2k5pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77480847-3271-4161-a833-e62b3bb4413e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fffmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fffmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2k5pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:34Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.003692 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.003864 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.003880 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:34Z","lastTransitionTime":"2026-02-17T13:05:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.021065 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9ecd4380a6ae8ff62e27e3946815170c2f02b4c3921bda595467345289c9e0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:34Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.034632 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0f4056e4b5a554f05004c194485d25231b67024e4376a73c715fcb32004a3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:34Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.106027 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.106080 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.106094 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.106112 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.106122 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:34Z","lastTransitionTime":"2026-02-17T13:05:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.209569 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.209666 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.209683 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.209718 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.209738 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:34Z","lastTransitionTime":"2026-02-17T13:05:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.222278 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:05:34 crc kubenswrapper[4955]: E0217 13:05:34.222529 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.251496 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 15:41:39.135241261 +0000 UTC Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.313076 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.313201 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.313277 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.313306 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.313327 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:34Z","lastTransitionTime":"2026-02-17T13:05:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.416593 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.416653 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.416667 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.416685 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.416698 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:34Z","lastTransitionTime":"2026-02-17T13:05:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.519683 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.519747 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.519758 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.519777 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.519806 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:34Z","lastTransitionTime":"2026-02-17T13:05:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.624403 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.624466 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.624480 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.624502 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.624513 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:34Z","lastTransitionTime":"2026-02-17T13:05:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.728894 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.728952 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.728972 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.728995 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.729012 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:34Z","lastTransitionTime":"2026-02-17T13:05:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.832102 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.832137 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.832149 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.832167 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.832180 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:34Z","lastTransitionTime":"2026-02-17T13:05:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.938993 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.939043 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.939052 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.939070 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:34 crc kubenswrapper[4955]: I0217 13:05:34.939081 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:34Z","lastTransitionTime":"2026-02-17T13:05:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.042231 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.042302 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.042325 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.042354 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.042373 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:35Z","lastTransitionTime":"2026-02-17T13:05:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.144986 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.145040 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.145058 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.145087 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.145105 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:35Z","lastTransitionTime":"2026-02-17T13:05:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.223029 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.223037 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.223289 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:05:35 crc kubenswrapper[4955]: E0217 13:05:35.223419 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:05:35 crc kubenswrapper[4955]: E0217 13:05:35.223658 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:05:35 crc kubenswrapper[4955]: E0217 13:05:35.223752 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.248184 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.248241 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.248251 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.248272 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.248283 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:35Z","lastTransitionTime":"2026-02-17T13:05:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.251689 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 06:16:22.714058957 +0000 UTC Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.352006 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.352067 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.352084 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.352111 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.352128 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:35Z","lastTransitionTime":"2026-02-17T13:05:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.455041 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.455109 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.455120 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.455140 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.455151 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:35Z","lastTransitionTime":"2026-02-17T13:05:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.557910 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.557961 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.557975 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.557992 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.558005 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:35Z","lastTransitionTime":"2026-02-17T13:05:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.660639 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.660696 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.660709 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.660729 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.660744 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:35Z","lastTransitionTime":"2026-02-17T13:05:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.763632 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.763671 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.763680 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.763693 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.763702 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:35Z","lastTransitionTime":"2026-02-17T13:05:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.866219 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.866268 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.866281 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.866302 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.866315 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:35Z","lastTransitionTime":"2026-02-17T13:05:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.968167 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.968215 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.968227 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.968242 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:35 crc kubenswrapper[4955]: I0217 13:05:35.968274 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:35Z","lastTransitionTime":"2026-02-17T13:05:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.071264 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.071336 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.071354 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.071385 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.071404 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:36Z","lastTransitionTime":"2026-02-17T13:05:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.174753 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.174819 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.174829 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.174847 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.174858 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:36Z","lastTransitionTime":"2026-02-17T13:05:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.222596 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:05:36 crc kubenswrapper[4955]: E0217 13:05:36.222742 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.252488 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 17:18:08.365406693 +0000 UTC Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.276976 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.277047 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.277068 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.277102 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.277127 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:36Z","lastTransitionTime":"2026-02-17T13:05:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.380703 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.380772 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.380796 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.380852 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.380871 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:36Z","lastTransitionTime":"2026-02-17T13:05:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.483095 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.483162 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.483179 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.483203 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.483221 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:36Z","lastTransitionTime":"2026-02-17T13:05:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.587205 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.587270 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.587286 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.587308 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.587335 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:36Z","lastTransitionTime":"2026-02-17T13:05:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.690104 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.690154 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.690165 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.690187 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.690204 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:36Z","lastTransitionTime":"2026-02-17T13:05:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.791796 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.791847 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.791869 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.791886 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.791896 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:36Z","lastTransitionTime":"2026-02-17T13:05:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.894586 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.894626 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.894638 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.894656 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.894673 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:36Z","lastTransitionTime":"2026-02-17T13:05:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.997602 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.997636 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.997644 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.997659 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:36 crc kubenswrapper[4955]: I0217 13:05:36.997669 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:36Z","lastTransitionTime":"2026-02-17T13:05:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.100723 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.100756 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.100764 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.100782 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.100794 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:37Z","lastTransitionTime":"2026-02-17T13:05:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.202424 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.202460 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.202468 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.202485 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.202496 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:37Z","lastTransitionTime":"2026-02-17T13:05:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.222493 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:05:37 crc kubenswrapper[4955]: E0217 13:05:37.222647 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.222817 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:05:37 crc kubenswrapper[4955]: E0217 13:05:37.222878 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.223036 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:05:37 crc kubenswrapper[4955]: E0217 13:05:37.223085 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.253542 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 04:47:16.710824553 +0000 UTC Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.305102 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.305140 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.305149 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.305163 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.305175 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:37Z","lastTransitionTime":"2026-02-17T13:05:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.407468 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.407515 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.407533 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.407555 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.407568 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:37Z","lastTransitionTime":"2026-02-17T13:05:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.510409 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.510480 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.510496 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.510526 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.510543 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:37Z","lastTransitionTime":"2026-02-17T13:05:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.614352 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.614408 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.614419 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.614435 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.614449 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:37Z","lastTransitionTime":"2026-02-17T13:05:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.717271 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.717345 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.717369 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.717399 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.717417 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:37Z","lastTransitionTime":"2026-02-17T13:05:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.820609 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.820680 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.820705 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.820738 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.820769 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:37Z","lastTransitionTime":"2026-02-17T13:05:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.925987 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.926054 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.926068 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.926110 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:37 crc kubenswrapper[4955]: I0217 13:05:37.926128 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:37Z","lastTransitionTime":"2026-02-17T13:05:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.035174 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.035259 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.035294 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.035328 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.035351 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:38Z","lastTransitionTime":"2026-02-17T13:05:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.138124 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.138193 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.138215 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.138246 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.138270 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:38Z","lastTransitionTime":"2026-02-17T13:05:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.222117 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:05:38 crc kubenswrapper[4955]: E0217 13:05:38.222325 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.240884 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.240926 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.240943 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.240964 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.240982 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:38Z","lastTransitionTime":"2026-02-17T13:05:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.254140 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 19:06:38.562609532 +0000 UTC Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.358288 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.358327 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.358339 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.358356 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.358368 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:38Z","lastTransitionTime":"2026-02-17T13:05:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.461073 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.461118 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.461128 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.461146 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.461158 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:38Z","lastTransitionTime":"2026-02-17T13:05:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.563592 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.563632 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.563641 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.563657 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.563669 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:38Z","lastTransitionTime":"2026-02-17T13:05:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.666325 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.666374 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.666388 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.666407 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.666420 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:38Z","lastTransitionTime":"2026-02-17T13:05:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.769399 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.769443 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.769460 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.769482 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.769498 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:38Z","lastTransitionTime":"2026-02-17T13:05:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.871681 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.871741 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.871759 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.871830 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.871855 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:38Z","lastTransitionTime":"2026-02-17T13:05:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.975167 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.975264 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.975281 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.975310 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:38 crc kubenswrapper[4955]: I0217 13:05:38.975332 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:38Z","lastTransitionTime":"2026-02-17T13:05:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.078476 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.078522 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.078534 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.078554 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.078565 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:39Z","lastTransitionTime":"2026-02-17T13:05:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.181768 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.181819 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.181829 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.181843 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.181853 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:39Z","lastTransitionTime":"2026-02-17T13:05:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.222995 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.223043 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.223168 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:05:39 crc kubenswrapper[4955]: E0217 13:05:39.223268 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:05:39 crc kubenswrapper[4955]: E0217 13:05:39.223436 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:05:39 crc kubenswrapper[4955]: E0217 13:05:39.223540 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.254424 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 12:46:50.670691284 +0000 UTC Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.285064 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.285098 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.285110 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.285127 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.285143 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:39Z","lastTransitionTime":"2026-02-17T13:05:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.387631 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.387665 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.387678 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.387695 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.387705 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:39Z","lastTransitionTime":"2026-02-17T13:05:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.490641 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.490687 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.490699 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.490719 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.490732 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:39Z","lastTransitionTime":"2026-02-17T13:05:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.593411 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.593441 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.593451 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.593467 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.593478 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:39Z","lastTransitionTime":"2026-02-17T13:05:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.696509 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.696574 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.696596 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.696624 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.696654 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:39Z","lastTransitionTime":"2026-02-17T13:05:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.798426 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.798469 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.798479 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.798493 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.798502 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:39Z","lastTransitionTime":"2026-02-17T13:05:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.901732 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.901857 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.901877 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.901909 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:39 crc kubenswrapper[4955]: I0217 13:05:39.901934 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:39Z","lastTransitionTime":"2026-02-17T13:05:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.005749 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.005884 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.005912 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.005947 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.005971 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:40Z","lastTransitionTime":"2026-02-17T13:05:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.108877 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.108963 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.109002 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.109035 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.109057 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:40Z","lastTransitionTime":"2026-02-17T13:05:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.211127 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.211201 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.211226 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.211258 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.211283 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:40Z","lastTransitionTime":"2026-02-17T13:05:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.222476 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:05:40 crc kubenswrapper[4955]: E0217 13:05:40.222668 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.254837 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 17:28:38.986180096 +0000 UTC Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.314303 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.314351 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.314367 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.314410 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.314424 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:40Z","lastTransitionTime":"2026-02-17T13:05:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.417123 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.417157 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.417168 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.417187 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.417201 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:40Z","lastTransitionTime":"2026-02-17T13:05:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.502023 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.502089 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.502110 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.502141 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.502163 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:40Z","lastTransitionTime":"2026-02-17T13:05:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:40 crc kubenswrapper[4955]: E0217 13:05:40.527138 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7605a498-360b-4ce4-b7a1-dfbc354d798d\\\",\\\"systemUUID\\\":\\\"6d4602d9-c718-4393-8f91-3c381f800c9e\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:40Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.532369 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.532419 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.532437 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.532461 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.532477 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:40Z","lastTransitionTime":"2026-02-17T13:05:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:40 crc kubenswrapper[4955]: E0217 13:05:40.549112 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7605a498-360b-4ce4-b7a1-dfbc354d798d\\\",\\\"systemUUID\\\":\\\"6d4602d9-c718-4393-8f91-3c381f800c9e\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:40Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.554658 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.554710 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.554729 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.554754 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.554770 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:40Z","lastTransitionTime":"2026-02-17T13:05:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:40 crc kubenswrapper[4955]: E0217 13:05:40.571426 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7605a498-360b-4ce4-b7a1-dfbc354d798d\\\",\\\"systemUUID\\\":\\\"6d4602d9-c718-4393-8f91-3c381f800c9e\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:40Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.576229 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.576280 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.576296 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.576320 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.576339 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:40Z","lastTransitionTime":"2026-02-17T13:05:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:40 crc kubenswrapper[4955]: E0217 13:05:40.594392 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7605a498-360b-4ce4-b7a1-dfbc354d798d\\\",\\\"systemUUID\\\":\\\"6d4602d9-c718-4393-8f91-3c381f800c9e\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:40Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.599305 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.599346 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.599381 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.599399 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.599415 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:40Z","lastTransitionTime":"2026-02-17T13:05:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:40 crc kubenswrapper[4955]: E0217 13:05:40.617100 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7605a498-360b-4ce4-b7a1-dfbc354d798d\\\",\\\"systemUUID\\\":\\\"6d4602d9-c718-4393-8f91-3c381f800c9e\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:40Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:40 crc kubenswrapper[4955]: E0217 13:05:40.617453 4955 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.619489 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.619547 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.619561 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.619625 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.619636 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:40Z","lastTransitionTime":"2026-02-17T13:05:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.724309 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.724408 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.724499 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.724717 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.724868 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:40Z","lastTransitionTime":"2026-02-17T13:05:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.828775 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.828892 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.828915 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.828947 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.828973 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:40Z","lastTransitionTime":"2026-02-17T13:05:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.933310 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.933350 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.933362 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.933381 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:40 crc kubenswrapper[4955]: I0217 13:05:40.933392 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:40Z","lastTransitionTime":"2026-02-17T13:05:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.037115 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.037172 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.037189 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.037214 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.037232 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:41Z","lastTransitionTime":"2026-02-17T13:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.140853 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.140909 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.140920 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.140942 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.140953 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:41Z","lastTransitionTime":"2026-02-17T13:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.222902 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.222930 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.222950 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:05:41 crc kubenswrapper[4955]: E0217 13:05:41.224409 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:05:41 crc kubenswrapper[4955]: E0217 13:05:41.224178 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:05:41 crc kubenswrapper[4955]: E0217 13:05:41.224486 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.244912 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.245004 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.245032 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.245066 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.245091 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:41Z","lastTransitionTime":"2026-02-17T13:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.256088 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 00:41:26.173082268 +0000 UTC Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.347965 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.348029 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.348045 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.348069 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.348088 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:41Z","lastTransitionTime":"2026-02-17T13:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.451287 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.451331 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.451342 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.451361 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.451373 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:41Z","lastTransitionTime":"2026-02-17T13:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.554408 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.554445 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.554454 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.554468 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.554480 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:41Z","lastTransitionTime":"2026-02-17T13:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.657730 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.657838 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.657857 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.657883 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.657903 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:41Z","lastTransitionTime":"2026-02-17T13:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.761395 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.761465 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.761484 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.761512 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.761529 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:41Z","lastTransitionTime":"2026-02-17T13:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.864440 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.864525 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.864546 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.864577 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.864601 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:41Z","lastTransitionTime":"2026-02-17T13:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.967841 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.967895 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.967907 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.967926 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:41 crc kubenswrapper[4955]: I0217 13:05:41.967941 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:41Z","lastTransitionTime":"2026-02-17T13:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.071327 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.071377 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.071390 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.071410 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.071421 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:42Z","lastTransitionTime":"2026-02-17T13:05:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.174899 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.174959 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.174972 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.175025 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.175041 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:42Z","lastTransitionTime":"2026-02-17T13:05:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.222373 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:05:42 crc kubenswrapper[4955]: E0217 13:05:42.222562 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.242222 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9hx9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e57bdf8-e537-402a-9815-609fecdd68ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c81948fc2806ea8718909fd269c029efc6a26675e45359e971272354eccfa74d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbc9014d80aee39922161014aac8fafa7d93be3e4092fe705c76da4a487a5e5d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T13:05:31Z\\\",\\\"message\\\":\\\"2026-02-17T13:04:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c2764cb0-2a4e-4449-bac5-794fed2f307a\\\\n2026-02-17T13:04:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c2764cb0-2a4e-4449-bac5-794fed2f307a to /host/opt/cni/bin/\\\\n2026-02-17T13:04:46Z [verbose] multus-daemon started\\\\n2026-02-17T13:04:46Z [verbose] Readiness Indicator file check\\\\n2026-02-17T13:05:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:05:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q22d4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9hx9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:42Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.255653 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-msn6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://277a12d62c90ef420c31d8dc71f0d7750dfd5f5ab7e90fec0dcd00e54b60245d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2tvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-msn6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:42Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.256703 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 18:05:42.141934983 +0000 UTC Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.273977 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05aa39d-94e6-49d4-885e-2c1f986310f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T13:04:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 13:04:35.885968 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 13:04:35.888176 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3462920193/tls.crt::/tmp/serving-cert-3462920193/tls.key\\\\\\\"\\\\nI0217 13:04:42.370546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 13:04:42.384242 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 13:04:42.384284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 13:04:42.384317 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 13:04:42.384327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 13:04:42.394555 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 13:04:42.394614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394628 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 13:04:42.394646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 13:04:42.394654 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 13:04:42.394662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 13:04:42.394900 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 13:04:42.396494 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:42Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.277735 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.277833 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.277859 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.277890 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.277913 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:42Z","lastTransitionTime":"2026-02-17T13:05:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.292365 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:42Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.308999 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:42Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.328057 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bf36887-2a2b-4193-abdb-a15beaf46300\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11a8924eb6ce41d629feb2b42c7d4778a3ca9110a59c3ac25acbba4367272374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bec78f9e9f6416d6b600f6231f0d24652473b48b95bf0e0675cc85c9208d124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-29qxq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:42Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.362273 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcbc7a69-58e7-4b7a-bb84-96276819ce88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29376a23843b866c283dc983b136aead4139d8ce2a97ca45633e17ad6028ced3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29376a23843b866c283dc983b136aead4139d8ce2a97ca45633e17ad6028ced3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T13:05:16Z\\\",\\\"message\\\":\\\"licy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0217 13:05:16.060354 6583 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0217 13:05:16.060455 6583 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0217 13:05:16.060913 6583 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0217 13:05:16.061605 6583 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0217 13:05:16.061682 6583 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0217 13:05:16.061709 6583 factory.go:656] Stopping watch factory\\\\nI0217 13:05:16.061733 6583 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0217 13:05:16.061766 6583 handler.go:208] Removed *v1.Node event handler 2\\\\nI0217 13:05:16.088613 6583 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0217 13:05:16.088662 6583 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0217 13:05:16.088736 6583 ovnkube.go:599] Stopped ovnkube\\\\nI0217 13:05:16.088764 6583 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0217 13:05:16.088897 6583 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:05:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dcwgg_openshift-ovn-kubernetes(dcbc7a69-58e7-4b7a-bb84-96276819ce88)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dcwgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:42Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.380996 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.381057 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.381074 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.381102 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.381122 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:42Z","lastTransitionTime":"2026-02-17T13:05:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.387917 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b49c9af-148a-48ec-af37-f555f9186337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d6919c91918e8bacafd1442b06d14ffd9b547f0f37b2298e2d8d3acc5962d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e38466a067b20ca4323ab6983cb324bb63179f942d2c37c0daa8b03767e7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4830450cbada3d24c9747e9de5f48e53e118549f1dd67edf7fff8b2adc47128f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001dfb595f39d1da33b265474b6cca51c62f522ff061c9e7c52f93806034f4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f1fee7ac5f8d1442d9f63b0e8d285187129eb3a3e3b50a49a45c7a7a39ae8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:42Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.410340 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5c34a3d-82ad-4d03-abdd-0a20818ac2fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0261f41b82b492d324b5e50edd72cedd5e824de542a7b47104abd2ab757a7049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f43c7fe16acfd4d39301ee64665d4cd1d885397f981d70f73215da7f3aabe72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c871c1b454e4edd2f41076ecfd6a66592023d7161f094fbb4fd197fbf9c881f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:42Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.442893 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:42Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.482155 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"924c66b7-7c3c-4db7-a427-3f152efbac4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9beee007a179c3c2e04c2ce0de8ea6dbae9da3421ab24f23b718a23744c828a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fp5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:42Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.483481 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.483512 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.483524 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.483542 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.483558 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:42Z","lastTransitionTime":"2026-02-17T13:05:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.496830 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2k5pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77480847-3271-4161-a833-e62b3bb4413e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fffmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fffmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2k5pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:42Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.510559 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aea1a9-3570-4339-a1a4-db2dc358328f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3285726bf736fdd9d99c9974bef45c180410c4e3b795acbbf6c4ac576be3ee9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c8e77568b5ec88d39510e2cdf7b9a3438a9c7cd0ad83cb350c9d7c927f0dbc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bbbc1ab7b069d1abd31929ec33bd494a7f403437f0be70da380ea0a5ff2e711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25f110bf0e9d951f4d48a491daae897a8250cf77fe464f1914cb2f1cfcde1d72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25f110bf0e9d951f4d48a491daae897a8250cf77fe464f1914cb2f1cfcde1d72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:42Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.527922 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d0966253f623158053125dffca956000824438bdc548eba6dafe9a162d5674a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d07bfb361f14d817913fa2cabd30649447790068545dda2f3a2f71213ea0e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:42Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.542136 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2gsx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1face148-45d1-44b3-8775-209f9f8651fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82c5147255d021114a8639e9b60593f847896b613b5d862db588014f7485faa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnlcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2gsx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:42Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.559390 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lffhw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8155b25d-be0c-44ed-b3af-7bf25d55ec83\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad540421c8104f89cf02f413145e08de7c3e65e47f78cb2be5b40f746d6376c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gw5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52f6c4f214dd72b2df8ba07e3448bb4a2bedc698f47c16d027a774e2723c201\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gw5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lffhw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:42Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.575049 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9ecd4380a6ae8ff62e27e3946815170c2f02b4c3921bda595467345289c9e0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:42Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.586807 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.586863 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.586875 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.586897 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.586912 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:42Z","lastTransitionTime":"2026-02-17T13:05:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.588535 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0f4056e4b5a554f05004c194485d25231b67024e4376a73c715fcb32004a3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:42Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.689110 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.689199 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.689216 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.689242 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.689261 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:42Z","lastTransitionTime":"2026-02-17T13:05:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.792563 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.792612 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.792625 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.792647 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.792661 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:42Z","lastTransitionTime":"2026-02-17T13:05:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.896060 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.896123 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.896141 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.896168 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.896189 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:42Z","lastTransitionTime":"2026-02-17T13:05:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.998916 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.998975 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.998987 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.999005 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:42 crc kubenswrapper[4955]: I0217 13:05:42.999015 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:42Z","lastTransitionTime":"2026-02-17T13:05:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.101425 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.101486 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.101508 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.101536 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.101554 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:43Z","lastTransitionTime":"2026-02-17T13:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.208865 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.209013 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.209032 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.209067 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.209103 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:43Z","lastTransitionTime":"2026-02-17T13:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.222281 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:05:43 crc kubenswrapper[4955]: E0217 13:05:43.222393 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.222287 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:05:43 crc kubenswrapper[4955]: E0217 13:05:43.222549 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.222318 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:05:43 crc kubenswrapper[4955]: E0217 13:05:43.222907 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.257020 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 23:39:56.369475677 +0000 UTC Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.312278 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.312327 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.312344 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.312369 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.312387 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:43Z","lastTransitionTime":"2026-02-17T13:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.415433 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.415483 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.415494 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.415510 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.415524 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:43Z","lastTransitionTime":"2026-02-17T13:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.520045 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.520099 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.520112 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.520129 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.520141 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:43Z","lastTransitionTime":"2026-02-17T13:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.622539 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.622597 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.622609 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.622632 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.622645 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:43Z","lastTransitionTime":"2026-02-17T13:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.726375 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.726444 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.726455 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.726479 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.726493 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:43Z","lastTransitionTime":"2026-02-17T13:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.829479 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.829534 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.829545 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.829567 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.829582 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:43Z","lastTransitionTime":"2026-02-17T13:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.932701 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.932851 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.932871 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.932897 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:43 crc kubenswrapper[4955]: I0217 13:05:43.932917 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:43Z","lastTransitionTime":"2026-02-17T13:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.036368 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.036449 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.036473 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.036507 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.036585 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:44Z","lastTransitionTime":"2026-02-17T13:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.140115 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.140210 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.140249 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.140283 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.140306 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:44Z","lastTransitionTime":"2026-02-17T13:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.223147 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:05:44 crc kubenswrapper[4955]: E0217 13:05:44.223664 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.224127 4955 scope.go:117] "RemoveContainer" containerID="29376a23843b866c283dc983b136aead4139d8ce2a97ca45633e17ad6028ced3" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.242830 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.242900 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.242929 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.242973 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.242998 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:44Z","lastTransitionTime":"2026-02-17T13:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.258344 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 07:55:17.151095119 +0000 UTC Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.346876 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.347228 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.347244 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.347266 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.347280 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:44Z","lastTransitionTime":"2026-02-17T13:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.450905 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.450956 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.450967 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.450986 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.450998 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:44Z","lastTransitionTime":"2026-02-17T13:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.554099 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.554141 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.554189 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.554212 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.554223 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:44Z","lastTransitionTime":"2026-02-17T13:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.657356 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.657419 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.657439 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.657465 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.657484 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:44Z","lastTransitionTime":"2026-02-17T13:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.760416 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.760458 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.760470 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.760488 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.760501 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:44Z","lastTransitionTime":"2026-02-17T13:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.812303 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dcwgg_dcbc7a69-58e7-4b7a-bb84-96276819ce88/ovnkube-controller/2.log" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.815681 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" event={"ID":"dcbc7a69-58e7-4b7a-bb84-96276819ce88","Type":"ContainerStarted","Data":"528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6"} Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.816211 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.838610 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9ecd4380a6ae8ff62e27e3946815170c2f02b4c3921bda595467345289c9e0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:44Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.853929 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0f4056e4b5a554f05004c194485d25231b67024e4376a73c715fcb32004a3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:44Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.863975 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.864082 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.864097 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.864202 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.864217 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:44Z","lastTransitionTime":"2026-02-17T13:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.866609 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-msn6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://277a12d62c90ef420c31d8dc71f0d7750dfd5f5ab7e90fec0dcd00e54b60245d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2tvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-msn6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:44Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.889055 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05aa39d-94e6-49d4-885e-2c1f986310f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T13:04:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 13:04:35.885968 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 13:04:35.888176 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3462920193/tls.crt::/tmp/serving-cert-3462920193/tls.key\\\\\\\"\\\\nI0217 13:04:42.370546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 13:04:42.384242 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 13:04:42.384284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 13:04:42.384317 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 13:04:42.384327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 13:04:42.394555 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 13:04:42.394614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394628 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 13:04:42.394646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 13:04:42.394654 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 13:04:42.394662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 13:04:42.394900 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 13:04:42.396494 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:44Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.909190 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:44Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.923225 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:44Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.938821 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bf36887-2a2b-4193-abdb-a15beaf46300\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11a8924eb6ce41d629feb2b42c7d4778a3ca9110a59c3ac25acbba4367272374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bec78f9e9f6416d6b600f6231f0d24652473b48b95bf0e0675cc85c9208d124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-29qxq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:44Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.960312 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9hx9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e57bdf8-e537-402a-9815-609fecdd68ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c81948fc2806ea8718909fd269c029efc6a26675e45359e971272354eccfa74d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbc9014d80aee39922161014aac8fafa7d93be3e4092fe705c76da4a487a5e5d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T13:05:31Z\\\",\\\"message\\\":\\\"2026-02-17T13:04:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c2764cb0-2a4e-4449-bac5-794fed2f307a\\\\n2026-02-17T13:04:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c2764cb0-2a4e-4449-bac5-794fed2f307a to /host/opt/cni/bin/\\\\n2026-02-17T13:04:46Z [verbose] multus-daemon started\\\\n2026-02-17T13:04:46Z [verbose] Readiness Indicator file check\\\\n2026-02-17T13:05:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:05:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q22d4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9hx9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:44Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.974303 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.974360 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.974403 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.974454 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.974471 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:44Z","lastTransitionTime":"2026-02-17T13:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:44 crc kubenswrapper[4955]: I0217 13:05:44.992001 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b49c9af-148a-48ec-af37-f555f9186337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d6919c91918e8bacafd1442b06d14ffd9b547f0f37b2298e2d8d3acc5962d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e38466a067b20ca4323ab6983cb324bb63179f942d2c37c0daa8b03767e7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4830450cbada3d24c9747e9de5f48e53e118549f1dd67edf7fff8b2adc47128f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001dfb595f39d1da33b265474b6cca51c62f522ff061c9e7c52f93806034f4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f1fee7ac5f8d1442d9f63b0e8d285187129eb3a3e3b50a49a45c7a7a39ae8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:44Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.007852 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5c34a3d-82ad-4d03-abdd-0a20818ac2fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0261f41b82b492d324b5e50edd72cedd5e824de542a7b47104abd2ab757a7049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f43c7fe16acfd4d39301ee64665d4cd1d885397f981d70f73215da7f3aabe72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c871c1b454e4edd2f41076ecfd6a66592023d7161f094fbb4fd197fbf9c881f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.026996 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.047932 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"924c66b7-7c3c-4db7-a427-3f152efbac4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9beee007a179c3c2e04c2ce0de8ea6dbae9da3421ab24f23b718a23744c828a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fp5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.075258 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcbc7a69-58e7-4b7a-bb84-96276819ce88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29376a23843b866c283dc983b136aead4139d8ce2a97ca45633e17ad6028ced3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T13:05:16Z\\\",\\\"message\\\":\\\"licy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0217 13:05:16.060354 6583 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0217 13:05:16.060455 6583 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0217 13:05:16.060913 6583 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0217 13:05:16.061605 6583 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0217 13:05:16.061682 6583 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0217 13:05:16.061709 6583 factory.go:656] Stopping watch factory\\\\nI0217 13:05:16.061733 6583 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0217 13:05:16.061766 6583 handler.go:208] Removed *v1.Node event handler 2\\\\nI0217 13:05:16.088613 6583 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0217 13:05:16.088662 6583 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0217 13:05:16.088736 6583 ovnkube.go:599] Stopped ovnkube\\\\nI0217 13:05:16.088764 6583 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0217 13:05:16.088897 6583 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:05:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:05:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dcwgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.076976 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.077006 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.077016 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.077035 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.077047 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:45Z","lastTransitionTime":"2026-02-17T13:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.094387 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aea1a9-3570-4339-a1a4-db2dc358328f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3285726bf736fdd9d99c9974bef45c180410c4e3b795acbbf6c4ac576be3ee9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c8e77568b5ec88d39510e2cdf7b9a3438a9c7cd0ad83cb350c9d7c927f0dbc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bbbc1ab7b069d1abd31929ec33bd494a7f403437f0be70da380ea0a5ff2e711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25f110bf0e9d951f4d48a491daae897a8250cf77fe464f1914cb2f1cfcde1d72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25f110bf0e9d951f4d48a491daae897a8250cf77fe464f1914cb2f1cfcde1d72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.107975 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d0966253f623158053125dffca956000824438bdc548eba6dafe9a162d5674a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d07bfb361f14d817913fa2cabd30649447790068545dda2f3a2f71213ea0e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.120494 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2gsx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1face148-45d1-44b3-8775-209f9f8651fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82c5147255d021114a8639e9b60593f847896b613b5d862db588014f7485faa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnlcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2gsx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.135404 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lffhw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8155b25d-be0c-44ed-b3af-7bf25d55ec83\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad540421c8104f89cf02f413145e08de7c3e65e47f78cb2be5b40f746d6376c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gw5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52f6c4f214dd72b2df8ba07e3448bb4a2bedc698f47c16d027a774e2723c201\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gw5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lffhw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.148367 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2k5pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77480847-3271-4161-a833-e62b3bb4413e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fffmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fffmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2k5pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.179728 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.179804 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.179821 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.179845 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.179860 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:45Z","lastTransitionTime":"2026-02-17T13:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.221925 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.221997 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.222065 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:05:45 crc kubenswrapper[4955]: E0217 13:05:45.222215 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:05:45 crc kubenswrapper[4955]: E0217 13:05:45.222357 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:05:45 crc kubenswrapper[4955]: E0217 13:05:45.222465 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.259240 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 05:40:14.393366303 +0000 UTC Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.283143 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.283197 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.283209 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.283229 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.283242 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:45Z","lastTransitionTime":"2026-02-17T13:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.386644 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.386688 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.386700 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.386720 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.386734 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:45Z","lastTransitionTime":"2026-02-17T13:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.489843 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.489887 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.489899 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.489915 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.489926 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:45Z","lastTransitionTime":"2026-02-17T13:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.593677 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.593761 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.593776 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.593845 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.593862 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:45Z","lastTransitionTime":"2026-02-17T13:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.695893 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.695963 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.695995 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.696029 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.696051 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:45Z","lastTransitionTime":"2026-02-17T13:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.798436 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.798485 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.798497 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.798515 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.798529 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:45Z","lastTransitionTime":"2026-02-17T13:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.821682 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dcwgg_dcbc7a69-58e7-4b7a-bb84-96276819ce88/ovnkube-controller/3.log" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.822577 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dcwgg_dcbc7a69-58e7-4b7a-bb84-96276819ce88/ovnkube-controller/2.log" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.826488 4955 generic.go:334] "Generic (PLEG): container finished" podID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerID="528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6" exitCode=1 Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.826557 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" event={"ID":"dcbc7a69-58e7-4b7a-bb84-96276819ce88","Type":"ContainerDied","Data":"528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6"} Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.826628 4955 scope.go:117] "RemoveContainer" containerID="29376a23843b866c283dc983b136aead4139d8ce2a97ca45633e17ad6028ced3" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.827257 4955 scope.go:117] "RemoveContainer" containerID="528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6" Feb 17 13:05:45 crc kubenswrapper[4955]: E0217 13:05:45.827465 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dcwgg_openshift-ovn-kubernetes(dcbc7a69-58e7-4b7a-bb84-96276819ce88)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.848142 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lffhw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8155b25d-be0c-44ed-b3af-7bf25d55ec83\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad540421c8104f89cf02f413145e08de7c3e65e47f78cb2be5b40f746d6376c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gw5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52f6c4f214dd72b2df8ba07e3448bb4a2bedc698f47c16d027a774e2723c201\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gw5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lffhw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.863485 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2k5pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77480847-3271-4161-a833-e62b3bb4413e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fffmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fffmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2k5pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.879677 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aea1a9-3570-4339-a1a4-db2dc358328f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3285726bf736fdd9d99c9974bef45c180410c4e3b795acbbf6c4ac576be3ee9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c8e77568b5ec88d39510e2cdf7b9a3438a9c7cd0ad83cb350c9d7c927f0dbc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bbbc1ab7b069d1abd31929ec33bd494a7f403437f0be70da380ea0a5ff2e711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25f110bf0e9d951f4d48a491daae897a8250cf77fe464f1914cb2f1cfcde1d72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25f110bf0e9d951f4d48a491daae897a8250cf77fe464f1914cb2f1cfcde1d72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.899135 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d0966253f623158053125dffca956000824438bdc548eba6dafe9a162d5674a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d07bfb361f14d817913fa2cabd30649447790068545dda2f3a2f71213ea0e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.902205 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.902258 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.902282 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.902314 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.902337 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:45Z","lastTransitionTime":"2026-02-17T13:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.910258 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2gsx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1face148-45d1-44b3-8775-209f9f8651fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82c5147255d021114a8639e9b60593f847896b613b5d862db588014f7485faa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnlcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2gsx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.930063 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9ecd4380a6ae8ff62e27e3946815170c2f02b4c3921bda595467345289c9e0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.948022 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0f4056e4b5a554f05004c194485d25231b67024e4376a73c715fcb32004a3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.962021 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bf36887-2a2b-4193-abdb-a15beaf46300\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11a8924eb6ce41d629feb2b42c7d4778a3ca9110a59c3ac25acbba4367272374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bec78f9e9f6416d6b600f6231f0d24652473b48b95bf0e0675cc85c9208d124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-29qxq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.977382 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9hx9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e57bdf8-e537-402a-9815-609fecdd68ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c81948fc2806ea8718909fd269c029efc6a26675e45359e971272354eccfa74d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbc9014d80aee39922161014aac8fafa7d93be3e4092fe705c76da4a487a5e5d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T13:05:31Z\\\",\\\"message\\\":\\\"2026-02-17T13:04:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c2764cb0-2a4e-4449-bac5-794fed2f307a\\\\n2026-02-17T13:04:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c2764cb0-2a4e-4449-bac5-794fed2f307a to /host/opt/cni/bin/\\\\n2026-02-17T13:04:46Z [verbose] multus-daemon started\\\\n2026-02-17T13:04:46Z [verbose] Readiness Indicator file check\\\\n2026-02-17T13:05:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:05:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q22d4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9hx9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:45 crc kubenswrapper[4955]: I0217 13:05:45.990392 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-msn6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://277a12d62c90ef420c31d8dc71f0d7750dfd5f5ab7e90fec0dcd00e54b60245d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2tvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-msn6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:45Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.004902 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.004989 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.005009 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.005034 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.005052 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:46Z","lastTransitionTime":"2026-02-17T13:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.009202 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05aa39d-94e6-49d4-885e-2c1f986310f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T13:04:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 13:04:35.885968 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 13:04:35.888176 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3462920193/tls.crt::/tmp/serving-cert-3462920193/tls.key\\\\\\\"\\\\nI0217 13:04:42.370546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 13:04:42.384242 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 13:04:42.384284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 13:04:42.384317 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 13:04:42.384327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 13:04:42.394555 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 13:04:42.394614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394628 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 13:04:42.394646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 13:04:42.394654 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 13:04:42.394662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 13:04:42.394900 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 13:04:42.396494 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:46Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.030372 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:46Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.046383 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:46Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.067764 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"924c66b7-7c3c-4db7-a427-3f152efbac4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9beee007a179c3c2e04c2ce0de8ea6dbae9da3421ab24f23b718a23744c828a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fp5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:46Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.090835 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcbc7a69-58e7-4b7a-bb84-96276819ce88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29376a23843b866c283dc983b136aead4139d8ce2a97ca45633e17ad6028ced3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T13:05:16Z\\\",\\\"message\\\":\\\"licy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0217 13:05:16.060354 6583 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0217 13:05:16.060455 6583 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0217 13:05:16.060913 6583 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0217 13:05:16.061605 6583 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0217 13:05:16.061682 6583 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0217 13:05:16.061709 6583 factory.go:656] Stopping watch factory\\\\nI0217 13:05:16.061733 6583 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0217 13:05:16.061766 6583 handler.go:208] Removed *v1.Node event handler 2\\\\nI0217 13:05:16.088613 6583 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0217 13:05:16.088662 6583 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0217 13:05:16.088736 6583 ovnkube.go:599] Stopped ovnkube\\\\nI0217 13:05:16.088764 6583 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0217 13:05:16.088897 6583 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:05:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T13:05:45Z\\\",\\\"message\\\":\\\"m github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0217 13:05:45.479055 6997 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0217 13:05:45.478979 6997 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0217 13:05:45.479313 6997 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0217 13:05:45.479345 6997 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0217 13:05:45.479386 6997 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0217 13:05:45.479405 6997 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0217 13:05:45.479430 6997 factory.go:656] Stopping watch factory\\\\nI0217 13:05:45.479446 6997 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0217 13:05:45.479594 6997 handler.go:208] Removed *v1.Node event handler 2\\\\nI0217 13:05:45.479607 6997 handler.go:208] Removed *v1.Node event handler 7\\\\nI0217 13:05:45.479629 6997 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0217 13:05:45.479687 6997 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:05:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dcwgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:46Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.107891 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.108115 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.108203 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.108292 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.108364 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:46Z","lastTransitionTime":"2026-02-17T13:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.119024 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b49c9af-148a-48ec-af37-f555f9186337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d6919c91918e8bacafd1442b06d14ffd9b547f0f37b2298e2d8d3acc5962d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e38466a067b20ca4323ab6983cb324bb63179f942d2c37c0daa8b03767e7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4830450cbada3d24c9747e9de5f48e53e118549f1dd67edf7fff8b2adc47128f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001dfb595f39d1da33b265474b6cca51c62f522ff061c9e7c52f93806034f4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f1fee7ac5f8d1442d9f63b0e8d285187129eb3a3e3b50a49a45c7a7a39ae8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:46Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.135271 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5c34a3d-82ad-4d03-abdd-0a20818ac2fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0261f41b82b492d324b5e50edd72cedd5e824de542a7b47104abd2ab757a7049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f43c7fe16acfd4d39301ee64665d4cd1d885397f981d70f73215da7f3aabe72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c871c1b454e4edd2f41076ecfd6a66592023d7161f094fbb4fd197fbf9c881f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:46Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.151491 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:46Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.214850 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.214906 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.214923 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.214948 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.214965 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:46Z","lastTransitionTime":"2026-02-17T13:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.222681 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:05:46 crc kubenswrapper[4955]: E0217 13:05:46.222864 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.260037 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 20:58:44.511887299 +0000 UTC Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.317594 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.317975 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.318251 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.318565 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.318775 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:46Z","lastTransitionTime":"2026-02-17T13:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.422000 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.422826 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.423037 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.423445 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.423691 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:46Z","lastTransitionTime":"2026-02-17T13:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.526970 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.527755 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.528000 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.528170 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.528319 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:46Z","lastTransitionTime":"2026-02-17T13:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.631656 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.631710 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.631731 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.631758 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.631780 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:46Z","lastTransitionTime":"2026-02-17T13:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.734595 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.734676 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.734699 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.734732 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.734754 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:46Z","lastTransitionTime":"2026-02-17T13:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.832483 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dcwgg_dcbc7a69-58e7-4b7a-bb84-96276819ce88/ovnkube-controller/3.log" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.837079 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.837136 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.837161 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.837185 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.837204 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:46Z","lastTransitionTime":"2026-02-17T13:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.838080 4955 scope.go:117] "RemoveContainer" containerID="528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6" Feb 17 13:05:46 crc kubenswrapper[4955]: E0217 13:05:46.838347 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dcwgg_openshift-ovn-kubernetes(dcbc7a69-58e7-4b7a-bb84-96276819ce88)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.853209 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2k5pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77480847-3271-4161-a833-e62b3bb4413e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fffmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fffmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2k5pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:46Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.872379 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aea1a9-3570-4339-a1a4-db2dc358328f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3285726bf736fdd9d99c9974bef45c180410c4e3b795acbbf6c4ac576be3ee9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c8e77568b5ec88d39510e2cdf7b9a3438a9c7cd0ad83cb350c9d7c927f0dbc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bbbc1ab7b069d1abd31929ec33bd494a7f403437f0be70da380ea0a5ff2e711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25f110bf0e9d951f4d48a491daae897a8250cf77fe464f1914cb2f1cfcde1d72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25f110bf0e9d951f4d48a491daae897a8250cf77fe464f1914cb2f1cfcde1d72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:46Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.893373 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d0966253f623158053125dffca956000824438bdc548eba6dafe9a162d5674a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d07bfb361f14d817913fa2cabd30649447790068545dda2f3a2f71213ea0e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:46Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.909161 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2gsx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1face148-45d1-44b3-8775-209f9f8651fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82c5147255d021114a8639e9b60593f847896b613b5d862db588014f7485faa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnlcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2gsx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:46Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.925724 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lffhw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8155b25d-be0c-44ed-b3af-7bf25d55ec83\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad540421c8104f89cf02f413145e08de7c3e65e47f78cb2be5b40f746d6376c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gw5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52f6c4f214dd72b2df8ba07e3448bb4a2bedc698f47c16d027a774e2723c201\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gw5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lffhw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:46Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.941148 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.941231 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.941256 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.941287 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.941309 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:46Z","lastTransitionTime":"2026-02-17T13:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.945449 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9ecd4380a6ae8ff62e27e3946815170c2f02b4c3921bda595467345289c9e0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:46Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.963395 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0f4056e4b5a554f05004c194485d25231b67024e4376a73c715fcb32004a3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:46Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.983097 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9hx9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e57bdf8-e537-402a-9815-609fecdd68ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c81948fc2806ea8718909fd269c029efc6a26675e45359e971272354eccfa74d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbc9014d80aee39922161014aac8fafa7d93be3e4092fe705c76da4a487a5e5d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T13:05:31Z\\\",\\\"message\\\":\\\"2026-02-17T13:04:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c2764cb0-2a4e-4449-bac5-794fed2f307a\\\\n2026-02-17T13:04:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c2764cb0-2a4e-4449-bac5-794fed2f307a to /host/opt/cni/bin/\\\\n2026-02-17T13:04:46Z [verbose] multus-daemon started\\\\n2026-02-17T13:04:46Z [verbose] Readiness Indicator file check\\\\n2026-02-17T13:05:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:05:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q22d4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9hx9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:46Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:46 crc kubenswrapper[4955]: I0217 13:05:46.998618 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-msn6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://277a12d62c90ef420c31d8dc71f0d7750dfd5f5ab7e90fec0dcd00e54b60245d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2tvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-msn6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:46Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.015530 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05aa39d-94e6-49d4-885e-2c1f986310f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T13:04:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 13:04:35.885968 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 13:04:35.888176 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3462920193/tls.crt::/tmp/serving-cert-3462920193/tls.key\\\\\\\"\\\\nI0217 13:04:42.370546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 13:04:42.384242 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 13:04:42.384284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 13:04:42.384317 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 13:04:42.384327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 13:04:42.394555 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 13:04:42.394614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394628 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 13:04:42.394646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 13:04:42.394654 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 13:04:42.394662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 13:04:42.394900 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 13:04:42.396494 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:47Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.034272 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:47Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.045098 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.045166 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.045189 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.045222 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.045247 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:47Z","lastTransitionTime":"2026-02-17T13:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.055860 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:47Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.071222 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bf36887-2a2b-4193-abdb-a15beaf46300\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11a8924eb6ce41d629feb2b42c7d4778a3ca9110a59c3ac25acbba4367272374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bec78f9e9f6416d6b600f6231f0d24652473b48b95bf0e0675cc85c9208d124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-29qxq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:47Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.097414 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcbc7a69-58e7-4b7a-bb84-96276819ce88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T13:05:45Z\\\",\\\"message\\\":\\\"m github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0217 13:05:45.479055 6997 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0217 13:05:45.478979 6997 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0217 13:05:45.479313 6997 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0217 13:05:45.479345 6997 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0217 13:05:45.479386 6997 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0217 13:05:45.479405 6997 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0217 13:05:45.479430 6997 factory.go:656] Stopping watch factory\\\\nI0217 13:05:45.479446 6997 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0217 13:05:45.479594 6997 handler.go:208] Removed *v1.Node event handler 2\\\\nI0217 13:05:45.479607 6997 handler.go:208] Removed *v1.Node event handler 7\\\\nI0217 13:05:45.479629 6997 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0217 13:05:45.479687 6997 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:05:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dcwgg_openshift-ovn-kubernetes(dcbc7a69-58e7-4b7a-bb84-96276819ce88)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dcwgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:47Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.098706 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.098942 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:05:47 crc kubenswrapper[4955]: E0217 13:05:47.099085 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:51.099033998 +0000 UTC m=+149.621763561 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:05:47 crc kubenswrapper[4955]: E0217 13:05:47.099156 4955 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 17 13:05:47 crc kubenswrapper[4955]: E0217 13:05:47.099191 4955 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 17 13:05:47 crc kubenswrapper[4955]: E0217 13:05:47.099218 4955 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 13:05:47 crc kubenswrapper[4955]: E0217 13:05:47.099333 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-17 13:06:51.099303586 +0000 UTC m=+149.622033169 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.099332 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.099444 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.099517 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:05:47 crc kubenswrapper[4955]: E0217 13:05:47.099461 4955 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 17 13:05:47 crc kubenswrapper[4955]: E0217 13:05:47.099560 4955 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 17 13:05:47 crc kubenswrapper[4955]: E0217 13:05:47.099582 4955 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 17 13:05:47 crc kubenswrapper[4955]: E0217 13:05:47.099607 4955 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 13:05:47 crc kubenswrapper[4955]: E0217 13:05:47.099622 4955 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 17 13:05:47 crc kubenswrapper[4955]: E0217 13:05:47.099637 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-17 13:06:51.099597594 +0000 UTC m=+149.622327137 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 17 13:05:47 crc kubenswrapper[4955]: E0217 13:05:47.099657 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-17 13:06:51.099646475 +0000 UTC m=+149.622376028 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 13:05:47 crc kubenswrapper[4955]: E0217 13:05:47.099679 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-17 13:06:51.099669926 +0000 UTC m=+149.622399479 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.128746 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b49c9af-148a-48ec-af37-f555f9186337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d6919c91918e8bacafd1442b06d14ffd9b547f0f37b2298e2d8d3acc5962d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e38466a067b20ca4323ab6983cb324bb63179f942d2c37c0daa8b03767e7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4830450cbada3d24c9747e9de5f48e53e118549f1dd67edf7fff8b2adc47128f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001dfb595f39d1da33b265474b6cca51c62f522ff061c9e7c52f93806034f4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f1fee7ac5f8d1442d9f63b0e8d285187129eb3a3e3b50a49a45c7a7a39ae8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:47Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.148012 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.148082 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.148105 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.148140 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.148169 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:47Z","lastTransitionTime":"2026-02-17T13:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.151008 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5c34a3d-82ad-4d03-abdd-0a20818ac2fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0261f41b82b492d324b5e50edd72cedd5e824de542a7b47104abd2ab757a7049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f43c7fe16acfd4d39301ee64665d4cd1d885397f981d70f73215da7f3aabe72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c871c1b454e4edd2f41076ecfd6a66592023d7161f094fbb4fd197fbf9c881f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:47Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.172970 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:47Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.195203 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"924c66b7-7c3c-4db7-a427-3f152efbac4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9beee007a179c3c2e04c2ce0de8ea6dbae9da3421ab24f23b718a23744c828a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fp5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:47Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.223071 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.223148 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.223307 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:05:47 crc kubenswrapper[4955]: E0217 13:05:47.223486 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:05:47 crc kubenswrapper[4955]: E0217 13:05:47.223663 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:05:47 crc kubenswrapper[4955]: E0217 13:05:47.223824 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.251863 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.251916 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.251937 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.251967 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.251990 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:47Z","lastTransitionTime":"2026-02-17T13:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.261117 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 21:48:55.931767085 +0000 UTC Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.355156 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.355202 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.355213 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.355232 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.355246 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:47Z","lastTransitionTime":"2026-02-17T13:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.457876 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.457944 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.457963 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.457990 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.458008 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:47Z","lastTransitionTime":"2026-02-17T13:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.562009 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.562069 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.562087 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.562118 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.562137 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:47Z","lastTransitionTime":"2026-02-17T13:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.665630 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.665688 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.665705 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.665729 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.665747 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:47Z","lastTransitionTime":"2026-02-17T13:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.768830 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.768900 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.768922 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.768950 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.768976 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:47Z","lastTransitionTime":"2026-02-17T13:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.871703 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.872092 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.872104 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.872124 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.872136 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:47Z","lastTransitionTime":"2026-02-17T13:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.975393 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.975455 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.975466 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.975486 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:47 crc kubenswrapper[4955]: I0217 13:05:47.975501 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:47Z","lastTransitionTime":"2026-02-17T13:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.079171 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.079220 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.079231 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.079251 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.079261 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:48Z","lastTransitionTime":"2026-02-17T13:05:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.183165 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.183233 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.183244 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.183262 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.183276 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:48Z","lastTransitionTime":"2026-02-17T13:05:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.223854 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:05:48 crc kubenswrapper[4955]: E0217 13:05:48.224171 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.261286 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 04:50:52.642208717 +0000 UTC Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.287308 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.287357 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.287369 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.287389 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.287401 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:48Z","lastTransitionTime":"2026-02-17T13:05:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.390699 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.390764 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.390782 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.390857 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.390878 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:48Z","lastTransitionTime":"2026-02-17T13:05:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.494327 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.494414 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.494433 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.494464 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.494483 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:48Z","lastTransitionTime":"2026-02-17T13:05:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.598521 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.598575 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.598588 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.598609 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.598626 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:48Z","lastTransitionTime":"2026-02-17T13:05:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.701374 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.701425 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.701437 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.701455 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.701469 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:48Z","lastTransitionTime":"2026-02-17T13:05:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.804492 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.804540 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.804551 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.804567 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.804578 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:48Z","lastTransitionTime":"2026-02-17T13:05:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.908989 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.909124 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.909145 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.909178 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:48 crc kubenswrapper[4955]: I0217 13:05:48.909204 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:48Z","lastTransitionTime":"2026-02-17T13:05:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.011463 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.011502 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.011513 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.011529 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.011561 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:49Z","lastTransitionTime":"2026-02-17T13:05:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.114222 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.114298 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.114311 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.114333 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.114348 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:49Z","lastTransitionTime":"2026-02-17T13:05:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.216488 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.216528 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.216539 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.216556 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.216568 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:49Z","lastTransitionTime":"2026-02-17T13:05:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.222920 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.222943 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:05:49 crc kubenswrapper[4955]: E0217 13:05:49.223131 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.222950 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:05:49 crc kubenswrapper[4955]: E0217 13:05:49.223228 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:05:49 crc kubenswrapper[4955]: E0217 13:05:49.223408 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.261902 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 00:19:11.939229642 +0000 UTC Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.319679 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.319730 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.319748 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.319775 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.319839 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:49Z","lastTransitionTime":"2026-02-17T13:05:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.424126 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.424180 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.424197 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.424225 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.424244 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:49Z","lastTransitionTime":"2026-02-17T13:05:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.527287 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.527348 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.527361 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.527395 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.527414 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:49Z","lastTransitionTime":"2026-02-17T13:05:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.630598 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.630681 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.630701 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.630730 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.630756 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:49Z","lastTransitionTime":"2026-02-17T13:05:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.733558 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.733587 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.733598 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.733615 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.733626 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:49Z","lastTransitionTime":"2026-02-17T13:05:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.836047 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.836135 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.836158 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.836193 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.836214 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:49Z","lastTransitionTime":"2026-02-17T13:05:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.941073 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.941129 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.941145 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.941166 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:49 crc kubenswrapper[4955]: I0217 13:05:49.941181 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:49Z","lastTransitionTime":"2026-02-17T13:05:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.044157 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.044211 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.044226 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.044244 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.044257 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:50Z","lastTransitionTime":"2026-02-17T13:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.147455 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.147522 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.147544 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.147572 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.147591 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:50Z","lastTransitionTime":"2026-02-17T13:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.222550 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:05:50 crc kubenswrapper[4955]: E0217 13:05:50.222815 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.250760 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.250831 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.250845 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.250862 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.250876 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:50Z","lastTransitionTime":"2026-02-17T13:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.262328 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 03:45:23.031862755 +0000 UTC Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.353626 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.353713 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.353732 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.353760 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.353830 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:50Z","lastTransitionTime":"2026-02-17T13:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.457191 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.457246 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.457259 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.457282 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.457305 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:50Z","lastTransitionTime":"2026-02-17T13:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.560893 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.560958 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.560975 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.561004 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.561024 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:50Z","lastTransitionTime":"2026-02-17T13:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.664720 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.664830 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.664858 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.664889 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.665096 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:50Z","lastTransitionTime":"2026-02-17T13:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.758859 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.758915 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.758925 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.758956 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.758971 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:50Z","lastTransitionTime":"2026-02-17T13:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:50 crc kubenswrapper[4955]: E0217 13:05:50.774874 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7605a498-360b-4ce4-b7a1-dfbc354d798d\\\",\\\"systemUUID\\\":\\\"6d4602d9-c718-4393-8f91-3c381f800c9e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:50Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.779533 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.779581 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.779610 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.779636 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.779648 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:50Z","lastTransitionTime":"2026-02-17T13:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:50 crc kubenswrapper[4955]: E0217 13:05:50.796634 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7605a498-360b-4ce4-b7a1-dfbc354d798d\\\",\\\"systemUUID\\\":\\\"6d4602d9-c718-4393-8f91-3c381f800c9e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:50Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.801878 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.801933 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.801949 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.801972 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.801986 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:50Z","lastTransitionTime":"2026-02-17T13:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:50 crc kubenswrapper[4955]: E0217 13:05:50.816599 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7605a498-360b-4ce4-b7a1-dfbc354d798d\\\",\\\"systemUUID\\\":\\\"6d4602d9-c718-4393-8f91-3c381f800c9e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:50Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.820666 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.820771 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.820809 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.820869 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.820884 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:50Z","lastTransitionTime":"2026-02-17T13:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:50 crc kubenswrapper[4955]: E0217 13:05:50.837913 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7605a498-360b-4ce4-b7a1-dfbc354d798d\\\",\\\"systemUUID\\\":\\\"6d4602d9-c718-4393-8f91-3c381f800c9e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:50Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.843212 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.843264 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.843282 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.843308 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.843326 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:50Z","lastTransitionTime":"2026-02-17T13:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:50 crc kubenswrapper[4955]: E0217 13:05:50.860343 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:05:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7605a498-360b-4ce4-b7a1-dfbc354d798d\\\",\\\"systemUUID\\\":\\\"6d4602d9-c718-4393-8f91-3c381f800c9e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:50Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:50 crc kubenswrapper[4955]: E0217 13:05:50.860514 4955 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.862316 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.862376 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.862386 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.862403 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.862485 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:50Z","lastTransitionTime":"2026-02-17T13:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.965483 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.965529 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.965539 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.965554 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:50 crc kubenswrapper[4955]: I0217 13:05:50.965565 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:50Z","lastTransitionTime":"2026-02-17T13:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.067957 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.068042 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.068065 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.068098 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.068123 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:51Z","lastTransitionTime":"2026-02-17T13:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.171540 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.171608 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.171627 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.171652 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.171674 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:51Z","lastTransitionTime":"2026-02-17T13:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.222915 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.223071 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:05:51 crc kubenswrapper[4955]: E0217 13:05:51.223217 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.223274 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:05:51 crc kubenswrapper[4955]: E0217 13:05:51.223477 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:05:51 crc kubenswrapper[4955]: E0217 13:05:51.223889 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.262870 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 02:38:14.244807038 +0000 UTC Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.274280 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.274368 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.274390 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.274420 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.274440 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:51Z","lastTransitionTime":"2026-02-17T13:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.377724 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.377777 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.377813 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.377837 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.377851 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:51Z","lastTransitionTime":"2026-02-17T13:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.480477 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.480524 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.480534 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.480549 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.480561 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:51Z","lastTransitionTime":"2026-02-17T13:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.582591 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.582625 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.582635 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.582653 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.582665 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:51Z","lastTransitionTime":"2026-02-17T13:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.686024 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.686126 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.686151 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.686189 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.686210 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:51Z","lastTransitionTime":"2026-02-17T13:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.790163 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.790232 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.790251 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.790277 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.790297 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:51Z","lastTransitionTime":"2026-02-17T13:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.894098 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.894145 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.894154 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.894171 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.894183 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:51Z","lastTransitionTime":"2026-02-17T13:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.997284 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.997347 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.997365 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.997393 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:51 crc kubenswrapper[4955]: I0217 13:05:51.997414 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:51Z","lastTransitionTime":"2026-02-17T13:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.100231 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.100337 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.100357 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.100384 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.100403 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:52Z","lastTransitionTime":"2026-02-17T13:05:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.203713 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.203753 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.203763 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.203803 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.203816 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:52Z","lastTransitionTime":"2026-02-17T13:05:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.222391 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:05:52 crc kubenswrapper[4955]: E0217 13:05:52.222696 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.236985 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bf36887-2a2b-4193-abdb-a15beaf46300\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11a8924eb6ce41d629feb2b42c7d4778a3ca9110a59c3ac25acbba4367272374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bec78f9e9f6416d6b600f6231f0d24652473b48b95bf0e0675cc85c9208d124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-29qxq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.237152 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.250743 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9hx9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e57bdf8-e537-402a-9815-609fecdd68ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c81948fc2806ea8718909fd269c029efc6a26675e45359e971272354eccfa74d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbc9014d80aee39922161014aac8fafa7d93be3e4092fe705c76da4a487a5e5d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T13:05:31Z\\\",\\\"message\\\":\\\"2026-02-17T13:04:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c2764cb0-2a4e-4449-bac5-794fed2f307a\\\\n2026-02-17T13:04:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c2764cb0-2a4e-4449-bac5-794fed2f307a to /host/opt/cni/bin/\\\\n2026-02-17T13:04:46Z [verbose] multus-daemon started\\\\n2026-02-17T13:04:46Z [verbose] Readiness Indicator file check\\\\n2026-02-17T13:05:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:05:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q22d4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9hx9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.262730 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-msn6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://277a12d62c90ef420c31d8dc71f0d7750dfd5f5ab7e90fec0dcd00e54b60245d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2tvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-msn6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.263735 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 06:51:19.822656312 +0000 UTC Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.287279 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05aa39d-94e6-49d4-885e-2c1f986310f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T13:04:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 13:04:35.885968 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 13:04:35.888176 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3462920193/tls.crt::/tmp/serving-cert-3462920193/tls.key\\\\\\\"\\\\nI0217 13:04:42.370546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 13:04:42.384242 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 13:04:42.384284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 13:04:42.384317 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 13:04:42.384327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 13:04:42.394555 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 13:04:42.394614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394628 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 13:04:42.394646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 13:04:42.394654 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 13:04:42.394662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 13:04:42.394900 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 13:04:42.396494 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.307514 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.307571 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.307589 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.307614 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.307632 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:52Z","lastTransitionTime":"2026-02-17T13:05:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.308298 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.327327 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.348247 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"924c66b7-7c3c-4db7-a427-3f152efbac4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9beee007a179c3c2e04c2ce0de8ea6dbae9da3421ab24f23b718a23744c828a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5541ac9dde0b9631d89f3e08c50961ab99e52f769c9125ba1f6c8615ca15eb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32eb79d8df797e7ac90160e2875382873d6d2507b0b179f4ce534cda1fa1c08e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1ec7a3bd9cae4f26879f83ed5f4f83a9f189740917dfdcc1e993b1ea641f693\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35bc6e4315baa590f4bd4e19716d18dc2c2640c38714dd497b3072462eec0461\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36944f893f06e7263143515ead9e9b867731a243226ba29335081a34d87160af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a2083d4fd81b933d9733011ab623ceb16807ec7ac0f526c1b29cecfb29e21d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6g4fg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-fp5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.373095 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcbc7a69-58e7-4b7a-bb84-96276819ce88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T13:05:45Z\\\",\\\"message\\\":\\\"m github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0217 13:05:45.479055 6997 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0217 13:05:45.478979 6997 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0217 13:05:45.479313 6997 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0217 13:05:45.479345 6997 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0217 13:05:45.479386 6997 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0217 13:05:45.479405 6997 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0217 13:05:45.479430 6997 factory.go:656] Stopping watch factory\\\\nI0217 13:05:45.479446 6997 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0217 13:05:45.479594 6997 handler.go:208] Removed *v1.Node event handler 2\\\\nI0217 13:05:45.479607 6997 handler.go:208] Removed *v1.Node event handler 7\\\\nI0217 13:05:45.479629 6997 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0217 13:05:45.479687 6997 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:05:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dcwgg_openshift-ovn-kubernetes(dcbc7a69-58e7-4b7a-bb84-96276819ce88)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2m6wz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dcwgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.401265 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b49c9af-148a-48ec-af37-f555f9186337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d6919c91918e8bacafd1442b06d14ffd9b547f0f37b2298e2d8d3acc5962d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e38466a067b20ca4323ab6983cb324bb63179f942d2c37c0daa8b03767e7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4830450cbada3d24c9747e9de5f48e53e118549f1dd67edf7fff8b2adc47128f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001dfb595f39d1da33b265474b6cca51c62f522ff061c9e7c52f93806034f4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f1fee7ac5f8d1442d9f63b0e8d285187129eb3a3e3b50a49a45c7a7a39ae8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7cf78ddd5f64739a5163e3b65c7b022d4ec271a40acb6abea702642dd47efa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f4deb40be0b0923c2eb73f0c30eebc58a638f86a7249877319bda6769ecde0a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://deeec9deddcb35af2989ef3239d010ad82fb95532d5cf832a72c19be8b945350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.410507 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.410545 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.410556 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.410573 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.410585 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:52Z","lastTransitionTime":"2026-02-17T13:05:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.421640 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5c34a3d-82ad-4d03-abdd-0a20818ac2fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0261f41b82b492d324b5e50edd72cedd5e824de542a7b47104abd2ab757a7049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f43c7fe16acfd4d39301ee64665d4cd1d885397f981d70f73215da7f3aabe72\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c871c1b454e4edd2f41076ecfd6a66592023d7161f094fbb4fd197fbf9c881f3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.440457 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.456291 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lffhw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8155b25d-be0c-44ed-b3af-7bf25d55ec83\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad540421c8104f89cf02f413145e08de7c3e65e47f78cb2be5b40f746d6376c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gw5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52f6c4f214dd72b2df8ba07e3448bb4a2bedc698f47c16d027a774e2723c201\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gw5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lffhw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.474272 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2k5pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77480847-3271-4161-a833-e62b3bb4413e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fffmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fffmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2k5pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.488325 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72aea1a9-3570-4339-a1a4-db2dc358328f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3285726bf736fdd9d99c9974bef45c180410c4e3b795acbbf6c4ac576be3ee9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c8e77568b5ec88d39510e2cdf7b9a3438a9c7cd0ad83cb350c9d7c927f0dbc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bbbc1ab7b069d1abd31929ec33bd494a7f403437f0be70da380ea0a5ff2e711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25f110bf0e9d951f4d48a491daae897a8250cf77fe464f1914cb2f1cfcde1d72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25f110bf0e9d951f4d48a491daae897a8250cf77fe464f1914cb2f1cfcde1d72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.507750 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d0966253f623158053125dffca956000824438bdc548eba6dafe9a162d5674a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d07bfb361f14d817913fa2cabd30649447790068545dda2f3a2f71213ea0e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.515298 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.515364 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.515378 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.515401 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.515416 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:52Z","lastTransitionTime":"2026-02-17T13:05:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.527646 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2gsx6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1face148-45d1-44b3-8775-209f9f8651fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82c5147255d021114a8639e9b60593f847896b613b5d862db588014f7485faa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnlcr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2gsx6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.548274 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9ecd4380a6ae8ff62e27e3946815170c2f02b4c3921bda595467345289c9e0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.566087 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0f4056e4b5a554f05004c194485d25231b67024e4376a73c715fcb32004a3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:05:52Z is after 2025-08-24T17:21:41Z" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.618614 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.618680 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.618707 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.618742 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.618765 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:52Z","lastTransitionTime":"2026-02-17T13:05:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.721152 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.721195 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.721205 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.721225 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.721237 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:52Z","lastTransitionTime":"2026-02-17T13:05:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.823522 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.823623 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.823648 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.823678 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.823706 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:52Z","lastTransitionTime":"2026-02-17T13:05:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.927262 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.927333 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.927356 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.927388 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:52 crc kubenswrapper[4955]: I0217 13:05:52.927414 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:52Z","lastTransitionTime":"2026-02-17T13:05:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.030755 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.030854 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.030874 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.030900 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.030917 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:53Z","lastTransitionTime":"2026-02-17T13:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.133712 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.133769 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.133813 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.133839 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.133857 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:53Z","lastTransitionTime":"2026-02-17T13:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.222394 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.222426 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:05:53 crc kubenswrapper[4955]: E0217 13:05:53.222535 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.222394 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:05:53 crc kubenswrapper[4955]: E0217 13:05:53.222662 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:05:53 crc kubenswrapper[4955]: E0217 13:05:53.222755 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.236429 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.236474 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.236491 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.236513 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.236531 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:53Z","lastTransitionTime":"2026-02-17T13:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.264966 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 15:58:40.937544072 +0000 UTC Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.339477 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.339530 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.339540 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.339558 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.339569 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:53Z","lastTransitionTime":"2026-02-17T13:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.443638 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.443712 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.443723 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.443745 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.443758 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:53Z","lastTransitionTime":"2026-02-17T13:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.547928 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.548017 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.548036 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.548065 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.548084 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:53Z","lastTransitionTime":"2026-02-17T13:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.651081 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.651135 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.651146 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.651170 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.651183 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:53Z","lastTransitionTime":"2026-02-17T13:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.753974 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.754145 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.754160 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.754180 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.754192 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:53Z","lastTransitionTime":"2026-02-17T13:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.858096 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.858172 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.858194 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.858239 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.858261 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:53Z","lastTransitionTime":"2026-02-17T13:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.961022 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.961092 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.961116 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.961144 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:53 crc kubenswrapper[4955]: I0217 13:05:53.961164 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:53Z","lastTransitionTime":"2026-02-17T13:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.064117 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.064421 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.064531 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.064625 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.064713 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:54Z","lastTransitionTime":"2026-02-17T13:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.168449 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.168490 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.168503 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.168521 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.168535 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:54Z","lastTransitionTime":"2026-02-17T13:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.223008 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:05:54 crc kubenswrapper[4955]: E0217 13:05:54.223556 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.266177 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 18:30:48.1600868 +0000 UTC Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.271252 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.271300 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.271317 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.271341 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.271359 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:54Z","lastTransitionTime":"2026-02-17T13:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.374041 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.374100 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.374122 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.374148 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.374167 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:54Z","lastTransitionTime":"2026-02-17T13:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.477438 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.477891 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.478087 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.478549 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.478748 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:54Z","lastTransitionTime":"2026-02-17T13:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.582274 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.582323 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.582333 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.582350 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.582362 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:54Z","lastTransitionTime":"2026-02-17T13:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.684751 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.684816 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.684826 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.684842 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.684853 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:54Z","lastTransitionTime":"2026-02-17T13:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.787489 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.787549 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.787566 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.787592 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.787610 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:54Z","lastTransitionTime":"2026-02-17T13:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.890056 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.890100 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.890114 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.890128 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.890138 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:54Z","lastTransitionTime":"2026-02-17T13:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.993328 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.993441 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.993470 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.993497 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:54 crc kubenswrapper[4955]: I0217 13:05:54.993516 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:54Z","lastTransitionTime":"2026-02-17T13:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.096409 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.096460 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.096476 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.096501 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.096518 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:55Z","lastTransitionTime":"2026-02-17T13:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.199666 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.199715 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.199731 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.199753 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.199772 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:55Z","lastTransitionTime":"2026-02-17T13:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.222408 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.222448 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.222468 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:05:55 crc kubenswrapper[4955]: E0217 13:05:55.222563 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:05:55 crc kubenswrapper[4955]: E0217 13:05:55.222676 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:05:55 crc kubenswrapper[4955]: E0217 13:05:55.222901 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.266333 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 14:19:30.906444635 +0000 UTC Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.302519 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.302556 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.302566 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.302583 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.302594 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:55Z","lastTransitionTime":"2026-02-17T13:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.408272 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.408326 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.408346 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.408376 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.408395 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:55Z","lastTransitionTime":"2026-02-17T13:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.511017 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.511069 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.511084 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.511111 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.511132 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:55Z","lastTransitionTime":"2026-02-17T13:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.616442 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.616498 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.616515 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.616539 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.616556 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:55Z","lastTransitionTime":"2026-02-17T13:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.718766 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.719234 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.719383 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.719518 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.719642 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:55Z","lastTransitionTime":"2026-02-17T13:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.823187 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.823582 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.823644 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.823734 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.824201 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:55Z","lastTransitionTime":"2026-02-17T13:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.927328 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.927372 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.927382 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.927400 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:55 crc kubenswrapper[4955]: I0217 13:05:55.927416 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:55Z","lastTransitionTime":"2026-02-17T13:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.031946 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.032014 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.032034 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.032064 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.032083 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:56Z","lastTransitionTime":"2026-02-17T13:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.135074 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.135431 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.135644 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.135867 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.136148 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:56Z","lastTransitionTime":"2026-02-17T13:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.222888 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:05:56 crc kubenswrapper[4955]: E0217 13:05:56.223140 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.239593 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.239762 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.239870 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.239943 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.240042 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:56Z","lastTransitionTime":"2026-02-17T13:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.267059 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 15:59:51.482884031 +0000 UTC Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.343152 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.343220 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.343237 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.343265 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.343288 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:56Z","lastTransitionTime":"2026-02-17T13:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.447554 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.447611 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.447623 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.447645 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.447658 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:56Z","lastTransitionTime":"2026-02-17T13:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.550751 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.550848 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.550867 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.550893 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.550913 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:56Z","lastTransitionTime":"2026-02-17T13:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.653581 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.653635 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.653647 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.653667 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.653680 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:56Z","lastTransitionTime":"2026-02-17T13:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.756649 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.756689 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.756699 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.756717 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.756729 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:56Z","lastTransitionTime":"2026-02-17T13:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.860470 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.860548 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.860571 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.860602 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.860622 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:56Z","lastTransitionTime":"2026-02-17T13:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.964170 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.964258 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.964282 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.964315 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:56 crc kubenswrapper[4955]: I0217 13:05:56.964339 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:56Z","lastTransitionTime":"2026-02-17T13:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.068362 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.068412 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.068427 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.068448 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.068459 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:57Z","lastTransitionTime":"2026-02-17T13:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.171702 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.171748 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.171761 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.171793 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.171806 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:57Z","lastTransitionTime":"2026-02-17T13:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.222102 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:05:57 crc kubenswrapper[4955]: E0217 13:05:57.222707 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.222205 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:05:57 crc kubenswrapper[4955]: E0217 13:05:57.222976 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.222187 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:05:57 crc kubenswrapper[4955]: E0217 13:05:57.223185 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.268043 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 15:36:16.663935085 +0000 UTC Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.274463 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.274520 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.274540 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.274564 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.274581 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:57Z","lastTransitionTime":"2026-02-17T13:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.377114 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.377171 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.377191 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.377214 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.377232 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:57Z","lastTransitionTime":"2026-02-17T13:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.480518 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.480589 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.480609 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.480633 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.480652 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:57Z","lastTransitionTime":"2026-02-17T13:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.583574 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.583625 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.583639 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.583659 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.583675 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:57Z","lastTransitionTime":"2026-02-17T13:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.687544 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.687894 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.688001 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.688108 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.688198 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:57Z","lastTransitionTime":"2026-02-17T13:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.790034 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.790322 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.790414 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.790506 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.790585 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:57Z","lastTransitionTime":"2026-02-17T13:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.893421 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.893470 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.893489 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.893515 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.893534 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:57Z","lastTransitionTime":"2026-02-17T13:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.996947 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.996988 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.997005 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.997022 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:57 crc kubenswrapper[4955]: I0217 13:05:57.997033 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:57Z","lastTransitionTime":"2026-02-17T13:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.100986 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.101033 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.101051 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.101073 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.101091 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:58Z","lastTransitionTime":"2026-02-17T13:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.204837 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.204889 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.204906 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.204932 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.204952 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:58Z","lastTransitionTime":"2026-02-17T13:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.222648 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:05:58 crc kubenswrapper[4955]: E0217 13:05:58.222935 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.268938 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 05:31:05.38720285 +0000 UTC Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.308134 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.308193 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.308206 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.308231 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.308246 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:58Z","lastTransitionTime":"2026-02-17T13:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.412055 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.412111 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.412124 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.412144 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.412162 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:58Z","lastTransitionTime":"2026-02-17T13:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.515531 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.515598 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.515620 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.515647 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.515668 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:58Z","lastTransitionTime":"2026-02-17T13:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.619356 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.619414 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.619432 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.619459 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.619479 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:58Z","lastTransitionTime":"2026-02-17T13:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.723144 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.723215 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.723226 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.723253 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.723662 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:58Z","lastTransitionTime":"2026-02-17T13:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.827375 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.827450 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.827468 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.827496 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.827517 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:58Z","lastTransitionTime":"2026-02-17T13:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.934457 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.934511 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.934525 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.934548 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:58 crc kubenswrapper[4955]: I0217 13:05:58.934564 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:58Z","lastTransitionTime":"2026-02-17T13:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.037804 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.037866 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.037878 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.037903 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.037917 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:59Z","lastTransitionTime":"2026-02-17T13:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.141176 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.141236 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.141248 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.141267 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.141281 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:59Z","lastTransitionTime":"2026-02-17T13:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.222114 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.222117 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:05:59 crc kubenswrapper[4955]: E0217 13:05:59.222280 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.222133 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:05:59 crc kubenswrapper[4955]: E0217 13:05:59.222394 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:05:59 crc kubenswrapper[4955]: E0217 13:05:59.222487 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.244455 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.244508 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.244529 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.244553 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.244572 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:59Z","lastTransitionTime":"2026-02-17T13:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.269476 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 18:32:30.813162428 +0000 UTC Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.349504 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.349567 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.349586 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.349657 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.349678 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:59Z","lastTransitionTime":"2026-02-17T13:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.453157 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.453416 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.453434 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.453461 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.453482 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:59Z","lastTransitionTime":"2026-02-17T13:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.556231 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.556284 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.556298 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.556318 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.556332 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:59Z","lastTransitionTime":"2026-02-17T13:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.659541 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.659637 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.659649 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.659664 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.659674 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:59Z","lastTransitionTime":"2026-02-17T13:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.763264 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.763327 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.763341 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.763362 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.763375 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:59Z","lastTransitionTime":"2026-02-17T13:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.866677 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.866742 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.866764 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.866828 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.866854 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:59Z","lastTransitionTime":"2026-02-17T13:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.969496 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.970056 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.970804 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.970824 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:05:59 crc kubenswrapper[4955]: I0217 13:05:59.970834 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:05:59Z","lastTransitionTime":"2026-02-17T13:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.073676 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.073713 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.073724 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.073744 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.073756 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:00Z","lastTransitionTime":"2026-02-17T13:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.177254 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.177324 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.177341 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.177367 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.177385 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:00Z","lastTransitionTime":"2026-02-17T13:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.222607 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:06:00 crc kubenswrapper[4955]: E0217 13:06:00.222776 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.224557 4955 scope.go:117] "RemoveContainer" containerID="528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6" Feb 17 13:06:00 crc kubenswrapper[4955]: E0217 13:06:00.224952 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dcwgg_openshift-ovn-kubernetes(dcbc7a69-58e7-4b7a-bb84-96276819ce88)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.270477 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 05:36:24.188091924 +0000 UTC Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.279896 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.280140 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.280273 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.280412 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.280544 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:00Z","lastTransitionTime":"2026-02-17T13:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.383925 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.383999 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.384022 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.384056 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.384080 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:00Z","lastTransitionTime":"2026-02-17T13:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.487164 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.487654 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.487875 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.488027 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.488206 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:00Z","lastTransitionTime":"2026-02-17T13:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.591968 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.592055 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.592077 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.592110 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.592134 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:00Z","lastTransitionTime":"2026-02-17T13:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.695194 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.695243 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.695258 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.695280 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.695292 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:00Z","lastTransitionTime":"2026-02-17T13:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.798570 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.798812 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.798876 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.798942 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.799013 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:00Z","lastTransitionTime":"2026-02-17T13:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.901698 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.901988 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.902113 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.902205 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:00 crc kubenswrapper[4955]: I0217 13:06:00.902290 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:00Z","lastTransitionTime":"2026-02-17T13:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.005333 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.005414 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.005446 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.005479 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.005502 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:01Z","lastTransitionTime":"2026-02-17T13:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.108040 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.108080 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.108096 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.108114 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.108126 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:01Z","lastTransitionTime":"2026-02-17T13:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.210892 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.210932 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.210942 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.210957 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.210967 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:01Z","lastTransitionTime":"2026-02-17T13:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.219289 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.219373 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.219391 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.219445 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.219462 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:01Z","lastTransitionTime":"2026-02-17T13:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.222475 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.222558 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:06:01 crc kubenswrapper[4955]: E0217 13:06:01.222589 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:06:01 crc kubenswrapper[4955]: E0217 13:06:01.222774 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.222829 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:06:01 crc kubenswrapper[4955]: E0217 13:06:01.223008 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:06:01 crc kubenswrapper[4955]: E0217 13:06:01.238481 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:06:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:06:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:06:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:06:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7605a498-360b-4ce4-b7a1-dfbc354d798d\\\",\\\"systemUUID\\\":\\\"6d4602d9-c718-4393-8f91-3c381f800c9e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:06:01Z is after 2025-08-24T17:21:41Z" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.243705 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.243738 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.243752 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.243767 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.243799 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:01Z","lastTransitionTime":"2026-02-17T13:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:01 crc kubenswrapper[4955]: E0217 13:06:01.262253 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:06:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:06:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:06:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:06:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7605a498-360b-4ce4-b7a1-dfbc354d798d\\\",\\\"systemUUID\\\":\\\"6d4602d9-c718-4393-8f91-3c381f800c9e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:06:01Z is after 2025-08-24T17:21:41Z" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.266520 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.266546 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.266558 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.266573 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.266585 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:01Z","lastTransitionTime":"2026-02-17T13:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.271471 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 23:45:52.39181047 +0000 UTC Feb 17 13:06:01 crc kubenswrapper[4955]: E0217 13:06:01.284943 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:06:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:06:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:06:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:06:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7605a498-360b-4ce4-b7a1-dfbc354d798d\\\",\\\"systemUUID\\\":\\\"6d4602d9-c718-4393-8f91-3c381f800c9e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:06:01Z is after 2025-08-24T17:21:41Z" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.290769 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.290814 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.290828 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.290844 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.290856 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:01Z","lastTransitionTime":"2026-02-17T13:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:01 crc kubenswrapper[4955]: E0217 13:06:01.306925 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:06:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:06:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:06:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:06:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7605a498-360b-4ce4-b7a1-dfbc354d798d\\\",\\\"systemUUID\\\":\\\"6d4602d9-c718-4393-8f91-3c381f800c9e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:06:01Z is after 2025-08-24T17:21:41Z" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.317461 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.317525 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.317545 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.317572 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.317591 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:01Z","lastTransitionTime":"2026-02-17T13:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:01 crc kubenswrapper[4955]: E0217 13:06:01.339492 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:06:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:06:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:06:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-17T13:06:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7605a498-360b-4ce4-b7a1-dfbc354d798d\\\",\\\"systemUUID\\\":\\\"6d4602d9-c718-4393-8f91-3c381f800c9e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:06:01Z is after 2025-08-24T17:21:41Z" Feb 17 13:06:01 crc kubenswrapper[4955]: E0217 13:06:01.339708 4955 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.342609 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.342684 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.342699 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.342728 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.342744 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:01Z","lastTransitionTime":"2026-02-17T13:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.446440 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.446498 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.446516 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.446540 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.446560 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:01Z","lastTransitionTime":"2026-02-17T13:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.549568 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.549628 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.549638 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.549658 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.549669 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:01Z","lastTransitionTime":"2026-02-17T13:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.653512 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.653577 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.653592 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.653613 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.653627 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:01Z","lastTransitionTime":"2026-02-17T13:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.756667 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.756716 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.756727 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.756747 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.756760 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:01Z","lastTransitionTime":"2026-02-17T13:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.860590 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.860647 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.860658 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.860678 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.860692 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:01Z","lastTransitionTime":"2026-02-17T13:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.963763 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.963852 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.963871 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.963900 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:01 crc kubenswrapper[4955]: I0217 13:06:01.963919 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:01Z","lastTransitionTime":"2026-02-17T13:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.067576 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.067633 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.067646 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.067664 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.067676 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:02Z","lastTransitionTime":"2026-02-17T13:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.170342 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.170402 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.170415 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.170434 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.170447 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:02Z","lastTransitionTime":"2026-02-17T13:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.222309 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:06:02 crc kubenswrapper[4955]: E0217 13:06:02.222892 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.238561 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0e4cef3-bd0e-495f-9ba3-5415afeb528b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd99814f5b3ea3e8ad492e07f6050234b9394d3f775077207e41d44db4d63898\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de99882bedf59812423fca8186a527f03eec5abe5777ff95b421b9d0330747d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de99882bedf59812423fca8186a527f03eec5abe5777ff95b421b9d0330747d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:06:02Z is after 2025-08-24T17:21:41Z" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.253960 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b9ecd4380a6ae8ff62e27e3946815170c2f02b4c3921bda595467345289c9e0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:06:02Z is after 2025-08-24T17:21:41Z" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.270567 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0f4056e4b5a554f05004c194485d25231b67024e4376a73c715fcb32004a3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:06:02Z is after 2025-08-24T17:21:41Z" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.271628 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 02:19:07.170600063 +0000 UTC Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.273451 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.273485 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.273496 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.273539 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.273553 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:02Z","lastTransitionTime":"2026-02-17T13:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.289904 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d05aa39d-94e6-49d4-885e-2c1f986310f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-17T13:04:42Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0217 13:04:35.885968 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0217 13:04:35.888176 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3462920193/tls.crt::/tmp/serving-cert-3462920193/tls.key\\\\\\\"\\\\nI0217 13:04:42.370546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0217 13:04:42.384242 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0217 13:04:42.384284 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0217 13:04:42.384317 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0217 13:04:42.384327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0217 13:04:42.394555 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0217 13:04:42.394614 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394628 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0217 13:04:42.394638 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0217 13:04:42.394646 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0217 13:04:42.394654 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0217 13:04:42.394662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0217 13:04:42.394900 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0217 13:04:42.396494 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-17T13:04:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:06:02Z is after 2025-08-24T17:21:41Z" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.306758 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:06:02Z is after 2025-08-24T17:21:41Z" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.319675 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:06:02Z is after 2025-08-24T17:21:41Z" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.334111 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5bf36887-2a2b-4193-abdb-a15beaf46300\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11a8924eb6ce41d629feb2b42c7d4778a3ca9110a59c3ac25acbba4367272374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bec78f9e9f6416d6b600f6231f0d24652473b48b95bf0e0675cc85c9208d124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l299f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-29qxq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:06:02Z is after 2025-08-24T17:21:41Z" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.354856 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9hx9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e57bdf8-e537-402a-9815-609fecdd68ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:05:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c81948fc2806ea8718909fd269c029efc6a26675e45359e971272354eccfa74d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dbc9014d80aee39922161014aac8fafa7d93be3e4092fe705c76da4a487a5e5d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-17T13:05:31Z\\\",\\\"message\\\":\\\"2026-02-17T13:04:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c2764cb0-2a4e-4449-bac5-794fed2f307a\\\\n2026-02-17T13:04:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c2764cb0-2a4e-4449-bac5-794fed2f307a to /host/opt/cni/bin/\\\\n2026-02-17T13:04:46Z [verbose] multus-daemon started\\\\n2026-02-17T13:04:46Z [verbose] Readiness Indicator file check\\\\n2026-02-17T13:05:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:05:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q22d4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9hx9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:06:02Z is after 2025-08-24T17:21:41Z" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.371049 4955 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-msn6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8df54e9e-2c74-4b95-bbc8-0fdad6efc5fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-17T13:04:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://277a12d62c90ef420c31d8dc71f0d7750dfd5f5ab7e90fec0dcd00e54b60245d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-17T13:04:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2tvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-17T13:04:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-msn6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-17T13:06:02Z is after 2025-08-24T17:21:41Z" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.375883 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.375944 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.375962 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.375990 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.376014 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:02Z","lastTransitionTime":"2026-02-17T13:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.418905 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=77.418885986 podStartE2EDuration="1m17.418885986s" podCreationTimestamp="2026-02-17 13:04:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:02.416661984 +0000 UTC m=+100.939391517" watchObservedRunningTime="2026-02-17 13:06:02.418885986 +0000 UTC m=+100.941615529" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.464531 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=79.464505246 podStartE2EDuration="1m19.464505246s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:02.442697749 +0000 UTC m=+100.965427312" watchObservedRunningTime="2026-02-17 13:06:02.464505246 +0000 UTC m=+100.987234779" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.480044 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.480136 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.480153 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.480194 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.480212 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:02Z","lastTransitionTime":"2026-02-17T13:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.491275 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-fp5j6" podStartSLOduration=79.491261111 podStartE2EDuration="1m19.491261111s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:02.488766522 +0000 UTC m=+101.011496075" watchObservedRunningTime="2026-02-17 13:06:02.491261111 +0000 UTC m=+101.013990654" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.562437 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=53.562401473 podStartE2EDuration="53.562401473s" podCreationTimestamp="2026-02-17 13:05:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:02.543110275 +0000 UTC m=+101.065839858" watchObservedRunningTime="2026-02-17 13:06:02.562401473 +0000 UTC m=+101.085131046" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.577727 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-2gsx6" podStartSLOduration=79.577700879 podStartE2EDuration="1m19.577700879s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:02.577335469 +0000 UTC m=+101.100065052" watchObservedRunningTime="2026-02-17 13:06:02.577700879 +0000 UTC m=+101.100430452" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.585928 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.585979 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.585990 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.586011 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.586026 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:02Z","lastTransitionTime":"2026-02-17T13:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.588640 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/77480847-3271-4161-a833-e62b3bb4413e-metrics-certs\") pod \"network-metrics-daemon-2k5pz\" (UID: \"77480847-3271-4161-a833-e62b3bb4413e\") " pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:06:02 crc kubenswrapper[4955]: E0217 13:06:02.588849 4955 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 13:06:02 crc kubenswrapper[4955]: E0217 13:06:02.588956 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/77480847-3271-4161-a833-e62b3bb4413e-metrics-certs podName:77480847-3271-4161-a833-e62b3bb4413e nodeName:}" failed. No retries permitted until 2026-02-17 13:07:06.588928241 +0000 UTC m=+165.111657814 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/77480847-3271-4161-a833-e62b3bb4413e-metrics-certs") pod "network-metrics-daemon-2k5pz" (UID: "77480847-3271-4161-a833-e62b3bb4413e") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.596705 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lffhw" podStartSLOduration=79.596686907 podStartE2EDuration="1m19.596686907s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:02.59571925 +0000 UTC m=+101.118448823" watchObservedRunningTime="2026-02-17 13:06:02.596686907 +0000 UTC m=+101.119416490" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.688884 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.688926 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.688935 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.688950 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.688961 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:02Z","lastTransitionTime":"2026-02-17T13:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.790992 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.791078 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.791095 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.791121 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.791140 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:02Z","lastTransitionTime":"2026-02-17T13:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.893845 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.893906 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.893928 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.893956 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.893976 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:02Z","lastTransitionTime":"2026-02-17T13:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.997191 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.997237 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.997253 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.997276 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:02 crc kubenswrapper[4955]: I0217 13:06:02.997294 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:02Z","lastTransitionTime":"2026-02-17T13:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.100063 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.100115 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.100134 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.100158 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.100177 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:03Z","lastTransitionTime":"2026-02-17T13:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.203641 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.203727 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.203763 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.203827 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.203854 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:03Z","lastTransitionTime":"2026-02-17T13:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.222142 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.222182 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.222202 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:06:03 crc kubenswrapper[4955]: E0217 13:06:03.222333 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:06:03 crc kubenswrapper[4955]: E0217 13:06:03.222505 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:06:03 crc kubenswrapper[4955]: E0217 13:06:03.222657 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.272884 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 03:01:52.116258246 +0000 UTC Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.307153 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.307225 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.307248 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.307280 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.307304 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:03Z","lastTransitionTime":"2026-02-17T13:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.410417 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.410489 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.410501 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.410519 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.410530 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:03Z","lastTransitionTime":"2026-02-17T13:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.514348 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.514454 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.514492 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.514529 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.514569 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:03Z","lastTransitionTime":"2026-02-17T13:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.618221 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.618286 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.618310 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.618347 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.618369 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:03Z","lastTransitionTime":"2026-02-17T13:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.722290 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.722334 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.722359 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.722377 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.722392 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:03Z","lastTransitionTime":"2026-02-17T13:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.825226 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.825295 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.825316 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.825348 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.825369 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:03Z","lastTransitionTime":"2026-02-17T13:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.928463 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.928543 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.928560 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.928587 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:03 crc kubenswrapper[4955]: I0217 13:06:03.928613 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:03Z","lastTransitionTime":"2026-02-17T13:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.032120 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.032196 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.032214 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.032241 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.032259 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:04Z","lastTransitionTime":"2026-02-17T13:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.135553 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.135625 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.135645 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.135673 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.135693 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:04Z","lastTransitionTime":"2026-02-17T13:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.222586 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:06:04 crc kubenswrapper[4955]: E0217 13:06:04.222850 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.238190 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.238232 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.238242 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.238261 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.238272 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:04Z","lastTransitionTime":"2026-02-17T13:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.273907 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 04:16:49.900600552 +0000 UTC Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.342408 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.342480 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.342499 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.342525 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.342546 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:04Z","lastTransitionTime":"2026-02-17T13:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.446044 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.446100 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.446118 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.446143 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.446159 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:04Z","lastTransitionTime":"2026-02-17T13:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.549313 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.549386 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.549410 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.549443 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.549467 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:04Z","lastTransitionTime":"2026-02-17T13:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.652547 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.652642 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.652660 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.652689 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.652715 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:04Z","lastTransitionTime":"2026-02-17T13:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.755224 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.755501 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.755532 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.755562 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.755585 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:04Z","lastTransitionTime":"2026-02-17T13:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.858189 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.858238 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.858248 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.858267 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.858278 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:04Z","lastTransitionTime":"2026-02-17T13:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.960420 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.960488 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.960512 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.960550 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:04 crc kubenswrapper[4955]: I0217 13:06:04.960575 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:04Z","lastTransitionTime":"2026-02-17T13:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.063227 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.063272 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.063283 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.063303 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.063315 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:05Z","lastTransitionTime":"2026-02-17T13:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.166269 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.166350 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.166379 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.166417 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.166441 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:05Z","lastTransitionTime":"2026-02-17T13:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.222425 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.222465 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.222463 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:06:05 crc kubenswrapper[4955]: E0217 13:06:05.222745 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:06:05 crc kubenswrapper[4955]: E0217 13:06:05.222923 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:06:05 crc kubenswrapper[4955]: E0217 13:06:05.223083 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.269823 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.269875 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.269884 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.269901 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.269911 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:05Z","lastTransitionTime":"2026-02-17T13:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.275077 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 04:12:40.511262793 +0000 UTC Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.372668 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.373011 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.373023 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.373041 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.373055 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:05Z","lastTransitionTime":"2026-02-17T13:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.475426 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.475463 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.475473 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.475488 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.475498 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:05Z","lastTransitionTime":"2026-02-17T13:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.578504 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.578566 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.578586 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.578611 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.578630 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:05Z","lastTransitionTime":"2026-02-17T13:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.681485 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.681540 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.681553 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.681574 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.681588 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:05Z","lastTransitionTime":"2026-02-17T13:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.784063 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.784112 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.784122 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.784147 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.784170 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:05Z","lastTransitionTime":"2026-02-17T13:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.887499 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.887536 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.887553 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.887575 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.887586 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:05Z","lastTransitionTime":"2026-02-17T13:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.990221 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.990267 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.990278 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.990295 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:05 crc kubenswrapper[4955]: I0217 13:06:05.990305 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:05Z","lastTransitionTime":"2026-02-17T13:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.093216 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.093257 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.093270 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.093289 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.093301 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:06Z","lastTransitionTime":"2026-02-17T13:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.196082 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.196127 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.196141 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.196159 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.196171 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:06Z","lastTransitionTime":"2026-02-17T13:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.222668 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:06:06 crc kubenswrapper[4955]: E0217 13:06:06.222862 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.275480 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 06:40:32.289480777 +0000 UTC Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.299266 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.299325 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.299351 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.299373 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.299389 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:06Z","lastTransitionTime":"2026-02-17T13:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.402569 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.402627 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.402637 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.402654 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.402666 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:06Z","lastTransitionTime":"2026-02-17T13:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.505557 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.505614 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.505624 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.505643 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.505657 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:06Z","lastTransitionTime":"2026-02-17T13:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.614428 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.614554 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.614567 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.614588 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.614624 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:06Z","lastTransitionTime":"2026-02-17T13:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.717448 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.717521 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.717555 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.717589 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.717615 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:06Z","lastTransitionTime":"2026-02-17T13:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.820374 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.820421 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.820431 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.820446 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.820457 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:06Z","lastTransitionTime":"2026-02-17T13:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.923292 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.923352 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.923368 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.923392 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:06 crc kubenswrapper[4955]: I0217 13:06:06.923409 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:06Z","lastTransitionTime":"2026-02-17T13:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.026395 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.026456 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.026470 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.026490 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.026505 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:07Z","lastTransitionTime":"2026-02-17T13:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.129538 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.129601 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.129621 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.129645 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.129660 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:07Z","lastTransitionTime":"2026-02-17T13:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.222525 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.222623 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.222586 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:06:07 crc kubenswrapper[4955]: E0217 13:06:07.222765 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:06:07 crc kubenswrapper[4955]: E0217 13:06:07.223062 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:06:07 crc kubenswrapper[4955]: E0217 13:06:07.223121 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.232464 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.232522 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.232550 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.232583 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.232609 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:07Z","lastTransitionTime":"2026-02-17T13:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.276300 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-30 21:06:49.403890733 +0000 UTC Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.335895 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.335954 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.335969 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.335991 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.336011 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:07Z","lastTransitionTime":"2026-02-17T13:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.439300 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.439342 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.439353 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.439370 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.439381 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:07Z","lastTransitionTime":"2026-02-17T13:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.542050 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.542127 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.542146 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.542176 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.542199 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:07Z","lastTransitionTime":"2026-02-17T13:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.644783 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.644855 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.644869 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.644893 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.644910 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:07Z","lastTransitionTime":"2026-02-17T13:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.747374 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.747438 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.747456 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.747480 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.747498 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:07Z","lastTransitionTime":"2026-02-17T13:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.850705 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.850842 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.850862 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.850890 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.850912 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:07Z","lastTransitionTime":"2026-02-17T13:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.954476 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.954527 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.954541 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.954562 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:07 crc kubenswrapper[4955]: I0217 13:06:07.954576 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:07Z","lastTransitionTime":"2026-02-17T13:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.058597 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.058670 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.058690 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.058722 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.058746 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:08Z","lastTransitionTime":"2026-02-17T13:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.164013 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.164082 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.164100 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.164127 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.164144 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:08Z","lastTransitionTime":"2026-02-17T13:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.222570 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:06:08 crc kubenswrapper[4955]: E0217 13:06:08.223106 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.268343 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.268412 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.268423 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.268441 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.268482 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:08Z","lastTransitionTime":"2026-02-17T13:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.276872 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 01:51:46.712257489 +0000 UTC Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.374106 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.374147 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.374160 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.374177 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.374189 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:08Z","lastTransitionTime":"2026-02-17T13:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.476922 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.476989 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.477010 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.477038 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.477058 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:08Z","lastTransitionTime":"2026-02-17T13:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.579878 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.579946 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.579956 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.579973 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.579984 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:08Z","lastTransitionTime":"2026-02-17T13:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.683161 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.683214 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.683226 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.683244 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.683256 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:08Z","lastTransitionTime":"2026-02-17T13:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.786082 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.786131 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.786143 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.786162 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.786176 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:08Z","lastTransitionTime":"2026-02-17T13:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.888477 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.888544 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.888573 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.888607 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.888644 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:08Z","lastTransitionTime":"2026-02-17T13:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.994878 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.994952 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.994971 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.995000 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:08 crc kubenswrapper[4955]: I0217 13:06:08.995019 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:08Z","lastTransitionTime":"2026-02-17T13:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.097508 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.097570 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.097592 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.097624 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.097648 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:09Z","lastTransitionTime":"2026-02-17T13:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.200372 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.200429 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.200440 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.200460 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.200474 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:09Z","lastTransitionTime":"2026-02-17T13:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.222733 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.222823 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.222751 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:06:09 crc kubenswrapper[4955]: E0217 13:06:09.222987 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:06:09 crc kubenswrapper[4955]: E0217 13:06:09.223181 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:06:09 crc kubenswrapper[4955]: E0217 13:06:09.223367 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.277323 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-04 19:03:09.279776114 +0000 UTC Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.303923 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.303977 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.303991 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.304013 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.304028 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:09Z","lastTransitionTime":"2026-02-17T13:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.407373 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.407414 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.407440 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.407458 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.407469 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:09Z","lastTransitionTime":"2026-02-17T13:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.511344 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.511410 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.511420 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.511443 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.511460 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:09Z","lastTransitionTime":"2026-02-17T13:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.614929 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.614994 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.615011 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.615039 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.615058 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:09Z","lastTransitionTime":"2026-02-17T13:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.717610 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.717685 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.717707 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.717731 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.717749 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:09Z","lastTransitionTime":"2026-02-17T13:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.821559 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.821617 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.821635 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.821661 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.821682 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:09Z","lastTransitionTime":"2026-02-17T13:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.923962 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.924006 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.924016 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.924034 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:09 crc kubenswrapper[4955]: I0217 13:06:09.924045 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:09Z","lastTransitionTime":"2026-02-17T13:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.026587 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.026631 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.026641 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.026658 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.026671 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:10Z","lastTransitionTime":"2026-02-17T13:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.130148 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.130224 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.130257 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.130290 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.130315 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:10Z","lastTransitionTime":"2026-02-17T13:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.222201 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:06:10 crc kubenswrapper[4955]: E0217 13:06:10.222427 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.232660 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.232704 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.232714 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.232728 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.232741 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:10Z","lastTransitionTime":"2026-02-17T13:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.278387 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 14:39:43.680474802 +0000 UTC Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.336039 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.336106 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.336129 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.336157 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.336177 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:10Z","lastTransitionTime":"2026-02-17T13:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.440235 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.440316 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.440334 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.440364 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.440386 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:10Z","lastTransitionTime":"2026-02-17T13:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.544310 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.544394 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.544407 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.544426 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.544443 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:10Z","lastTransitionTime":"2026-02-17T13:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.647437 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.647491 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.647507 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.647530 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.647545 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:10Z","lastTransitionTime":"2026-02-17T13:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.751075 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.751127 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.751139 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.751158 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.751173 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:10Z","lastTransitionTime":"2026-02-17T13:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.854405 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.854459 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.854472 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.854488 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.854500 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:10Z","lastTransitionTime":"2026-02-17T13:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.957327 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.957409 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.957433 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.957464 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:10 crc kubenswrapper[4955]: I0217 13:06:10.957486 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:10Z","lastTransitionTime":"2026-02-17T13:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.060549 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.060589 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.060600 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.060618 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.060630 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:11Z","lastTransitionTime":"2026-02-17T13:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.164074 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.164137 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.164155 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.164182 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.164203 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:11Z","lastTransitionTime":"2026-02-17T13:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.222667 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.222700 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.222742 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:06:11 crc kubenswrapper[4955]: E0217 13:06:11.222929 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:06:11 crc kubenswrapper[4955]: E0217 13:06:11.223092 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:06:11 crc kubenswrapper[4955]: E0217 13:06:11.223254 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.267283 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.267320 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.267330 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.267346 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.267362 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:11Z","lastTransitionTime":"2026-02-17T13:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.279075 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 10:03:09.07257798 +0000 UTC Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.371002 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.371069 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.371088 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.371112 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.371132 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:11Z","lastTransitionTime":"2026-02-17T13:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.457839 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.457924 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.457958 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.457991 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.458018 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:11Z","lastTransitionTime":"2026-02-17T13:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.478726 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.478798 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.478818 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.478841 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.478857 4955 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-17T13:06:11Z","lastTransitionTime":"2026-02-17T13:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.517782 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-p5hd6"] Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.518198 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p5hd6" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.522083 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.522575 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.522957 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.523656 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.536104 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-msn6p" podStartSLOduration=89.53606676 podStartE2EDuration="1m29.53606676s" podCreationTimestamp="2026-02-17 13:04:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:11.536028319 +0000 UTC m=+110.058757852" watchObservedRunningTime="2026-02-17 13:06:11.53606676 +0000 UTC m=+110.058796333" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.569979 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=88.569952393 podStartE2EDuration="1m28.569952393s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:11.565573331 +0000 UTC m=+110.088302874" watchObservedRunningTime="2026-02-17 13:06:11.569952393 +0000 UTC m=+110.092681936" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.577103 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3bb34a49-00a8-4e19-bd70-423102da93e0-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-p5hd6\" (UID: \"3bb34a49-00a8-4e19-bd70-423102da93e0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p5hd6" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.577168 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/3bb34a49-00a8-4e19-bd70-423102da93e0-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-p5hd6\" (UID: \"3bb34a49-00a8-4e19-bd70-423102da93e0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p5hd6" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.577190 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/3bb34a49-00a8-4e19-bd70-423102da93e0-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-p5hd6\" (UID: \"3bb34a49-00a8-4e19-bd70-423102da93e0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p5hd6" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.577403 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3bb34a49-00a8-4e19-bd70-423102da93e0-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-p5hd6\" (UID: \"3bb34a49-00a8-4e19-bd70-423102da93e0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p5hd6" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.577471 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3bb34a49-00a8-4e19-bd70-423102da93e0-service-ca\") pod \"cluster-version-operator-5c965bbfc6-p5hd6\" (UID: \"3bb34a49-00a8-4e19-bd70-423102da93e0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p5hd6" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.640627 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-9hx9f" podStartSLOduration=88.640602831 podStartE2EDuration="1m28.640602831s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:11.640057846 +0000 UTC m=+110.162787399" watchObservedRunningTime="2026-02-17 13:06:11.640602831 +0000 UTC m=+110.163332374" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.640901 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podStartSLOduration=88.640894849 podStartE2EDuration="1m28.640894849s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:11.621706325 +0000 UTC m=+110.144435878" watchObservedRunningTime="2026-02-17 13:06:11.640894849 +0000 UTC m=+110.163624392" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.677905 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3bb34a49-00a8-4e19-bd70-423102da93e0-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-p5hd6\" (UID: \"3bb34a49-00a8-4e19-bd70-423102da93e0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p5hd6" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.677963 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/3bb34a49-00a8-4e19-bd70-423102da93e0-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-p5hd6\" (UID: \"3bb34a49-00a8-4e19-bd70-423102da93e0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p5hd6" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.677986 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/3bb34a49-00a8-4e19-bd70-423102da93e0-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-p5hd6\" (UID: \"3bb34a49-00a8-4e19-bd70-423102da93e0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p5hd6" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.678024 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3bb34a49-00a8-4e19-bd70-423102da93e0-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-p5hd6\" (UID: \"3bb34a49-00a8-4e19-bd70-423102da93e0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p5hd6" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.678053 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3bb34a49-00a8-4e19-bd70-423102da93e0-service-ca\") pod \"cluster-version-operator-5c965bbfc6-p5hd6\" (UID: \"3bb34a49-00a8-4e19-bd70-423102da93e0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p5hd6" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.678955 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3bb34a49-00a8-4e19-bd70-423102da93e0-service-ca\") pod \"cluster-version-operator-5c965bbfc6-p5hd6\" (UID: \"3bb34a49-00a8-4e19-bd70-423102da93e0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p5hd6" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.679499 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/3bb34a49-00a8-4e19-bd70-423102da93e0-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-p5hd6\" (UID: \"3bb34a49-00a8-4e19-bd70-423102da93e0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p5hd6" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.679594 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/3bb34a49-00a8-4e19-bd70-423102da93e0-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-p5hd6\" (UID: \"3bb34a49-00a8-4e19-bd70-423102da93e0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p5hd6" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.683947 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3bb34a49-00a8-4e19-bd70-423102da93e0-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-p5hd6\" (UID: \"3bb34a49-00a8-4e19-bd70-423102da93e0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p5hd6" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.699363 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=19.699344277 podStartE2EDuration="19.699344277s" podCreationTimestamp="2026-02-17 13:05:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:11.676661495 +0000 UTC m=+110.199391048" watchObservedRunningTime="2026-02-17 13:06:11.699344277 +0000 UTC m=+110.222073820" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.700040 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3bb34a49-00a8-4e19-bd70-423102da93e0-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-p5hd6\" (UID: \"3bb34a49-00a8-4e19-bd70-423102da93e0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p5hd6" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.836924 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p5hd6" Feb 17 13:06:11 crc kubenswrapper[4955]: I0217 13:06:11.931753 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p5hd6" event={"ID":"3bb34a49-00a8-4e19-bd70-423102da93e0","Type":"ContainerStarted","Data":"86e513b23bdd289783f0f389708e0f2670b1d34ccffa64d13a873b4b879c4347"} Feb 17 13:06:12 crc kubenswrapper[4955]: I0217 13:06:12.222775 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:06:12 crc kubenswrapper[4955]: E0217 13:06:12.223890 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:06:12 crc kubenswrapper[4955]: I0217 13:06:12.224820 4955 scope.go:117] "RemoveContainer" containerID="528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6" Feb 17 13:06:12 crc kubenswrapper[4955]: E0217 13:06:12.224995 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dcwgg_openshift-ovn-kubernetes(dcbc7a69-58e7-4b7a-bb84-96276819ce88)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" Feb 17 13:06:12 crc kubenswrapper[4955]: I0217 13:06:12.279725 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-03 23:54:14.738563135 +0000 UTC Feb 17 13:06:12 crc kubenswrapper[4955]: I0217 13:06:12.279860 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Feb 17 13:06:12 crc kubenswrapper[4955]: I0217 13:06:12.291199 4955 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 17 13:06:12 crc kubenswrapper[4955]: I0217 13:06:12.935694 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p5hd6" event={"ID":"3bb34a49-00a8-4e19-bd70-423102da93e0","Type":"ContainerStarted","Data":"d4a6ed616c1e44f9c982579df5c7876d7978c767d069d40b93c459f641edacab"} Feb 17 13:06:12 crc kubenswrapper[4955]: I0217 13:06:12.959832 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p5hd6" podStartSLOduration=89.959763516 podStartE2EDuration="1m29.959763516s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:12.9563191 +0000 UTC m=+111.479048643" watchObservedRunningTime="2026-02-17 13:06:12.959763516 +0000 UTC m=+111.482493079" Feb 17 13:06:13 crc kubenswrapper[4955]: I0217 13:06:13.222344 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:06:13 crc kubenswrapper[4955]: I0217 13:06:13.222412 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:06:13 crc kubenswrapper[4955]: I0217 13:06:13.222416 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:06:13 crc kubenswrapper[4955]: E0217 13:06:13.222592 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:06:13 crc kubenswrapper[4955]: E0217 13:06:13.222765 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:06:13 crc kubenswrapper[4955]: E0217 13:06:13.222856 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:06:14 crc kubenswrapper[4955]: I0217 13:06:14.222320 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:06:14 crc kubenswrapper[4955]: E0217 13:06:14.222583 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:06:15 crc kubenswrapper[4955]: I0217 13:06:15.222584 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:06:15 crc kubenswrapper[4955]: I0217 13:06:15.222640 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:06:15 crc kubenswrapper[4955]: E0217 13:06:15.223051 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:06:15 crc kubenswrapper[4955]: I0217 13:06:15.222668 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:06:15 crc kubenswrapper[4955]: E0217 13:06:15.223150 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:06:15 crc kubenswrapper[4955]: E0217 13:06:15.223528 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:06:16 crc kubenswrapper[4955]: I0217 13:06:16.222318 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:06:16 crc kubenswrapper[4955]: E0217 13:06:16.222495 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:06:17 crc kubenswrapper[4955]: I0217 13:06:17.222913 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:06:17 crc kubenswrapper[4955]: I0217 13:06:17.222991 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:06:17 crc kubenswrapper[4955]: I0217 13:06:17.222930 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:06:17 crc kubenswrapper[4955]: E0217 13:06:17.223202 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:06:17 crc kubenswrapper[4955]: E0217 13:06:17.223363 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:06:17 crc kubenswrapper[4955]: E0217 13:06:17.223512 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:06:18 crc kubenswrapper[4955]: I0217 13:06:18.222999 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:06:18 crc kubenswrapper[4955]: E0217 13:06:18.223242 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:06:18 crc kubenswrapper[4955]: I0217 13:06:18.957115 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9hx9f_1e57bdf8-e537-402a-9815-609fecdd68ea/kube-multus/1.log" Feb 17 13:06:18 crc kubenswrapper[4955]: I0217 13:06:18.958006 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9hx9f_1e57bdf8-e537-402a-9815-609fecdd68ea/kube-multus/0.log" Feb 17 13:06:18 crc kubenswrapper[4955]: I0217 13:06:18.958073 4955 generic.go:334] "Generic (PLEG): container finished" podID="1e57bdf8-e537-402a-9815-609fecdd68ea" containerID="c81948fc2806ea8718909fd269c029efc6a26675e45359e971272354eccfa74d" exitCode=1 Feb 17 13:06:18 crc kubenswrapper[4955]: I0217 13:06:18.958119 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-9hx9f" event={"ID":"1e57bdf8-e537-402a-9815-609fecdd68ea","Type":"ContainerDied","Data":"c81948fc2806ea8718909fd269c029efc6a26675e45359e971272354eccfa74d"} Feb 17 13:06:18 crc kubenswrapper[4955]: I0217 13:06:18.958169 4955 scope.go:117] "RemoveContainer" containerID="dbc9014d80aee39922161014aac8fafa7d93be3e4092fe705c76da4a487a5e5d" Feb 17 13:06:18 crc kubenswrapper[4955]: I0217 13:06:18.959209 4955 scope.go:117] "RemoveContainer" containerID="c81948fc2806ea8718909fd269c029efc6a26675e45359e971272354eccfa74d" Feb 17 13:06:18 crc kubenswrapper[4955]: E0217 13:06:18.959535 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-9hx9f_openshift-multus(1e57bdf8-e537-402a-9815-609fecdd68ea)\"" pod="openshift-multus/multus-9hx9f" podUID="1e57bdf8-e537-402a-9815-609fecdd68ea" Feb 17 13:06:19 crc kubenswrapper[4955]: I0217 13:06:19.222271 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:06:19 crc kubenswrapper[4955]: I0217 13:06:19.222326 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:06:19 crc kubenswrapper[4955]: I0217 13:06:19.222397 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:06:19 crc kubenswrapper[4955]: E0217 13:06:19.222565 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:06:19 crc kubenswrapper[4955]: E0217 13:06:19.222642 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:06:19 crc kubenswrapper[4955]: E0217 13:06:19.222458 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:06:19 crc kubenswrapper[4955]: I0217 13:06:19.964071 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9hx9f_1e57bdf8-e537-402a-9815-609fecdd68ea/kube-multus/1.log" Feb 17 13:06:20 crc kubenswrapper[4955]: I0217 13:06:20.222513 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:06:20 crc kubenswrapper[4955]: E0217 13:06:20.222732 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:06:21 crc kubenswrapper[4955]: I0217 13:06:21.222463 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:06:21 crc kubenswrapper[4955]: I0217 13:06:21.222547 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:06:21 crc kubenswrapper[4955]: I0217 13:06:21.222463 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:06:21 crc kubenswrapper[4955]: E0217 13:06:21.222646 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:06:21 crc kubenswrapper[4955]: E0217 13:06:21.222728 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:06:21 crc kubenswrapper[4955]: E0217 13:06:21.222863 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:06:22 crc kubenswrapper[4955]: E0217 13:06:22.212450 4955 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Feb 17 13:06:22 crc kubenswrapper[4955]: I0217 13:06:22.222309 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:06:22 crc kubenswrapper[4955]: E0217 13:06:22.224403 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:06:22 crc kubenswrapper[4955]: E0217 13:06:22.355827 4955 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 17 13:06:23 crc kubenswrapper[4955]: I0217 13:06:23.221943 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:06:23 crc kubenswrapper[4955]: I0217 13:06:23.222024 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:06:23 crc kubenswrapper[4955]: I0217 13:06:23.222097 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:06:23 crc kubenswrapper[4955]: E0217 13:06:23.222264 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:06:23 crc kubenswrapper[4955]: E0217 13:06:23.222500 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:06:23 crc kubenswrapper[4955]: E0217 13:06:23.222408 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:06:24 crc kubenswrapper[4955]: I0217 13:06:24.222653 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:06:24 crc kubenswrapper[4955]: E0217 13:06:24.223550 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:06:25 crc kubenswrapper[4955]: I0217 13:06:25.222705 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:06:25 crc kubenswrapper[4955]: I0217 13:06:25.222698 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:06:25 crc kubenswrapper[4955]: I0217 13:06:25.222741 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:06:25 crc kubenswrapper[4955]: E0217 13:06:25.223043 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:06:25 crc kubenswrapper[4955]: E0217 13:06:25.223176 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:06:25 crc kubenswrapper[4955]: E0217 13:06:25.223640 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:06:26 crc kubenswrapper[4955]: I0217 13:06:26.222005 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:06:26 crc kubenswrapper[4955]: E0217 13:06:26.222637 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:06:26 crc kubenswrapper[4955]: I0217 13:06:26.222835 4955 scope.go:117] "RemoveContainer" containerID="528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6" Feb 17 13:06:26 crc kubenswrapper[4955]: I0217 13:06:26.992090 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dcwgg_dcbc7a69-58e7-4b7a-bb84-96276819ce88/ovnkube-controller/3.log" Feb 17 13:06:26 crc kubenswrapper[4955]: I0217 13:06:26.995853 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" event={"ID":"dcbc7a69-58e7-4b7a-bb84-96276819ce88","Type":"ContainerStarted","Data":"0bf8e3c7005480c233ef81b5a37bc7d925eadebc225ecd7d5345fa4e96383b90"} Feb 17 13:06:26 crc kubenswrapper[4955]: I0217 13:06:26.996404 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:06:27 crc kubenswrapper[4955]: I0217 13:06:27.077523 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" podStartSLOduration=104.077495268 podStartE2EDuration="1m44.077495268s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:27.033279107 +0000 UTC m=+125.556008670" watchObservedRunningTime="2026-02-17 13:06:27.077495268 +0000 UTC m=+125.600224811" Feb 17 13:06:27 crc kubenswrapper[4955]: I0217 13:06:27.078751 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-2k5pz"] Feb 17 13:06:27 crc kubenswrapper[4955]: I0217 13:06:27.078879 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:06:27 crc kubenswrapper[4955]: E0217 13:06:27.078993 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:06:27 crc kubenswrapper[4955]: I0217 13:06:27.222259 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:06:27 crc kubenswrapper[4955]: I0217 13:06:27.222333 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:06:27 crc kubenswrapper[4955]: E0217 13:06:27.222417 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:06:27 crc kubenswrapper[4955]: E0217 13:06:27.222513 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:06:27 crc kubenswrapper[4955]: I0217 13:06:27.222333 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:06:27 crc kubenswrapper[4955]: E0217 13:06:27.222640 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:06:27 crc kubenswrapper[4955]: E0217 13:06:27.357655 4955 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 17 13:06:29 crc kubenswrapper[4955]: I0217 13:06:29.222388 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:06:29 crc kubenswrapper[4955]: E0217 13:06:29.222871 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:06:29 crc kubenswrapper[4955]: I0217 13:06:29.223313 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:06:29 crc kubenswrapper[4955]: E0217 13:06:29.223424 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:06:29 crc kubenswrapper[4955]: I0217 13:06:29.223657 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:06:29 crc kubenswrapper[4955]: E0217 13:06:29.223772 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:06:29 crc kubenswrapper[4955]: I0217 13:06:29.224106 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:06:29 crc kubenswrapper[4955]: E0217 13:06:29.224265 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:06:31 crc kubenswrapper[4955]: I0217 13:06:31.222584 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:06:31 crc kubenswrapper[4955]: E0217 13:06:31.222726 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:06:31 crc kubenswrapper[4955]: I0217 13:06:31.223508 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:06:31 crc kubenswrapper[4955]: E0217 13:06:31.223568 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:06:31 crc kubenswrapper[4955]: I0217 13:06:31.223684 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:06:31 crc kubenswrapper[4955]: E0217 13:06:31.223737 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:06:31 crc kubenswrapper[4955]: I0217 13:06:31.223865 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:06:31 crc kubenswrapper[4955]: E0217 13:06:31.223944 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:06:31 crc kubenswrapper[4955]: I0217 13:06:31.224240 4955 scope.go:117] "RemoveContainer" containerID="c81948fc2806ea8718909fd269c029efc6a26675e45359e971272354eccfa74d" Feb 17 13:06:32 crc kubenswrapper[4955]: I0217 13:06:32.017511 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9hx9f_1e57bdf8-e537-402a-9815-609fecdd68ea/kube-multus/1.log" Feb 17 13:06:32 crc kubenswrapper[4955]: I0217 13:06:32.017615 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-9hx9f" event={"ID":"1e57bdf8-e537-402a-9815-609fecdd68ea","Type":"ContainerStarted","Data":"7db7f26cfeecd2fd564d165a58ac1b15757ebb5167a56c82a8fc8def9bdca610"} Feb 17 13:06:32 crc kubenswrapper[4955]: E0217 13:06:32.358247 4955 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 17 13:06:33 crc kubenswrapper[4955]: I0217 13:06:33.223034 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:06:33 crc kubenswrapper[4955]: I0217 13:06:33.223060 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:06:33 crc kubenswrapper[4955]: I0217 13:06:33.223237 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:06:33 crc kubenswrapper[4955]: I0217 13:06:33.223297 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:06:33 crc kubenswrapper[4955]: E0217 13:06:33.223555 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:06:33 crc kubenswrapper[4955]: E0217 13:06:33.223945 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:06:33 crc kubenswrapper[4955]: E0217 13:06:33.224064 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:06:33 crc kubenswrapper[4955]: E0217 13:06:33.224579 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:06:35 crc kubenswrapper[4955]: I0217 13:06:35.221967 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:06:35 crc kubenswrapper[4955]: E0217 13:06:35.222155 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:06:35 crc kubenswrapper[4955]: I0217 13:06:35.222208 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:06:35 crc kubenswrapper[4955]: I0217 13:06:35.222263 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:06:35 crc kubenswrapper[4955]: I0217 13:06:35.222208 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:06:35 crc kubenswrapper[4955]: E0217 13:06:35.222391 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:06:35 crc kubenswrapper[4955]: E0217 13:06:35.222592 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:06:35 crc kubenswrapper[4955]: E0217 13:06:35.222909 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:06:37 crc kubenswrapper[4955]: I0217 13:06:37.222747 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:06:37 crc kubenswrapper[4955]: I0217 13:06:37.222865 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:06:37 crc kubenswrapper[4955]: E0217 13:06:37.222930 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 17 13:06:37 crc kubenswrapper[4955]: I0217 13:06:37.222949 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:06:37 crc kubenswrapper[4955]: I0217 13:06:37.222873 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:06:37 crc kubenswrapper[4955]: E0217 13:06:37.223067 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2k5pz" podUID="77480847-3271-4161-a833-e62b3bb4413e" Feb 17 13:06:37 crc kubenswrapper[4955]: E0217 13:06:37.223126 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 17 13:06:37 crc kubenswrapper[4955]: E0217 13:06:37.223298 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 17 13:06:39 crc kubenswrapper[4955]: I0217 13:06:39.222249 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:06:39 crc kubenswrapper[4955]: I0217 13:06:39.222300 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:06:39 crc kubenswrapper[4955]: I0217 13:06:39.222333 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:06:39 crc kubenswrapper[4955]: I0217 13:06:39.222356 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:06:39 crc kubenswrapper[4955]: I0217 13:06:39.226061 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 17 13:06:39 crc kubenswrapper[4955]: I0217 13:06:39.226145 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 17 13:06:39 crc kubenswrapper[4955]: I0217 13:06:39.226722 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 17 13:06:39 crc kubenswrapper[4955]: I0217 13:06:39.226890 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 17 13:06:39 crc kubenswrapper[4955]: I0217 13:06:39.228451 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 17 13:06:39 crc kubenswrapper[4955]: I0217 13:06:39.230760 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.071868 4955 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.108824 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-prtk8"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.109516 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-prtk8" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.115845 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.116198 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.116397 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.117087 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.117346 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.121474 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-d5sv6"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.122324 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.122980 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-d5sv6" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.123105 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p72m9"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.124153 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p72m9" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.127639 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.127649 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.128057 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.128249 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.128426 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.128430 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.128719 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-982dt"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.129335 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-982dt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.133293 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-trrhn"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.133690 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-5x7zc"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.134133 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5x7zc" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.134576 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-trrhn" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.136040 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.136829 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.139496 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.139571 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.139510 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.141932 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-8dx57"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.143245 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-fpwvv"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.143989 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-fpwvv" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.144631 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-8dx57" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.146973 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-mvfqt"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.147740 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mvfqt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.156946 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.157267 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.158504 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.158734 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.158773 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.160130 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-npb7n"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.160478 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.160726 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-qz68q"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.160777 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.161286 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-rxsb4"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.161299 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.161332 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.161487 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.161507 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.161621 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.161638 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.161744 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.161754 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-rxsb4" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.161775 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.161858 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.161948 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.162233 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.162451 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-npb7n" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.162592 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-qz68q" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.164305 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.164434 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.164549 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.165201 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vjrvb"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.165763 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vjrvb" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.166144 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-nskw2"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.166174 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.166198 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.171088 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.171624 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.185677 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-nskw2" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.189159 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.189298 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.189313 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.189263 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.189514 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.189537 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.189620 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.189204 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.189631 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.189992 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.190260 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.190853 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.190974 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.191171 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.194045 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.194285 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.194969 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8mkhz"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.195632 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.201232 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t5v8v"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.201973 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t5v8v" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.207281 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.207553 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.208052 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.208245 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.208648 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.208923 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.212970 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-vfbvk"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.221902 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qbxwp"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.222361 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.222502 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qbxwp" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.210079 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.222727 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.210209 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.210262 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.210333 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.210367 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.210417 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.213619 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.213747 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.213857 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.215864 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.215936 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.215976 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.216018 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.216055 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.216093 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.216133 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.216248 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.216298 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.217596 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.217658 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.217699 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.217876 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.219414 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.220451 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.222570 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.222613 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.222626 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.228222 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.232053 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqclt\" (UniqueName: \"kubernetes.io/projected/01aa5d18-e942-4a3d-8af5-1e41b503f85a-kube-api-access-vqclt\") pod \"downloads-7954f5f757-qz68q\" (UID: \"01aa5d18-e942-4a3d-8af5-1e41b503f85a\") " pod="openshift-console/downloads-7954f5f757-qz68q" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.232090 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32ea5b76-e8e5-4906-bc55-99d5ec780ff2-serving-cert\") pod \"route-controller-manager-6576b87f9c-trrhn\" (UID: \"32ea5b76-e8e5-4906-bc55-99d5ec780ff2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-trrhn" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.232127 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7db0eeb4-4ffe-4375-a854-0f47efce6a26-config\") pod \"controller-manager-879f6c89f-prtk8\" (UID: \"7db0eeb4-4ffe-4375-a854-0f47efce6a26\") " pod="openshift-controller-manager/controller-manager-879f6c89f-prtk8" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.232152 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/792abe2b-bcce-4af6-828b-7a6726b87976-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-p72m9\" (UID: \"792abe2b-bcce-4af6-828b-7a6726b87976\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p72m9" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.232174 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/922b208d-fc86-4243-86f1-eacc2797118e-etcd-client\") pod \"apiserver-76f77b778f-8dx57\" (UID: \"922b208d-fc86-4243-86f1-eacc2797118e\") " pod="openshift-apiserver/apiserver-76f77b778f-8dx57" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.232193 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26690efc-79a1-4bc4-bf0a-78b46fb95223-config\") pod \"machine-approver-56656f9798-mvfqt\" (UID: \"26690efc-79a1-4bc4-bf0a-78b46fb95223\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mvfqt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.232218 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6b4e02e4-f80e-44a5-a2a7-f75feaf2c1ce-serving-cert\") pod \"authentication-operator-69f744f599-fpwvv\" (UID: \"6b4e02e4-f80e-44a5-a2a7-f75feaf2c1ce\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fpwvv" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.232241 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/792abe2b-bcce-4af6-828b-7a6726b87976-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-p72m9\" (UID: \"792abe2b-bcce-4af6-828b-7a6726b87976\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p72m9" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.232268 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmtdf\" (UniqueName: \"kubernetes.io/projected/7db0eeb4-4ffe-4375-a854-0f47efce6a26-kube-api-access-cmtdf\") pod \"controller-manager-879f6c89f-prtk8\" (UID: \"7db0eeb4-4ffe-4375-a854-0f47efce6a26\") " pod="openshift-controller-manager/controller-manager-879f6c89f-prtk8" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.232290 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/58bab2f1-e131-46cb-b705-6bee97eb8452-audit-dir\") pod \"apiserver-7bbb656c7d-5x7zc\" (UID: \"58bab2f1-e131-46cb-b705-6bee97eb8452\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5x7zc" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.232320 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92cx5\" (UniqueName: \"kubernetes.io/projected/792abe2b-bcce-4af6-828b-7a6726b87976-kube-api-access-92cx5\") pod \"cluster-image-registry-operator-dc59b4c8b-p72m9\" (UID: \"792abe2b-bcce-4af6-828b-7a6726b87976\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p72m9" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.232340 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/26690efc-79a1-4bc4-bf0a-78b46fb95223-machine-approver-tls\") pod \"machine-approver-56656f9798-mvfqt\" (UID: \"26690efc-79a1-4bc4-bf0a-78b46fb95223\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mvfqt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.232361 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/58bab2f1-e131-46cb-b705-6bee97eb8452-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-5x7zc\" (UID: \"58bab2f1-e131-46cb-b705-6bee97eb8452\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5x7zc" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.232381 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6b4e02e4-f80e-44a5-a2a7-f75feaf2c1ce-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-fpwvv\" (UID: \"6b4e02e4-f80e-44a5-a2a7-f75feaf2c1ce\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fpwvv" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.232400 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rf2mz\" (UniqueName: \"kubernetes.io/projected/9d7d0f56-00a3-4eae-8ff2-a35e1f20a377-kube-api-access-rf2mz\") pod \"openshift-config-operator-7777fb866f-nskw2\" (UID: \"9d7d0f56-00a3-4eae-8ff2-a35e1f20a377\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-nskw2" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.232421 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/de868b10-12d3-4ec1-87d6-f4fcb4994aa3-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-982dt\" (UID: \"de868b10-12d3-4ec1-87d6-f4fcb4994aa3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-982dt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.232440 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nczq7\" (UniqueName: \"kubernetes.io/projected/58bab2f1-e131-46cb-b705-6bee97eb8452-kube-api-access-nczq7\") pod \"apiserver-7bbb656c7d-5x7zc\" (UID: \"58bab2f1-e131-46cb-b705-6bee97eb8452\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5x7zc" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.232471 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4p86\" (UniqueName: \"kubernetes.io/projected/93264a11-973b-435d-a51f-999be787ed0a-kube-api-access-d4p86\") pod \"openshift-controller-manager-operator-756b6f6bc6-vjrvb\" (UID: \"93264a11-973b-435d-a51f-999be787ed0a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vjrvb" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.232493 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgznr\" (UniqueName: \"kubernetes.io/projected/6b4e02e4-f80e-44a5-a2a7-f75feaf2c1ce-kube-api-access-fgznr\") pod \"authentication-operator-69f744f599-fpwvv\" (UID: \"6b4e02e4-f80e-44a5-a2a7-f75feaf2c1ce\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fpwvv" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.232516 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d7d0f56-00a3-4eae-8ff2-a35e1f20a377-serving-cert\") pod \"openshift-config-operator-7777fb866f-nskw2\" (UID: \"9d7d0f56-00a3-4eae-8ff2-a35e1f20a377\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-nskw2" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.232538 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93264a11-973b-435d-a51f-999be787ed0a-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-vjrvb\" (UID: \"93264a11-973b-435d-a51f-999be787ed0a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vjrvb" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.232558 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/847ef69d-8cd1-4831-8250-05f2cf789d88-config\") pod \"machine-api-operator-5694c8668f-d5sv6\" (UID: \"847ef69d-8cd1-4831-8250-05f2cf789d88\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-d5sv6" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.232577 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6b4e02e4-f80e-44a5-a2a7-f75feaf2c1ce-service-ca-bundle\") pod \"authentication-operator-69f744f599-fpwvv\" (UID: \"6b4e02e4-f80e-44a5-a2a7-f75feaf2c1ce\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fpwvv" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.232694 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/922b208d-fc86-4243-86f1-eacc2797118e-encryption-config\") pod \"apiserver-76f77b778f-8dx57\" (UID: \"922b208d-fc86-4243-86f1-eacc2797118e\") " pod="openshift-apiserver/apiserver-76f77b778f-8dx57" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.232717 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/32ea5b76-e8e5-4906-bc55-99d5ec780ff2-client-ca\") pod \"route-controller-manager-6576b87f9c-trrhn\" (UID: \"32ea5b76-e8e5-4906-bc55-99d5ec780ff2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-trrhn" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.232763 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/44192b09-31c7-4e49-b472-5ff0a67d30f5-trusted-ca\") pod \"console-operator-58897d9998-rxsb4\" (UID: \"44192b09-31c7-4e49-b472-5ff0a67d30f5\") " pod="openshift-console-operator/console-operator-58897d9998-rxsb4" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.232812 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/58bab2f1-e131-46cb-b705-6bee97eb8452-serving-cert\") pod \"apiserver-7bbb656c7d-5x7zc\" (UID: \"58bab2f1-e131-46cb-b705-6bee97eb8452\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5x7zc" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.232857 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e690bd42-0ab1-4b3a-802b-4511117b6fe0-service-ca\") pod \"console-f9d7485db-npb7n\" (UID: \"e690bd42-0ab1-4b3a-802b-4511117b6fe0\") " pod="openshift-console/console-f9d7485db-npb7n" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.232904 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/44192b09-31c7-4e49-b472-5ff0a67d30f5-serving-cert\") pod \"console-operator-58897d9998-rxsb4\" (UID: \"44192b09-31c7-4e49-b472-5ff0a67d30f5\") " pod="openshift-console-operator/console-operator-58897d9998-rxsb4" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.232929 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b4e02e4-f80e-44a5-a2a7-f75feaf2c1ce-config\") pod \"authentication-operator-69f744f599-fpwvv\" (UID: \"6b4e02e4-f80e-44a5-a2a7-f75feaf2c1ce\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fpwvv" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.232954 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/58bab2f1-e131-46cb-b705-6bee97eb8452-audit-policies\") pod \"apiserver-7bbb656c7d-5x7zc\" (UID: \"58bab2f1-e131-46cb-b705-6bee97eb8452\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5x7zc" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.233003 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbbcg\" (UniqueName: \"kubernetes.io/projected/de868b10-12d3-4ec1-87d6-f4fcb4994aa3-kube-api-access-cbbcg\") pod \"openshift-apiserver-operator-796bbdcf4f-982dt\" (UID: \"de868b10-12d3-4ec1-87d6-f4fcb4994aa3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-982dt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.233027 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xns4k\" (UniqueName: \"kubernetes.io/projected/26690efc-79a1-4bc4-bf0a-78b46fb95223-kube-api-access-xns4k\") pod \"machine-approver-56656f9798-mvfqt\" (UID: \"26690efc-79a1-4bc4-bf0a-78b46fb95223\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mvfqt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.233088 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7db0eeb4-4ffe-4375-a854-0f47efce6a26-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-prtk8\" (UID: \"7db0eeb4-4ffe-4375-a854-0f47efce6a26\") " pod="openshift-controller-manager/controller-manager-879f6c89f-prtk8" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.233134 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bb8c5\" (UniqueName: \"kubernetes.io/projected/847ef69d-8cd1-4831-8250-05f2cf789d88-kube-api-access-bb8c5\") pod \"machine-api-operator-5694c8668f-d5sv6\" (UID: \"847ef69d-8cd1-4831-8250-05f2cf789d88\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-d5sv6" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.233173 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/922b208d-fc86-4243-86f1-eacc2797118e-node-pullsecrets\") pod \"apiserver-76f77b778f-8dx57\" (UID: \"922b208d-fc86-4243-86f1-eacc2797118e\") " pod="openshift-apiserver/apiserver-76f77b778f-8dx57" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.233217 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/792abe2b-bcce-4af6-828b-7a6726b87976-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-p72m9\" (UID: \"792abe2b-bcce-4af6-828b-7a6726b87976\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p72m9" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.233249 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/922b208d-fc86-4243-86f1-eacc2797118e-config\") pod \"apiserver-76f77b778f-8dx57\" (UID: \"922b208d-fc86-4243-86f1-eacc2797118e\") " pod="openshift-apiserver/apiserver-76f77b778f-8dx57" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.233280 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44192b09-31c7-4e49-b472-5ff0a67d30f5-config\") pod \"console-operator-58897d9998-rxsb4\" (UID: \"44192b09-31c7-4e49-b472-5ff0a67d30f5\") " pod="openshift-console-operator/console-operator-58897d9998-rxsb4" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.233310 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dfbc\" (UniqueName: \"kubernetes.io/projected/44192b09-31c7-4e49-b472-5ff0a67d30f5-kube-api-access-2dfbc\") pod \"console-operator-58897d9998-rxsb4\" (UID: \"44192b09-31c7-4e49-b472-5ff0a67d30f5\") " pod="openshift-console-operator/console-operator-58897d9998-rxsb4" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.233330 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/26690efc-79a1-4bc4-bf0a-78b46fb95223-auth-proxy-config\") pod \"machine-approver-56656f9798-mvfqt\" (UID: \"26690efc-79a1-4bc4-bf0a-78b46fb95223\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mvfqt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.233380 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/847ef69d-8cd1-4831-8250-05f2cf789d88-images\") pod \"machine-api-operator-5694c8668f-d5sv6\" (UID: \"847ef69d-8cd1-4831-8250-05f2cf789d88\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-d5sv6" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.233407 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e690bd42-0ab1-4b3a-802b-4511117b6fe0-trusted-ca-bundle\") pod \"console-f9d7485db-npb7n\" (UID: \"e690bd42-0ab1-4b3a-802b-4511117b6fe0\") " pod="openshift-console/console-f9d7485db-npb7n" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.233445 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/93264a11-973b-435d-a51f-999be787ed0a-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-vjrvb\" (UID: \"93264a11-973b-435d-a51f-999be787ed0a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vjrvb" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.233470 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/58bab2f1-e131-46cb-b705-6bee97eb8452-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-5x7zc\" (UID: \"58bab2f1-e131-46cb-b705-6bee97eb8452\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5x7zc" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.233516 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e690bd42-0ab1-4b3a-802b-4511117b6fe0-console-oauth-config\") pod \"console-f9d7485db-npb7n\" (UID: \"e690bd42-0ab1-4b3a-802b-4511117b6fe0\") " pod="openshift-console/console-f9d7485db-npb7n" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.233540 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e690bd42-0ab1-4b3a-802b-4511117b6fe0-oauth-serving-cert\") pod \"console-f9d7485db-npb7n\" (UID: \"e690bd42-0ab1-4b3a-802b-4511117b6fe0\") " pod="openshift-console/console-f9d7485db-npb7n" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.233561 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qn6qd\" (UniqueName: \"kubernetes.io/projected/32ea5b76-e8e5-4906-bc55-99d5ec780ff2-kube-api-access-qn6qd\") pod \"route-controller-manager-6576b87f9c-trrhn\" (UID: \"32ea5b76-e8e5-4906-bc55-99d5ec780ff2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-trrhn" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.233593 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/847ef69d-8cd1-4831-8250-05f2cf789d88-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-d5sv6\" (UID: \"847ef69d-8cd1-4831-8250-05f2cf789d88\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-d5sv6" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.233620 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/922b208d-fc86-4243-86f1-eacc2797118e-audit\") pod \"apiserver-76f77b778f-8dx57\" (UID: \"922b208d-fc86-4243-86f1-eacc2797118e\") " pod="openshift-apiserver/apiserver-76f77b778f-8dx57" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.233643 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e690bd42-0ab1-4b3a-802b-4511117b6fe0-console-config\") pod \"console-f9d7485db-npb7n\" (UID: \"e690bd42-0ab1-4b3a-802b-4511117b6fe0\") " pod="openshift-console/console-f9d7485db-npb7n" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.233670 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/58bab2f1-e131-46cb-b705-6bee97eb8452-encryption-config\") pod \"apiserver-7bbb656c7d-5x7zc\" (UID: \"58bab2f1-e131-46cb-b705-6bee97eb8452\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5x7zc" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.233700 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/922b208d-fc86-4243-86f1-eacc2797118e-image-import-ca\") pod \"apiserver-76f77b778f-8dx57\" (UID: \"922b208d-fc86-4243-86f1-eacc2797118e\") " pod="openshift-apiserver/apiserver-76f77b778f-8dx57" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.233806 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7db0eeb4-4ffe-4375-a854-0f47efce6a26-client-ca\") pod \"controller-manager-879f6c89f-prtk8\" (UID: \"7db0eeb4-4ffe-4375-a854-0f47efce6a26\") " pod="openshift-controller-manager/controller-manager-879f6c89f-prtk8" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.233840 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de868b10-12d3-4ec1-87d6-f4fcb4994aa3-config\") pod \"openshift-apiserver-operator-796bbdcf4f-982dt\" (UID: \"de868b10-12d3-4ec1-87d6-f4fcb4994aa3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-982dt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.233876 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/9d7d0f56-00a3-4eae-8ff2-a35e1f20a377-available-featuregates\") pod \"openshift-config-operator-7777fb866f-nskw2\" (UID: \"9d7d0f56-00a3-4eae-8ff2-a35e1f20a377\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-nskw2" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.233991 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/58bab2f1-e131-46cb-b705-6bee97eb8452-etcd-client\") pod \"apiserver-7bbb656c7d-5x7zc\" (UID: \"58bab2f1-e131-46cb-b705-6bee97eb8452\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5x7zc" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.234026 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7db0eeb4-4ffe-4375-a854-0f47efce6a26-serving-cert\") pod \"controller-manager-879f6c89f-prtk8\" (UID: \"7db0eeb4-4ffe-4375-a854-0f47efce6a26\") " pod="openshift-controller-manager/controller-manager-879f6c89f-prtk8" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.234050 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/922b208d-fc86-4243-86f1-eacc2797118e-serving-cert\") pod \"apiserver-76f77b778f-8dx57\" (UID: \"922b208d-fc86-4243-86f1-eacc2797118e\") " pod="openshift-apiserver/apiserver-76f77b778f-8dx57" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.234083 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9bjs\" (UniqueName: \"kubernetes.io/projected/e690bd42-0ab1-4b3a-802b-4511117b6fe0-kube-api-access-b9bjs\") pod \"console-f9d7485db-npb7n\" (UID: \"e690bd42-0ab1-4b3a-802b-4511117b6fe0\") " pod="openshift-console/console-f9d7485db-npb7n" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.234108 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32ea5b76-e8e5-4906-bc55-99d5ec780ff2-config\") pod \"route-controller-manager-6576b87f9c-trrhn\" (UID: \"32ea5b76-e8e5-4906-bc55-99d5ec780ff2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-trrhn" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.234155 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/922b208d-fc86-4243-86f1-eacc2797118e-etcd-serving-ca\") pod \"apiserver-76f77b778f-8dx57\" (UID: \"922b208d-fc86-4243-86f1-eacc2797118e\") " pod="openshift-apiserver/apiserver-76f77b778f-8dx57" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.234189 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/922b208d-fc86-4243-86f1-eacc2797118e-audit-dir\") pod \"apiserver-76f77b778f-8dx57\" (UID: \"922b208d-fc86-4243-86f1-eacc2797118e\") " pod="openshift-apiserver/apiserver-76f77b778f-8dx57" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.234215 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-st8cn\" (UniqueName: \"kubernetes.io/projected/922b208d-fc86-4243-86f1-eacc2797118e-kube-api-access-st8cn\") pod \"apiserver-76f77b778f-8dx57\" (UID: \"922b208d-fc86-4243-86f1-eacc2797118e\") " pod="openshift-apiserver/apiserver-76f77b778f-8dx57" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.234236 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/922b208d-fc86-4243-86f1-eacc2797118e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-8dx57\" (UID: \"922b208d-fc86-4243-86f1-eacc2797118e\") " pod="openshift-apiserver/apiserver-76f77b778f-8dx57" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.234266 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e690bd42-0ab1-4b3a-802b-4511117b6fe0-console-serving-cert\") pod \"console-f9d7485db-npb7n\" (UID: \"e690bd42-0ab1-4b3a-802b-4511117b6fe0\") " pod="openshift-console/console-f9d7485db-npb7n" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.238043 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.246141 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.246349 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.246531 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.246888 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.247054 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.247344 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.251551 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.255116 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.256377 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.266966 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-prtk8"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.267035 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8t8vm"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.274359 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-qdg4k"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.274855 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-wdpjf"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.275184 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8t8vm" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.275231 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-vnwxg"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.275369 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-wdpjf" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.275447 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qdg4k" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.276493 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cb48c"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.276809 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8l455"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.277093 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4hhxl"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.277361 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-4xvzn"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.277404 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-vnwxg" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.277516 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.277827 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8l455" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.277881 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cb48c" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.277911 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4hhxl" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.278717 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vtd5z"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.279353 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4xvzn" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.279433 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-sv4jj"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.283942 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vtd5z" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.285162 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5264p"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.285530 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sv4jj" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.286031 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5264p" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.287672 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.294705 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-m7gfn"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.295537 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-m7gfn" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.296173 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f7ccb"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.296735 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f7ccb" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.306279 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.311977 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-zc8xw"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.312771 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-zc8xw" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.313186 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-9tpfm"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.313957 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tpfm" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.314154 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-brmrn"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.314939 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-brmrn" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.322947 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29522220-v5twm"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.323711 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29522220-v5twm" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.323962 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9c9zp"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.324731 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-9c9zp" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.325749 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-6hmm5"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.336631 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/922b208d-fc86-4243-86f1-eacc2797118e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-8dx57\" (UID: \"922b208d-fc86-4243-86f1-eacc2797118e\") " pod="openshift-apiserver/apiserver-76f77b778f-8dx57" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.336683 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e690bd42-0ab1-4b3a-802b-4511117b6fe0-console-serving-cert\") pod \"console-f9d7485db-npb7n\" (UID: \"e690bd42-0ab1-4b3a-802b-4511117b6fe0\") " pod="openshift-console/console-f9d7485db-npb7n" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.336715 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3955ee14-c2d4-434e-a9b5-baf95abcbd02-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-5264p\" (UID: \"3955ee14-c2d4-434e-a9b5-baf95abcbd02\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5264p" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.336745 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/792abe2b-bcce-4af6-828b-7a6726b87976-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-p72m9\" (UID: \"792abe2b-bcce-4af6-828b-7a6726b87976\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p72m9" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.336774 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqclt\" (UniqueName: \"kubernetes.io/projected/01aa5d18-e942-4a3d-8af5-1e41b503f85a-kube-api-access-vqclt\") pod \"downloads-7954f5f757-qz68q\" (UID: \"01aa5d18-e942-4a3d-8af5-1e41b503f85a\") " pod="openshift-console/downloads-7954f5f757-qz68q" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.336820 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32ea5b76-e8e5-4906-bc55-99d5ec780ff2-serving-cert\") pod \"route-controller-manager-6576b87f9c-trrhn\" (UID: \"32ea5b76-e8e5-4906-bc55-99d5ec780ff2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-trrhn" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.336856 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7db0eeb4-4ffe-4375-a854-0f47efce6a26-config\") pod \"controller-manager-879f6c89f-prtk8\" (UID: \"7db0eeb4-4ffe-4375-a854-0f47efce6a26\") " pod="openshift-controller-manager/controller-manager-879f6c89f-prtk8" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.336880 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/792abe2b-bcce-4af6-828b-7a6726b87976-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-p72m9\" (UID: \"792abe2b-bcce-4af6-828b-7a6726b87976\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p72m9" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.336902 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/922b208d-fc86-4243-86f1-eacc2797118e-etcd-client\") pod \"apiserver-76f77b778f-8dx57\" (UID: \"922b208d-fc86-4243-86f1-eacc2797118e\") " pod="openshift-apiserver/apiserver-76f77b778f-8dx57" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.336923 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26690efc-79a1-4bc4-bf0a-78b46fb95223-config\") pod \"machine-approver-56656f9798-mvfqt\" (UID: \"26690efc-79a1-4bc4-bf0a-78b46fb95223\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mvfqt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.336948 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6b4e02e4-f80e-44a5-a2a7-f75feaf2c1ce-serving-cert\") pod \"authentication-operator-69f744f599-fpwvv\" (UID: \"6b4e02e4-f80e-44a5-a2a7-f75feaf2c1ce\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fpwvv" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.336969 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/58bab2f1-e131-46cb-b705-6bee97eb8452-audit-dir\") pod \"apiserver-7bbb656c7d-5x7zc\" (UID: \"58bab2f1-e131-46cb-b705-6bee97eb8452\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5x7zc" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.336990 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmtdf\" (UniqueName: \"kubernetes.io/projected/7db0eeb4-4ffe-4375-a854-0f47efce6a26-kube-api-access-cmtdf\") pod \"controller-manager-879f6c89f-prtk8\" (UID: \"7db0eeb4-4ffe-4375-a854-0f47efce6a26\") " pod="openshift-controller-manager/controller-manager-879f6c89f-prtk8" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.337025 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92cx5\" (UniqueName: \"kubernetes.io/projected/792abe2b-bcce-4af6-828b-7a6726b87976-kube-api-access-92cx5\") pod \"cluster-image-registry-operator-dc59b4c8b-p72m9\" (UID: \"792abe2b-bcce-4af6-828b-7a6726b87976\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p72m9" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.337048 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/26690efc-79a1-4bc4-bf0a-78b46fb95223-machine-approver-tls\") pod \"machine-approver-56656f9798-mvfqt\" (UID: \"26690efc-79a1-4bc4-bf0a-78b46fb95223\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mvfqt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.337075 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g28kf\" (UniqueName: \"kubernetes.io/projected/ff950861-7bca-4bfa-9857-cd906f67d5cb-kube-api-access-g28kf\") pod \"migrator-59844c95c7-4xvzn\" (UID: \"ff950861-7bca-4bfa-9857-cd906f67d5cb\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4xvzn" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.337098 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/58bab2f1-e131-46cb-b705-6bee97eb8452-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-5x7zc\" (UID: \"58bab2f1-e131-46cb-b705-6bee97eb8452\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5x7zc" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.337119 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nczq7\" (UniqueName: \"kubernetes.io/projected/58bab2f1-e131-46cb-b705-6bee97eb8452-kube-api-access-nczq7\") pod \"apiserver-7bbb656c7d-5x7zc\" (UID: \"58bab2f1-e131-46cb-b705-6bee97eb8452\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5x7zc" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.337142 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b3dbb323-561b-4d26-886c-bef115c0dca0-metrics-tls\") pod \"ingress-operator-5b745b69d9-qdg4k\" (UID: \"b3dbb323-561b-4d26-886c-bef115c0dca0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qdg4k" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.337166 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6b4e02e4-f80e-44a5-a2a7-f75feaf2c1ce-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-fpwvv\" (UID: \"6b4e02e4-f80e-44a5-a2a7-f75feaf2c1ce\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fpwvv" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.337189 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rf2mz\" (UniqueName: \"kubernetes.io/projected/9d7d0f56-00a3-4eae-8ff2-a35e1f20a377-kube-api-access-rf2mz\") pod \"openshift-config-operator-7777fb866f-nskw2\" (UID: \"9d7d0f56-00a3-4eae-8ff2-a35e1f20a377\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-nskw2" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.337213 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/de868b10-12d3-4ec1-87d6-f4fcb4994aa3-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-982dt\" (UID: \"de868b10-12d3-4ec1-87d6-f4fcb4994aa3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-982dt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.337246 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4p86\" (UniqueName: \"kubernetes.io/projected/93264a11-973b-435d-a51f-999be787ed0a-kube-api-access-d4p86\") pod \"openshift-controller-manager-operator-756b6f6bc6-vjrvb\" (UID: \"93264a11-973b-435d-a51f-999be787ed0a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vjrvb" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.337268 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgznr\" (UniqueName: \"kubernetes.io/projected/6b4e02e4-f80e-44a5-a2a7-f75feaf2c1ce-kube-api-access-fgznr\") pod \"authentication-operator-69f744f599-fpwvv\" (UID: \"6b4e02e4-f80e-44a5-a2a7-f75feaf2c1ce\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fpwvv" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.337306 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d7d0f56-00a3-4eae-8ff2-a35e1f20a377-serving-cert\") pod \"openshift-config-operator-7777fb866f-nskw2\" (UID: \"9d7d0f56-00a3-4eae-8ff2-a35e1f20a377\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-nskw2" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.337328 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93264a11-973b-435d-a51f-999be787ed0a-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-vjrvb\" (UID: \"93264a11-973b-435d-a51f-999be787ed0a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vjrvb" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.337347 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/847ef69d-8cd1-4831-8250-05f2cf789d88-config\") pod \"machine-api-operator-5694c8668f-d5sv6\" (UID: \"847ef69d-8cd1-4831-8250-05f2cf789d88\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-d5sv6" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.337366 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/32ea5b76-e8e5-4906-bc55-99d5ec780ff2-client-ca\") pod \"route-controller-manager-6576b87f9c-trrhn\" (UID: \"32ea5b76-e8e5-4906-bc55-99d5ec780ff2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-trrhn" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.337435 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b3dbb323-561b-4d26-886c-bef115c0dca0-trusted-ca\") pod \"ingress-operator-5b745b69d9-qdg4k\" (UID: \"b3dbb323-561b-4d26-886c-bef115c0dca0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qdg4k" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.337467 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b3dbb323-561b-4d26-886c-bef115c0dca0-bound-sa-token\") pod \"ingress-operator-5b745b69d9-qdg4k\" (UID: \"b3dbb323-561b-4d26-886c-bef115c0dca0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qdg4k" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.337491 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6b4e02e4-f80e-44a5-a2a7-f75feaf2c1ce-service-ca-bundle\") pod \"authentication-operator-69f744f599-fpwvv\" (UID: \"6b4e02e4-f80e-44a5-a2a7-f75feaf2c1ce\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fpwvv" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.337509 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/922b208d-fc86-4243-86f1-eacc2797118e-encryption-config\") pod \"apiserver-76f77b778f-8dx57\" (UID: \"922b208d-fc86-4243-86f1-eacc2797118e\") " pod="openshift-apiserver/apiserver-76f77b778f-8dx57" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.337529 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/f84f960e-7741-4d13-bd6f-ec2072cd4d14-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-t5v8v\" (UID: \"f84f960e-7741-4d13-bd6f-ec2072cd4d14\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t5v8v" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.337557 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/44192b09-31c7-4e49-b472-5ff0a67d30f5-trusted-ca\") pod \"console-operator-58897d9998-rxsb4\" (UID: \"44192b09-31c7-4e49-b472-5ff0a67d30f5\") " pod="openshift-console-operator/console-operator-58897d9998-rxsb4" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.337580 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/58bab2f1-e131-46cb-b705-6bee97eb8452-serving-cert\") pod \"apiserver-7bbb656c7d-5x7zc\" (UID: \"58bab2f1-e131-46cb-b705-6bee97eb8452\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5x7zc" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.337606 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e690bd42-0ab1-4b3a-802b-4511117b6fe0-service-ca\") pod \"console-f9d7485db-npb7n\" (UID: \"e690bd42-0ab1-4b3a-802b-4511117b6fe0\") " pod="openshift-console/console-f9d7485db-npb7n" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.337631 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/44192b09-31c7-4e49-b472-5ff0a67d30f5-serving-cert\") pod \"console-operator-58897d9998-rxsb4\" (UID: \"44192b09-31c7-4e49-b472-5ff0a67d30f5\") " pod="openshift-console-operator/console-operator-58897d9998-rxsb4" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.337649 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b4e02e4-f80e-44a5-a2a7-f75feaf2c1ce-config\") pod \"authentication-operator-69f744f599-fpwvv\" (UID: \"6b4e02e4-f80e-44a5-a2a7-f75feaf2c1ce\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fpwvv" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.337670 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thvn9\" (UniqueName: \"kubernetes.io/projected/b3dbb323-561b-4d26-886c-bef115c0dca0-kube-api-access-thvn9\") pod \"ingress-operator-5b745b69d9-qdg4k\" (UID: \"b3dbb323-561b-4d26-886c-bef115c0dca0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qdg4k" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.337690 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbbcg\" (UniqueName: \"kubernetes.io/projected/de868b10-12d3-4ec1-87d6-f4fcb4994aa3-kube-api-access-cbbcg\") pod \"openshift-apiserver-operator-796bbdcf4f-982dt\" (UID: \"de868b10-12d3-4ec1-87d6-f4fcb4994aa3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-982dt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.337709 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/58bab2f1-e131-46cb-b705-6bee97eb8452-audit-policies\") pod \"apiserver-7bbb656c7d-5x7zc\" (UID: \"58bab2f1-e131-46cb-b705-6bee97eb8452\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5x7zc" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.338349 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26690efc-79a1-4bc4-bf0a-78b46fb95223-config\") pod \"machine-approver-56656f9798-mvfqt\" (UID: \"26690efc-79a1-4bc4-bf0a-78b46fb95223\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mvfqt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.339079 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6b4e02e4-f80e-44a5-a2a7-f75feaf2c1ce-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-fpwvv\" (UID: \"6b4e02e4-f80e-44a5-a2a7-f75feaf2c1ce\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fpwvv" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.339227 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/922b208d-fc86-4243-86f1-eacc2797118e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-8dx57\" (UID: \"922b208d-fc86-4243-86f1-eacc2797118e\") " pod="openshift-apiserver/apiserver-76f77b778f-8dx57" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.340250 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/922b208d-fc86-4243-86f1-eacc2797118e-node-pullsecrets\") pod \"apiserver-76f77b778f-8dx57\" (UID: \"922b208d-fc86-4243-86f1-eacc2797118e\") " pod="openshift-apiserver/apiserver-76f77b778f-8dx57" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.340545 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xns4k\" (UniqueName: \"kubernetes.io/projected/26690efc-79a1-4bc4-bf0a-78b46fb95223-kube-api-access-xns4k\") pod \"machine-approver-56656f9798-mvfqt\" (UID: \"26690efc-79a1-4bc4-bf0a-78b46fb95223\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mvfqt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.340570 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7db0eeb4-4ffe-4375-a854-0f47efce6a26-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-prtk8\" (UID: \"7db0eeb4-4ffe-4375-a854-0f47efce6a26\") " pod="openshift-controller-manager/controller-manager-879f6c89f-prtk8" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.340590 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bb8c5\" (UniqueName: \"kubernetes.io/projected/847ef69d-8cd1-4831-8250-05f2cf789d88-kube-api-access-bb8c5\") pod \"machine-api-operator-5694c8668f-d5sv6\" (UID: \"847ef69d-8cd1-4831-8250-05f2cf789d88\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-d5sv6" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.340612 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dfbc\" (UniqueName: \"kubernetes.io/projected/44192b09-31c7-4e49-b472-5ff0a67d30f5-kube-api-access-2dfbc\") pod \"console-operator-58897d9998-rxsb4\" (UID: \"44192b09-31c7-4e49-b472-5ff0a67d30f5\") " pod="openshift-console-operator/console-operator-58897d9998-rxsb4" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.340634 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/792abe2b-bcce-4af6-828b-7a6726b87976-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-p72m9\" (UID: \"792abe2b-bcce-4af6-828b-7a6726b87976\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p72m9" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.340653 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/922b208d-fc86-4243-86f1-eacc2797118e-config\") pod \"apiserver-76f77b778f-8dx57\" (UID: \"922b208d-fc86-4243-86f1-eacc2797118e\") " pod="openshift-apiserver/apiserver-76f77b778f-8dx57" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.340672 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44192b09-31c7-4e49-b472-5ff0a67d30f5-config\") pod \"console-operator-58897d9998-rxsb4\" (UID: \"44192b09-31c7-4e49-b472-5ff0a67d30f5\") " pod="openshift-console-operator/console-operator-58897d9998-rxsb4" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.340693 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/26690efc-79a1-4bc4-bf0a-78b46fb95223-auth-proxy-config\") pod \"machine-approver-56656f9798-mvfqt\" (UID: \"26690efc-79a1-4bc4-bf0a-78b46fb95223\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mvfqt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.340714 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vdrs\" (UniqueName: \"kubernetes.io/projected/3955ee14-c2d4-434e-a9b5-baf95abcbd02-kube-api-access-6vdrs\") pod \"package-server-manager-789f6589d5-5264p\" (UID: \"3955ee14-c2d4-434e-a9b5-baf95abcbd02\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5264p" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.340739 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/847ef69d-8cd1-4831-8250-05f2cf789d88-images\") pod \"machine-api-operator-5694c8668f-d5sv6\" (UID: \"847ef69d-8cd1-4831-8250-05f2cf789d88\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-d5sv6" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.340759 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e690bd42-0ab1-4b3a-802b-4511117b6fe0-trusted-ca-bundle\") pod \"console-f9d7485db-npb7n\" (UID: \"e690bd42-0ab1-4b3a-802b-4511117b6fe0\") " pod="openshift-console/console-f9d7485db-npb7n" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.341153 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/58bab2f1-e131-46cb-b705-6bee97eb8452-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-5x7zc\" (UID: \"58bab2f1-e131-46cb-b705-6bee97eb8452\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5x7zc" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.325861 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.348763 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/922b208d-fc86-4243-86f1-eacc2797118e-node-pullsecrets\") pod \"apiserver-76f77b778f-8dx57\" (UID: \"922b208d-fc86-4243-86f1-eacc2797118e\") " pod="openshift-apiserver/apiserver-76f77b778f-8dx57" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.340254 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/58bab2f1-e131-46cb-b705-6bee97eb8452-audit-dir\") pod \"apiserver-7bbb656c7d-5x7zc\" (UID: \"58bab2f1-e131-46cb-b705-6bee97eb8452\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5x7zc" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.349569 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/58bab2f1-e131-46cb-b705-6bee97eb8452-audit-policies\") pod \"apiserver-7bbb656c7d-5x7zc\" (UID: \"58bab2f1-e131-46cb-b705-6bee97eb8452\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5x7zc" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.349614 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44192b09-31c7-4e49-b472-5ff0a67d30f5-config\") pod \"console-operator-58897d9998-rxsb4\" (UID: \"44192b09-31c7-4e49-b472-5ff0a67d30f5\") " pod="openshift-console-operator/console-operator-58897d9998-rxsb4" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.350199 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/922b208d-fc86-4243-86f1-eacc2797118e-config\") pod \"apiserver-76f77b778f-8dx57\" (UID: \"922b208d-fc86-4243-86f1-eacc2797118e\") " pod="openshift-apiserver/apiserver-76f77b778f-8dx57" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.350536 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7db0eeb4-4ffe-4375-a854-0f47efce6a26-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-prtk8\" (UID: \"7db0eeb4-4ffe-4375-a854-0f47efce6a26\") " pod="openshift-controller-manager/controller-manager-879f6c89f-prtk8" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.350848 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/26690efc-79a1-4bc4-bf0a-78b46fb95223-auth-proxy-config\") pod \"machine-approver-56656f9798-mvfqt\" (UID: \"26690efc-79a1-4bc4-bf0a-78b46fb95223\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mvfqt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.351315 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/847ef69d-8cd1-4831-8250-05f2cf789d88-images\") pod \"machine-api-operator-5694c8668f-d5sv6\" (UID: \"847ef69d-8cd1-4831-8250-05f2cf789d88\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-d5sv6" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.353492 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/792abe2b-bcce-4af6-828b-7a6726b87976-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-p72m9\" (UID: \"792abe2b-bcce-4af6-828b-7a6726b87976\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p72m9" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.358395 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/32ea5b76-e8e5-4906-bc55-99d5ec780ff2-client-ca\") pod \"route-controller-manager-6576b87f9c-trrhn\" (UID: \"32ea5b76-e8e5-4906-bc55-99d5ec780ff2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-trrhn" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.361401 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b4e02e4-f80e-44a5-a2a7-f75feaf2c1ce-config\") pod \"authentication-operator-69f744f599-fpwvv\" (UID: \"6b4e02e4-f80e-44a5-a2a7-f75feaf2c1ce\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fpwvv" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.361496 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/93264a11-973b-435d-a51f-999be787ed0a-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-vjrvb\" (UID: \"93264a11-973b-435d-a51f-999be787ed0a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vjrvb" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.361550 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qn6qd\" (UniqueName: \"kubernetes.io/projected/32ea5b76-e8e5-4906-bc55-99d5ec780ff2-kube-api-access-qn6qd\") pod \"route-controller-manager-6576b87f9c-trrhn\" (UID: \"32ea5b76-e8e5-4906-bc55-99d5ec780ff2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-trrhn" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.361587 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/58bab2f1-e131-46cb-b705-6bee97eb8452-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-5x7zc\" (UID: \"58bab2f1-e131-46cb-b705-6bee97eb8452\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5x7zc" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.361618 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e690bd42-0ab1-4b3a-802b-4511117b6fe0-console-oauth-config\") pod \"console-f9d7485db-npb7n\" (UID: \"e690bd42-0ab1-4b3a-802b-4511117b6fe0\") " pod="openshift-console/console-f9d7485db-npb7n" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.361649 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e690bd42-0ab1-4b3a-802b-4511117b6fe0-oauth-serving-cert\") pod \"console-f9d7485db-npb7n\" (UID: \"e690bd42-0ab1-4b3a-802b-4511117b6fe0\") " pod="openshift-console/console-f9d7485db-npb7n" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.361680 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/847ef69d-8cd1-4831-8250-05f2cf789d88-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-d5sv6\" (UID: \"847ef69d-8cd1-4831-8250-05f2cf789d88\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-d5sv6" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.361706 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/922b208d-fc86-4243-86f1-eacc2797118e-audit\") pod \"apiserver-76f77b778f-8dx57\" (UID: \"922b208d-fc86-4243-86f1-eacc2797118e\") " pod="openshift-apiserver/apiserver-76f77b778f-8dx57" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.361733 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e690bd42-0ab1-4b3a-802b-4511117b6fe0-console-config\") pod \"console-f9d7485db-npb7n\" (UID: \"e690bd42-0ab1-4b3a-802b-4511117b6fe0\") " pod="openshift-console/console-f9d7485db-npb7n" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.361762 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/58bab2f1-e131-46cb-b705-6bee97eb8452-encryption-config\") pod \"apiserver-7bbb656c7d-5x7zc\" (UID: \"58bab2f1-e131-46cb-b705-6bee97eb8452\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5x7zc" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.361833 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/922b208d-fc86-4243-86f1-eacc2797118e-image-import-ca\") pod \"apiserver-76f77b778f-8dx57\" (UID: \"922b208d-fc86-4243-86f1-eacc2797118e\") " pod="openshift-apiserver/apiserver-76f77b778f-8dx57" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.361897 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de868b10-12d3-4ec1-87d6-f4fcb4994aa3-config\") pod \"openshift-apiserver-operator-796bbdcf4f-982dt\" (UID: \"de868b10-12d3-4ec1-87d6-f4fcb4994aa3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-982dt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.361931 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7db0eeb4-4ffe-4375-a854-0f47efce6a26-client-ca\") pod \"controller-manager-879f6c89f-prtk8\" (UID: \"7db0eeb4-4ffe-4375-a854-0f47efce6a26\") " pod="openshift-controller-manager/controller-manager-879f6c89f-prtk8" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.361969 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/9d7d0f56-00a3-4eae-8ff2-a35e1f20a377-available-featuregates\") pod \"openshift-config-operator-7777fb866f-nskw2\" (UID: \"9d7d0f56-00a3-4eae-8ff2-a35e1f20a377\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-nskw2" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.362003 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32ea5b76-e8e5-4906-bc55-99d5ec780ff2-config\") pod \"route-controller-manager-6576b87f9c-trrhn\" (UID: \"32ea5b76-e8e5-4906-bc55-99d5ec780ff2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-trrhn" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.362028 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/58bab2f1-e131-46cb-b705-6bee97eb8452-etcd-client\") pod \"apiserver-7bbb656c7d-5x7zc\" (UID: \"58bab2f1-e131-46cb-b705-6bee97eb8452\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5x7zc" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.362062 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pf4hn\" (UniqueName: \"kubernetes.io/projected/f84f960e-7741-4d13-bd6f-ec2072cd4d14-kube-api-access-pf4hn\") pod \"cluster-samples-operator-665b6dd947-t5v8v\" (UID: \"f84f960e-7741-4d13-bd6f-ec2072cd4d14\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t5v8v" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.362091 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7db0eeb4-4ffe-4375-a854-0f47efce6a26-serving-cert\") pod \"controller-manager-879f6c89f-prtk8\" (UID: \"7db0eeb4-4ffe-4375-a854-0f47efce6a26\") " pod="openshift-controller-manager/controller-manager-879f6c89f-prtk8" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.362115 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/922b208d-fc86-4243-86f1-eacc2797118e-serving-cert\") pod \"apiserver-76f77b778f-8dx57\" (UID: \"922b208d-fc86-4243-86f1-eacc2797118e\") " pod="openshift-apiserver/apiserver-76f77b778f-8dx57" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.362141 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9bjs\" (UniqueName: \"kubernetes.io/projected/e690bd42-0ab1-4b3a-802b-4511117b6fe0-kube-api-access-b9bjs\") pod \"console-f9d7485db-npb7n\" (UID: \"e690bd42-0ab1-4b3a-802b-4511117b6fe0\") " pod="openshift-console/console-f9d7485db-npb7n" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.362178 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/922b208d-fc86-4243-86f1-eacc2797118e-etcd-serving-ca\") pod \"apiserver-76f77b778f-8dx57\" (UID: \"922b208d-fc86-4243-86f1-eacc2797118e\") " pod="openshift-apiserver/apiserver-76f77b778f-8dx57" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.362207 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/922b208d-fc86-4243-86f1-eacc2797118e-audit-dir\") pod \"apiserver-76f77b778f-8dx57\" (UID: \"922b208d-fc86-4243-86f1-eacc2797118e\") " pod="openshift-apiserver/apiserver-76f77b778f-8dx57" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.362231 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-st8cn\" (UniqueName: \"kubernetes.io/projected/922b208d-fc86-4243-86f1-eacc2797118e-kube-api-access-st8cn\") pod \"apiserver-76f77b778f-8dx57\" (UID: \"922b208d-fc86-4243-86f1-eacc2797118e\") " pod="openshift-apiserver/apiserver-76f77b778f-8dx57" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.363226 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.363304 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/58bab2f1-e131-46cb-b705-6bee97eb8452-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-5x7zc\" (UID: \"58bab2f1-e131-46cb-b705-6bee97eb8452\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5x7zc" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.370867 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z7bll"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.372465 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/44192b09-31c7-4e49-b472-5ff0a67d30f5-trusted-ca\") pod \"console-operator-58897d9998-rxsb4\" (UID: \"44192b09-31c7-4e49-b472-5ff0a67d30f5\") " pod="openshift-console-operator/console-operator-58897d9998-rxsb4" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.374153 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e690bd42-0ab1-4b3a-802b-4511117b6fe0-trusted-ca-bundle\") pod \"console-f9d7485db-npb7n\" (UID: \"e690bd42-0ab1-4b3a-802b-4511117b6fe0\") " pod="openshift-console/console-f9d7485db-npb7n" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.374559 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/44192b09-31c7-4e49-b472-5ff0a67d30f5-serving-cert\") pod \"console-operator-58897d9998-rxsb4\" (UID: \"44192b09-31c7-4e49-b472-5ff0a67d30f5\") " pod="openshift-console-operator/console-operator-58897d9998-rxsb4" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.374828 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e690bd42-0ab1-4b3a-802b-4511117b6fe0-console-serving-cert\") pod \"console-f9d7485db-npb7n\" (UID: \"e690bd42-0ab1-4b3a-802b-4511117b6fe0\") " pod="openshift-console/console-f9d7485db-npb7n" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.375206 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32ea5b76-e8e5-4906-bc55-99d5ec780ff2-serving-cert\") pod \"route-controller-manager-6576b87f9c-trrhn\" (UID: \"32ea5b76-e8e5-4906-bc55-99d5ec780ff2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-trrhn" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.375324 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/de868b10-12d3-4ec1-87d6-f4fcb4994aa3-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-982dt\" (UID: \"de868b10-12d3-4ec1-87d6-f4fcb4994aa3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-982dt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.375616 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/26690efc-79a1-4bc4-bf0a-78b46fb95223-machine-approver-tls\") pod \"machine-approver-56656f9798-mvfqt\" (UID: \"26690efc-79a1-4bc4-bf0a-78b46fb95223\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mvfqt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.375839 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e690bd42-0ab1-4b3a-802b-4511117b6fe0-console-oauth-config\") pod \"console-f9d7485db-npb7n\" (UID: \"e690bd42-0ab1-4b3a-802b-4511117b6fe0\") " pod="openshift-console/console-f9d7485db-npb7n" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.375922 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/922b208d-fc86-4243-86f1-eacc2797118e-encryption-config\") pod \"apiserver-76f77b778f-8dx57\" (UID: \"922b208d-fc86-4243-86f1-eacc2797118e\") " pod="openshift-apiserver/apiserver-76f77b778f-8dx57" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.376286 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/93264a11-973b-435d-a51f-999be787ed0a-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-vjrvb\" (UID: \"93264a11-973b-435d-a51f-999be787ed0a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vjrvb" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.377591 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93264a11-973b-435d-a51f-999be787ed0a-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-vjrvb\" (UID: \"93264a11-973b-435d-a51f-999be787ed0a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vjrvb" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.379657 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-64cqc"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.381984 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e690bd42-0ab1-4b3a-802b-4511117b6fe0-oauth-serving-cert\") pod \"console-f9d7485db-npb7n\" (UID: \"e690bd42-0ab1-4b3a-802b-4511117b6fe0\") " pod="openshift-console/console-f9d7485db-npb7n" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.382122 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/58bab2f1-e131-46cb-b705-6bee97eb8452-serving-cert\") pod \"apiserver-7bbb656c7d-5x7zc\" (UID: \"58bab2f1-e131-46cb-b705-6bee97eb8452\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5x7zc" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.382880 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6b4e02e4-f80e-44a5-a2a7-f75feaf2c1ce-service-ca-bundle\") pod \"authentication-operator-69f744f599-fpwvv\" (UID: \"6b4e02e4-f80e-44a5-a2a7-f75feaf2c1ce\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fpwvv" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.383953 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e690bd42-0ab1-4b3a-802b-4511117b6fe0-service-ca\") pod \"console-f9d7485db-npb7n\" (UID: \"e690bd42-0ab1-4b3a-802b-4511117b6fe0\") " pod="openshift-console/console-f9d7485db-npb7n" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.384752 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6b4e02e4-f80e-44a5-a2a7-f75feaf2c1ce-serving-cert\") pod \"authentication-operator-69f744f599-fpwvv\" (UID: \"6b4e02e4-f80e-44a5-a2a7-f75feaf2c1ce\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fpwvv" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.386083 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7db0eeb4-4ffe-4375-a854-0f47efce6a26-config\") pod \"controller-manager-879f6c89f-prtk8\" (UID: \"7db0eeb4-4ffe-4375-a854-0f47efce6a26\") " pod="openshift-controller-manager/controller-manager-879f6c89f-prtk8" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.389767 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.390709 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/922b208d-fc86-4243-86f1-eacc2797118e-audit\") pod \"apiserver-76f77b778f-8dx57\" (UID: \"922b208d-fc86-4243-86f1-eacc2797118e\") " pod="openshift-apiserver/apiserver-76f77b778f-8dx57" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.391560 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/922b208d-fc86-4243-86f1-eacc2797118e-image-import-ca\") pod \"apiserver-76f77b778f-8dx57\" (UID: \"922b208d-fc86-4243-86f1-eacc2797118e\") " pod="openshift-apiserver/apiserver-76f77b778f-8dx57" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.392529 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de868b10-12d3-4ec1-87d6-f4fcb4994aa3-config\") pod \"openshift-apiserver-operator-796bbdcf4f-982dt\" (UID: \"de868b10-12d3-4ec1-87d6-f4fcb4994aa3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-982dt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.392731 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e690bd42-0ab1-4b3a-802b-4511117b6fe0-console-config\") pod \"console-f9d7485db-npb7n\" (UID: \"e690bd42-0ab1-4b3a-802b-4511117b6fe0\") " pod="openshift-console/console-f9d7485db-npb7n" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.396303 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32ea5b76-e8e5-4906-bc55-99d5ec780ff2-config\") pod \"route-controller-manager-6576b87f9c-trrhn\" (UID: \"32ea5b76-e8e5-4906-bc55-99d5ec780ff2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-trrhn" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.390314 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.397226 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d7d0f56-00a3-4eae-8ff2-a35e1f20a377-serving-cert\") pod \"openshift-config-operator-7777fb866f-nskw2\" (UID: \"9d7d0f56-00a3-4eae-8ff2-a35e1f20a377\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-nskw2" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.397767 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/847ef69d-8cd1-4831-8250-05f2cf789d88-config\") pod \"machine-api-operator-5694c8668f-d5sv6\" (UID: \"847ef69d-8cd1-4831-8250-05f2cf789d88\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-d5sv6" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.397930 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/792abe2b-bcce-4af6-828b-7a6726b87976-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-p72m9\" (UID: \"792abe2b-bcce-4af6-828b-7a6726b87976\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p72m9" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.398896 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/9d7d0f56-00a3-4eae-8ff2-a35e1f20a377-available-featuregates\") pod \"openshift-config-operator-7777fb866f-nskw2\" (UID: \"9d7d0f56-00a3-4eae-8ff2-a35e1f20a377\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-nskw2" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.399063 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/922b208d-fc86-4243-86f1-eacc2797118e-audit-dir\") pod \"apiserver-76f77b778f-8dx57\" (UID: \"922b208d-fc86-4243-86f1-eacc2797118e\") " pod="openshift-apiserver/apiserver-76f77b778f-8dx57" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.399702 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/922b208d-fc86-4243-86f1-eacc2797118e-etcd-serving-ca\") pod \"apiserver-76f77b778f-8dx57\" (UID: \"922b208d-fc86-4243-86f1-eacc2797118e\") " pod="openshift-apiserver/apiserver-76f77b778f-8dx57" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.402015 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-6hmm5" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.402858 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7db0eeb4-4ffe-4375-a854-0f47efce6a26-client-ca\") pod \"controller-manager-879f6c89f-prtk8\" (UID: \"7db0eeb4-4ffe-4375-a854-0f47efce6a26\") " pod="openshift-controller-manager/controller-manager-879f6c89f-prtk8" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.403514 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.406357 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z7bll" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.420012 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-fpwvv"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.420077 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-d5sv6"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.420088 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-8dx57"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.420097 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p72m9"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.420183 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-nskw2"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.420219 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-64cqc" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.420245 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-trrhn"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.420257 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qbxwp"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.420271 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-982dt"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.421144 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-npb7n"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.425114 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-vnwxg"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.425237 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vjrvb"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.427273 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-5x7zc"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.430756 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/847ef69d-8cd1-4831-8250-05f2cf789d88-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-d5sv6\" (UID: \"847ef69d-8cd1-4831-8250-05f2cf789d88\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-d5sv6" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.431685 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/922b208d-fc86-4243-86f1-eacc2797118e-etcd-client\") pod \"apiserver-76f77b778f-8dx57\" (UID: \"922b208d-fc86-4243-86f1-eacc2797118e\") " pod="openshift-apiserver/apiserver-76f77b778f-8dx57" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.432695 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/922b208d-fc86-4243-86f1-eacc2797118e-serving-cert\") pod \"apiserver-76f77b778f-8dx57\" (UID: \"922b208d-fc86-4243-86f1-eacc2797118e\") " pod="openshift-apiserver/apiserver-76f77b778f-8dx57" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.433070 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8mkhz"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.433365 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.436875 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-qdg4k"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.439474 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-vfbvk"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.441693 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cb48c"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.443049 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5264p"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.444719 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-brmrn"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.446168 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vtd5z"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.447831 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-5s2bx"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.448327 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7db0eeb4-4ffe-4375-a854-0f47efce6a26-serving-cert\") pod \"controller-manager-879f6c89f-prtk8\" (UID: \"7db0eeb4-4ffe-4375-a854-0f47efce6a26\") " pod="openshift-controller-manager/controller-manager-879f6c89f-prtk8" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.449258 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t5v8v"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.449405 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-5s2bx" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.450529 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-45lhd"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.451610 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.456561 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z7bll"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.456582 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-rxsb4"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.456592 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-wdpjf"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.456601 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9c9zp"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.456655 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-45lhd" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.456796 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8t8vm"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.457930 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8l455"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.459128 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-4xvzn"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.460412 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/58bab2f1-e131-46cb-b705-6bee97eb8452-encryption-config\") pod \"apiserver-7bbb656c7d-5x7zc\" (UID: \"58bab2f1-e131-46cb-b705-6bee97eb8452\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5x7zc" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.460559 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f7ccb"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.460560 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/58bab2f1-e131-46cb-b705-6bee97eb8452-etcd-client\") pod \"apiserver-7bbb656c7d-5x7zc\" (UID: \"58bab2f1-e131-46cb-b705-6bee97eb8452\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5x7zc" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.462273 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-sv4jj"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.462991 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b3dbb323-561b-4d26-886c-bef115c0dca0-trusted-ca\") pod \"ingress-operator-5b745b69d9-qdg4k\" (UID: \"b3dbb323-561b-4d26-886c-bef115c0dca0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qdg4k" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.463012 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b3dbb323-561b-4d26-886c-bef115c0dca0-bound-sa-token\") pod \"ingress-operator-5b745b69d9-qdg4k\" (UID: \"b3dbb323-561b-4d26-886c-bef115c0dca0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qdg4k" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.463031 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/f84f960e-7741-4d13-bd6f-ec2072cd4d14-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-t5v8v\" (UID: \"f84f960e-7741-4d13-bd6f-ec2072cd4d14\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t5v8v" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.463052 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thvn9\" (UniqueName: \"kubernetes.io/projected/b3dbb323-561b-4d26-886c-bef115c0dca0-kube-api-access-thvn9\") pod \"ingress-operator-5b745b69d9-qdg4k\" (UID: \"b3dbb323-561b-4d26-886c-bef115c0dca0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qdg4k" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.463098 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vdrs\" (UniqueName: \"kubernetes.io/projected/3955ee14-c2d4-434e-a9b5-baf95abcbd02-kube-api-access-6vdrs\") pod \"package-server-manager-789f6589d5-5264p\" (UID: \"3955ee14-c2d4-434e-a9b5-baf95abcbd02\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5264p" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.463151 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pf4hn\" (UniqueName: \"kubernetes.io/projected/f84f960e-7741-4d13-bd6f-ec2072cd4d14-kube-api-access-pf4hn\") pod \"cluster-samples-operator-665b6dd947-t5v8v\" (UID: \"f84f960e-7741-4d13-bd6f-ec2072cd4d14\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t5v8v" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.463179 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3955ee14-c2d4-434e-a9b5-baf95abcbd02-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-5264p\" (UID: \"3955ee14-c2d4-434e-a9b5-baf95abcbd02\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5264p" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.463244 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g28kf\" (UniqueName: \"kubernetes.io/projected/ff950861-7bca-4bfa-9857-cd906f67d5cb-kube-api-access-g28kf\") pod \"migrator-59844c95c7-4xvzn\" (UID: \"ff950861-7bca-4bfa-9857-cd906f67d5cb\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4xvzn" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.463289 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b3dbb323-561b-4d26-886c-bef115c0dca0-metrics-tls\") pod \"ingress-operator-5b745b69d9-qdg4k\" (UID: \"b3dbb323-561b-4d26-886c-bef115c0dca0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qdg4k" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.463636 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29522220-v5twm"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.464958 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4hhxl"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.466247 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-45lhd"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.467258 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/f84f960e-7741-4d13-bd6f-ec2072cd4d14-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-t5v8v\" (UID: \"f84f960e-7741-4d13-bd6f-ec2072cd4d14\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t5v8v" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.468121 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-qz68q"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.468597 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-64cqc"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.469767 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-9tpfm"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.470955 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-6hmm5"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.471973 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.472312 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-hs5dk"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.473312 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-hs5dk" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.473523 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-zc8xw"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.474717 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-hs5dk"] Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.491538 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.512398 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.533321 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.552760 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.572596 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.592167 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.612600 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.632386 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.652506 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.680432 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.692961 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.697559 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b3dbb323-561b-4d26-886c-bef115c0dca0-metrics-tls\") pod \"ingress-operator-5b745b69d9-qdg4k\" (UID: \"b3dbb323-561b-4d26-886c-bef115c0dca0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qdg4k" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.719739 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.726207 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b3dbb323-561b-4d26-886c-bef115c0dca0-trusted-ca\") pod \"ingress-operator-5b745b69d9-qdg4k\" (UID: \"b3dbb323-561b-4d26-886c-bef115c0dca0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qdg4k" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.733384 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.751750 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.773188 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.792398 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.813102 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.833320 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.853220 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.873586 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.903067 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.913438 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.942497 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.953387 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 17 13:06:42 crc kubenswrapper[4955]: I0217 13:06:42.973108 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.003555 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.013805 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.031688 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.053422 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.072215 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.092926 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.112698 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.132431 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.152609 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.172467 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.192612 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.212508 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.233084 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.252266 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.271320 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.291071 4955 request.go:700] Waited for 1.004799451s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/secrets?fieldSelector=metadata.name%3Dpackage-server-manager-serving-cert&limit=500&resourceVersion=0 Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.293203 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.299520 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3955ee14-c2d4-434e-a9b5-baf95abcbd02-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-5264p\" (UID: \"3955ee14-c2d4-434e-a9b5-baf95abcbd02\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5264p" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.332163 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.353184 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.372887 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.393236 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.412857 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.433825 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.452630 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.473352 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.492331 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.512963 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.533196 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.552831 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.573330 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.593684 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.614042 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.632820 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.652032 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.672198 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.692752 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.712522 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.732190 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.752379 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.772054 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.810690 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.813236 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.832561 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.852662 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.890695 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmtdf\" (UniqueName: \"kubernetes.io/projected/7db0eeb4-4ffe-4375-a854-0f47efce6a26-kube-api-access-cmtdf\") pod \"controller-manager-879f6c89f-prtk8\" (UID: \"7db0eeb4-4ffe-4375-a854-0f47efce6a26\") " pod="openshift-controller-manager/controller-manager-879f6c89f-prtk8" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.908174 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bb8c5\" (UniqueName: \"kubernetes.io/projected/847ef69d-8cd1-4831-8250-05f2cf789d88-kube-api-access-bb8c5\") pod \"machine-api-operator-5694c8668f-d5sv6\" (UID: \"847ef69d-8cd1-4831-8250-05f2cf789d88\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-d5sv6" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.930947 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/792abe2b-bcce-4af6-828b-7a6726b87976-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-p72m9\" (UID: \"792abe2b-bcce-4af6-828b-7a6726b87976\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p72m9" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.950001 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-prtk8" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.960562 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92cx5\" (UniqueName: \"kubernetes.io/projected/792abe2b-bcce-4af6-828b-7a6726b87976-kube-api-access-92cx5\") pod \"cluster-image-registry-operator-dc59b4c8b-p72m9\" (UID: \"792abe2b-bcce-4af6-828b-7a6726b87976\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p72m9" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.969877 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nczq7\" (UniqueName: \"kubernetes.io/projected/58bab2f1-e131-46cb-b705-6bee97eb8452-kube-api-access-nczq7\") pod \"apiserver-7bbb656c7d-5x7zc\" (UID: \"58bab2f1-e131-46cb-b705-6bee97eb8452\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5x7zc" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.972572 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-d5sv6" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.990234 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p72m9" Feb 17 13:06:43 crc kubenswrapper[4955]: I0217 13:06:43.993067 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbbcg\" (UniqueName: \"kubernetes.io/projected/de868b10-12d3-4ec1-87d6-f4fcb4994aa3-kube-api-access-cbbcg\") pod \"openshift-apiserver-operator-796bbdcf4f-982dt\" (UID: \"de868b10-12d3-4ec1-87d6-f4fcb4994aa3\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-982dt" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.000805 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-982dt" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.009988 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xns4k\" (UniqueName: \"kubernetes.io/projected/26690efc-79a1-4bc4-bf0a-78b46fb95223-kube-api-access-xns4k\") pod \"machine-approver-56656f9798-mvfqt\" (UID: \"26690efc-79a1-4bc4-bf0a-78b46fb95223\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mvfqt" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.019146 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5x7zc" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.046093 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dfbc\" (UniqueName: \"kubernetes.io/projected/44192b09-31c7-4e49-b472-5ff0a67d30f5-kube-api-access-2dfbc\") pod \"console-operator-58897d9998-rxsb4\" (UID: \"44192b09-31c7-4e49-b472-5ff0a67d30f5\") " pod="openshift-console-operator/console-operator-58897d9998-rxsb4" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.054450 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgznr\" (UniqueName: \"kubernetes.io/projected/6b4e02e4-f80e-44a5-a2a7-f75feaf2c1ce-kube-api-access-fgznr\") pod \"authentication-operator-69f744f599-fpwvv\" (UID: \"6b4e02e4-f80e-44a5-a2a7-f75feaf2c1ce\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fpwvv" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.067931 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-fpwvv" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.077923 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rf2mz\" (UniqueName: \"kubernetes.io/projected/9d7d0f56-00a3-4eae-8ff2-a35e1f20a377-kube-api-access-rf2mz\") pod \"openshift-config-operator-7777fb866f-nskw2\" (UID: \"9d7d0f56-00a3-4eae-8ff2-a35e1f20a377\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-nskw2" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.100820 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4p86\" (UniqueName: \"kubernetes.io/projected/93264a11-973b-435d-a51f-999be787ed0a-kube-api-access-d4p86\") pod \"openshift-controller-manager-operator-756b6f6bc6-vjrvb\" (UID: \"93264a11-973b-435d-a51f-999be787ed0a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vjrvb" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.101979 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mvfqt" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.116732 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-rxsb4" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.126302 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-st8cn\" (UniqueName: \"kubernetes.io/projected/922b208d-fc86-4243-86f1-eacc2797118e-kube-api-access-st8cn\") pod \"apiserver-76f77b778f-8dx57\" (UID: \"922b208d-fc86-4243-86f1-eacc2797118e\") " pod="openshift-apiserver/apiserver-76f77b778f-8dx57" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.145329 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qn6qd\" (UniqueName: \"kubernetes.io/projected/32ea5b76-e8e5-4906-bc55-99d5ec780ff2-kube-api-access-qn6qd\") pod \"route-controller-manager-6576b87f9c-trrhn\" (UID: \"32ea5b76-e8e5-4906-bc55-99d5ec780ff2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-trrhn" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.153412 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqclt\" (UniqueName: \"kubernetes.io/projected/01aa5d18-e942-4a3d-8af5-1e41b503f85a-kube-api-access-vqclt\") pod \"downloads-7954f5f757-qz68q\" (UID: \"01aa5d18-e942-4a3d-8af5-1e41b503f85a\") " pod="openshift-console/downloads-7954f5f757-qz68q" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.172138 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.174328 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9bjs\" (UniqueName: \"kubernetes.io/projected/e690bd42-0ab1-4b3a-802b-4511117b6fe0-kube-api-access-b9bjs\") pod \"console-f9d7485db-npb7n\" (UID: \"e690bd42-0ab1-4b3a-802b-4511117b6fe0\") " pod="openshift-console/console-f9d7485db-npb7n" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.192736 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.212998 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.230706 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-qz68q" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.233587 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.239744 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vjrvb" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.247247 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-nskw2" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.252961 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.272659 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.292569 4955 request.go:700] Waited for 1.858560274s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/pods/cluster-image-registry-operator-dc59b4c8b-p72m9 Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.313570 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.334101 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.349095 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-trrhn" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.361200 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.372532 4955 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.389515 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-8dx57" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.392462 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.418489 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.464671 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-npb7n" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.468996 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thvn9\" (UniqueName: \"kubernetes.io/projected/b3dbb323-561b-4d26-886c-bef115c0dca0-kube-api-access-thvn9\") pod \"ingress-operator-5b745b69d9-qdg4k\" (UID: \"b3dbb323-561b-4d26-886c-bef115c0dca0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qdg4k" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.469995 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pf4hn\" (UniqueName: \"kubernetes.io/projected/f84f960e-7741-4d13-bd6f-ec2072cd4d14-kube-api-access-pf4hn\") pod \"cluster-samples-operator-665b6dd947-t5v8v\" (UID: \"f84f960e-7741-4d13-bd6f-ec2072cd4d14\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t5v8v" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.489113 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b3dbb323-561b-4d26-886c-bef115c0dca0-bound-sa-token\") pod \"ingress-operator-5b745b69d9-qdg4k\" (UID: \"b3dbb323-561b-4d26-886c-bef115c0dca0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qdg4k" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.494132 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-rxsb4"] Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.496221 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p72m9"] Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.496506 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-prtk8"] Feb 17 13:06:44 crc kubenswrapper[4955]: W0217 13:06:44.509956 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod44192b09_31c7_4e49_b472_5ff0a67d30f5.slice/crio-1401cc5130f7a83a135b176df21f718ed3f9da1d2ca49816cb4fe1ecf0d43f98 WatchSource:0}: Error finding container 1401cc5130f7a83a135b176df21f718ed3f9da1d2ca49816cb4fe1ecf0d43f98: Status 404 returned error can't find the container with id 1401cc5130f7a83a135b176df21f718ed3f9da1d2ca49816cb4fe1ecf0d43f98 Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.523801 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vdrs\" (UniqueName: \"kubernetes.io/projected/3955ee14-c2d4-434e-a9b5-baf95abcbd02-kube-api-access-6vdrs\") pod \"package-server-manager-789f6589d5-5264p\" (UID: \"3955ee14-c2d4-434e-a9b5-baf95abcbd02\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5264p" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.525357 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-d5sv6"] Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.532303 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g28kf\" (UniqueName: \"kubernetes.io/projected/ff950861-7bca-4bfa-9857-cd906f67d5cb-kube-api-access-g28kf\") pod \"migrator-59844c95c7-4xvzn\" (UID: \"ff950861-7bca-4bfa-9857-cd906f67d5cb\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4xvzn" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.537543 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.547944 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-fpwvv"] Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.552337 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.567395 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t5v8v" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.575056 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.591662 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.607792 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qdg4k" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.614572 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-5x7zc"] Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.615622 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-982dt"] Feb 17 13:06:44 crc kubenswrapper[4955]: W0217 13:06:44.637997 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod58bab2f1_e131_46cb_b705_6bee97eb8452.slice/crio-0d4c8764bd6735cee843dc34a3a2304b432cccd7e81a6b1df4a329eb8550cfa3 WatchSource:0}: Error finding container 0d4c8764bd6735cee843dc34a3a2304b432cccd7e81a6b1df4a329eb8550cfa3: Status 404 returned error can't find the container with id 0d4c8764bd6735cee843dc34a3a2304b432cccd7e81a6b1df4a329eb8550cfa3 Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.651144 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4xvzn" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.671222 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5264p" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.683767 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-trrhn"] Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697009 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0b5c785-906c-4d00-93b3-36111e5d38da-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-8l455\" (UID: \"e0b5c785-906c-4d00-93b3-36111e5d38da\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8l455" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697037 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e3687198-a53d-4a80-baec-bafd1c5ef14d-trusted-ca\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697058 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1546cd3a-91cf-4e78-91e3-46ee4b8fbb43-config\") pod \"kube-controller-manager-operator-78b949d7b-8t8vm\" (UID: \"1546cd3a-91cf-4e78-91e3-46ee4b8fbb43\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8t8vm" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697108 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-vfbvk\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697125 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-vfbvk\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697145 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-vfbvk\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697168 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/89d41921-f56f-4e2e-8ef0-2d981933ad68-profile-collector-cert\") pod \"catalog-operator-68c6474976-4hhxl\" (UID: \"89d41921-f56f-4e2e-8ef0-2d981933ad68\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4hhxl" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697195 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/784c9c2c-64a9-4887-b52c-0949da50de62-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-qbxwp\" (UID: \"784c9c2c-64a9-4887-b52c-0949da50de62\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qbxwp" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697219 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a800c1f4-8f07-4d13-a112-e3e4073c0241-images\") pod \"machine-config-operator-74547568cd-sv4jj\" (UID: \"a800c1f4-8f07-4d13-a112-e3e4073c0241\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sv4jj" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697247 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc45d0df-2f2d-449e-9f66-3d6cb79d1821-config\") pod \"etcd-operator-b45778765-wdpjf\" (UID: \"fc45d0df-2f2d-449e-9f66-3d6cb79d1821\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wdpjf" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697264 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-vfbvk\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697305 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-vfbvk\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697322 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2vvx\" (UniqueName: \"kubernetes.io/projected/89d41921-f56f-4e2e-8ef0-2d981933ad68-kube-api-access-s2vvx\") pod \"catalog-operator-68c6474976-4hhxl\" (UID: \"89d41921-f56f-4e2e-8ef0-2d981933ad68\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4hhxl" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697346 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/fc45d0df-2f2d-449e-9f66-3d6cb79d1821-etcd-ca\") pod \"etcd-operator-b45778765-wdpjf\" (UID: \"fc45d0df-2f2d-449e-9f66-3d6cb79d1821\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wdpjf" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697362 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f86fa65-b079-4331-b597-63a6c9995849-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cb48c\" (UID: \"2f86fa65-b079-4331-b597-63a6c9995849\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cb48c" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697376 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/89d41921-f56f-4e2e-8ef0-2d981933ad68-srv-cert\") pod \"catalog-operator-68c6474976-4hhxl\" (UID: \"89d41921-f56f-4e2e-8ef0-2d981933ad68\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4hhxl" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697389 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e3687198-a53d-4a80-baec-bafd1c5ef14d-installation-pull-secrets\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697404 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/784c9c2c-64a9-4887-b52c-0949da50de62-config\") pod \"kube-apiserver-operator-766d6c64bb-qbxwp\" (UID: \"784c9c2c-64a9-4887-b52c-0949da50de62\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qbxwp" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697445 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e3687198-a53d-4a80-baec-bafd1c5ef14d-registry-tls\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697460 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-vfbvk\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697475 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f7b03b61-99a7-4db3-be68-e2ad2149f44b-audit-dir\") pod \"oauth-openshift-558db77b4-vfbvk\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697501 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/784c9c2c-64a9-4887-b52c-0949da50de62-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-qbxwp\" (UID: \"784c9c2c-64a9-4887-b52c-0949da50de62\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qbxwp" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697517 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3a2b80bf-ead8-48e2-95fe-837c9ce39f61-metrics-tls\") pod \"dns-operator-744455d44c-vnwxg\" (UID: \"3a2b80bf-ead8-48e2-95fe-837c9ce39f61\") " pod="openshift-dns-operator/dns-operator-744455d44c-vnwxg" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697532 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e3687198-a53d-4a80-baec-bafd1c5ef14d-bound-sa-token\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697573 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a800c1f4-8f07-4d13-a112-e3e4073c0241-auth-proxy-config\") pod \"machine-config-operator-74547568cd-sv4jj\" (UID: \"a800c1f4-8f07-4d13-a112-e3e4073c0241\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sv4jj" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697589 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f7b03b61-99a7-4db3-be68-e2ad2149f44b-audit-policies\") pod \"oauth-openshift-558db77b4-vfbvk\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697608 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9rzm\" (UniqueName: \"kubernetes.io/projected/8f5feb72-96ca-4e4d-9566-0757c7dd93f9-kube-api-access-g9rzm\") pod \"control-plane-machine-set-operator-78cbb6b69f-vtd5z\" (UID: \"8f5feb72-96ca-4e4d-9566-0757c7dd93f9\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vtd5z" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697624 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e3687198-a53d-4a80-baec-bafd1c5ef14d-ca-trust-extracted\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697653 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697671 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a800c1f4-8f07-4d13-a112-e3e4073c0241-proxy-tls\") pod \"machine-config-operator-74547568cd-sv4jj\" (UID: \"a800c1f4-8f07-4d13-a112-e3e4073c0241\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sv4jj" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697688 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-vfbvk\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697702 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-vfbvk\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697726 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e3687198-a53d-4a80-baec-bafd1c5ef14d-registry-certificates\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697743 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9vhk\" (UniqueName: \"kubernetes.io/projected/fc45d0df-2f2d-449e-9f66-3d6cb79d1821-kube-api-access-j9vhk\") pod \"etcd-operator-b45778765-wdpjf\" (UID: \"fc45d0df-2f2d-449e-9f66-3d6cb79d1821\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wdpjf" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697767 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0b5c785-906c-4d00-93b3-36111e5d38da-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-8l455\" (UID: \"e0b5c785-906c-4d00-93b3-36111e5d38da\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8l455" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697815 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-vfbvk\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697832 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/fc45d0df-2f2d-449e-9f66-3d6cb79d1821-etcd-service-ca\") pod \"etcd-operator-b45778765-wdpjf\" (UID: \"fc45d0df-2f2d-449e-9f66-3d6cb79d1821\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wdpjf" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697846 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f86fa65-b079-4331-b597-63a6c9995849-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cb48c\" (UID: \"2f86fa65-b079-4331-b597-63a6c9995849\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cb48c" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697860 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1546cd3a-91cf-4e78-91e3-46ee4b8fbb43-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-8t8vm\" (UID: \"1546cd3a-91cf-4e78-91e3-46ee4b8fbb43\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8t8vm" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697886 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnjvv\" (UniqueName: \"kubernetes.io/projected/3a2b80bf-ead8-48e2-95fe-837c9ce39f61-kube-api-access-fnjvv\") pod \"dns-operator-744455d44c-vnwxg\" (UID: \"3a2b80bf-ead8-48e2-95fe-837c9ce39f61\") " pod="openshift-dns-operator/dns-operator-744455d44c-vnwxg" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697903 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcnnq\" (UniqueName: \"kubernetes.io/projected/a800c1f4-8f07-4d13-a112-e3e4073c0241-kube-api-access-hcnnq\") pod \"machine-config-operator-74547568cd-sv4jj\" (UID: \"a800c1f4-8f07-4d13-a112-e3e4073c0241\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sv4jj" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697949 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-vfbvk\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697974 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1546cd3a-91cf-4e78-91e3-46ee4b8fbb43-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-8t8vm\" (UID: \"1546cd3a-91cf-4e78-91e3-46ee4b8fbb43\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8t8vm" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.697989 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdfkn\" (UniqueName: \"kubernetes.io/projected/e0b5c785-906c-4d00-93b3-36111e5d38da-kube-api-access-gdfkn\") pod \"kube-storage-version-migrator-operator-b67b599dd-8l455\" (UID: \"e0b5c785-906c-4d00-93b3-36111e5d38da\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8l455" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.698006 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-vfbvk\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.698022 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc45d0df-2f2d-449e-9f66-3d6cb79d1821-serving-cert\") pod \"etcd-operator-b45778765-wdpjf\" (UID: \"fc45d0df-2f2d-449e-9f66-3d6cb79d1821\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wdpjf" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.698055 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whm4d\" (UniqueName: \"kubernetes.io/projected/e3687198-a53d-4a80-baec-bafd1c5ef14d-kube-api-access-whm4d\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.698074 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f86fa65-b079-4331-b597-63a6c9995849-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cb48c\" (UID: \"2f86fa65-b079-4331-b597-63a6c9995849\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cb48c" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.698097 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/8f5feb72-96ca-4e4d-9566-0757c7dd93f9-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-vtd5z\" (UID: \"8f5feb72-96ca-4e4d-9566-0757c7dd93f9\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vtd5z" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.698123 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5ftl\" (UniqueName: \"kubernetes.io/projected/f7b03b61-99a7-4db3-be68-e2ad2149f44b-kube-api-access-g5ftl\") pod \"oauth-openshift-558db77b4-vfbvk\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.698136 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fc45d0df-2f2d-449e-9f66-3d6cb79d1821-etcd-client\") pod \"etcd-operator-b45778765-wdpjf\" (UID: \"fc45d0df-2f2d-449e-9f66-3d6cb79d1821\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wdpjf" Feb 17 13:06:44 crc kubenswrapper[4955]: E0217 13:06:44.702451 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:45.20243689 +0000 UTC m=+143.725166433 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.727451 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-npb7n"] Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.759187 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-nskw2"] Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.764860 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-qz68q"] Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.798937 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:44 crc kubenswrapper[4955]: E0217 13:06:44.799040 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:45.299022286 +0000 UTC m=+143.821751829 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.799205 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccx4h\" (UniqueName: \"kubernetes.io/projected/c46d5d3e-92cd-415f-a04d-59756dd57637-kube-api-access-ccx4h\") pod \"olm-operator-6b444d44fb-f7ccb\" (UID: \"c46d5d3e-92cd-415f-a04d-59756dd57637\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f7ccb" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.799227 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/cea310be-8fea-43cf-a000-69b614f260e4-default-certificate\") pod \"router-default-5444994796-m7gfn\" (UID: \"cea310be-8fea-43cf-a000-69b614f260e4\") " pod="openshift-ingress/router-default-5444994796-m7gfn" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.799253 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a800c1f4-8f07-4d13-a112-e3e4073c0241-auth-proxy-config\") pod \"machine-config-operator-74547568cd-sv4jj\" (UID: \"a800c1f4-8f07-4d13-a112-e3e4073c0241\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sv4jj" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.799282 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f7b03b61-99a7-4db3-be68-e2ad2149f44b-audit-policies\") pod \"oauth-openshift-558db77b4-vfbvk\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.799307 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9rzm\" (UniqueName: \"kubernetes.io/projected/8f5feb72-96ca-4e4d-9566-0757c7dd93f9-kube-api-access-g9rzm\") pod \"control-plane-machine-set-operator-78cbb6b69f-vtd5z\" (UID: \"8f5feb72-96ca-4e4d-9566-0757c7dd93f9\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vtd5z" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.799323 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e3687198-a53d-4a80-baec-bafd1c5ef14d-ca-trust-extracted\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.799355 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e8351820-e7b6-4477-9944-5875c76514ce-config-volume\") pod \"dns-default-64cqc\" (UID: \"e8351820-e7b6-4477-9944-5875c76514ce\") " pod="openshift-dns/dns-default-64cqc" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.799370 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/2d2365ff-c132-4977-b394-c7bdb5f9c02d-signing-cabundle\") pod \"service-ca-9c57cc56f-brmrn\" (UID: \"2d2365ff-c132-4977-b394-c7bdb5f9c02d\") " pod="openshift-service-ca/service-ca-9c57cc56f-brmrn" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.799397 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.799417 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98xl5\" (UniqueName: \"kubernetes.io/projected/93fe04d3-c430-4ed6-ae04-312cf0613834-kube-api-access-98xl5\") pod \"collect-profiles-29522220-v5twm\" (UID: \"93fe04d3-c430-4ed6-ae04-312cf0613834\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522220-v5twm" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.799435 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a800c1f4-8f07-4d13-a112-e3e4073c0241-proxy-tls\") pod \"machine-config-operator-74547568cd-sv4jj\" (UID: \"a800c1f4-8f07-4d13-a112-e3e4073c0241\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sv4jj" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.799454 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-vfbvk\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.799470 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-vfbvk\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.799490 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/84d8965a-d181-4f61-921f-1802347f6606-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-9c9zp\" (UID: \"84d8965a-d181-4f61-921f-1802347f6606\") " pod="openshift-marketplace/marketplace-operator-79b997595-9c9zp" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.799508 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ce6956ee-96b9-4fea-af68-90540d249ba8-csi-data-dir\") pod \"csi-hostpathplugin-45lhd\" (UID: \"ce6956ee-96b9-4fea-af68-90540d249ba8\") " pod="hostpath-provisioner/csi-hostpathplugin-45lhd" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.799524 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ce6956ee-96b9-4fea-af68-90540d249ba8-registration-dir\") pod \"csi-hostpathplugin-45lhd\" (UID: \"ce6956ee-96b9-4fea-af68-90540d249ba8\") " pod="hostpath-provisioner/csi-hostpathplugin-45lhd" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.799543 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e3687198-a53d-4a80-baec-bafd1c5ef14d-registry-certificates\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.799565 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9vhk\" (UniqueName: \"kubernetes.io/projected/fc45d0df-2f2d-449e-9f66-3d6cb79d1821-kube-api-access-j9vhk\") pod \"etcd-operator-b45778765-wdpjf\" (UID: \"fc45d0df-2f2d-449e-9f66-3d6cb79d1821\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wdpjf" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.799584 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0b5c785-906c-4d00-93b3-36111e5d38da-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-8l455\" (UID: \"e0b5c785-906c-4d00-93b3-36111e5d38da\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8l455" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.799602 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/fc45d0df-2f2d-449e-9f66-3d6cb79d1821-etcd-service-ca\") pod \"etcd-operator-b45778765-wdpjf\" (UID: \"fc45d0df-2f2d-449e-9f66-3d6cb79d1821\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wdpjf" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.799619 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-vfbvk\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.799636 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f86fa65-b079-4331-b597-63a6c9995849-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cb48c\" (UID: \"2f86fa65-b079-4331-b597-63a6c9995849\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cb48c" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.799653 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1546cd3a-91cf-4e78-91e3-46ee4b8fbb43-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-8t8vm\" (UID: \"1546cd3a-91cf-4e78-91e3-46ee4b8fbb43\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8t8vm" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.799670 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnjvv\" (UniqueName: \"kubernetes.io/projected/3a2b80bf-ead8-48e2-95fe-837c9ce39f61-kube-api-access-fnjvv\") pod \"dns-operator-744455d44c-vnwxg\" (UID: \"3a2b80bf-ead8-48e2-95fe-837c9ce39f61\") " pod="openshift-dns-operator/dns-operator-744455d44c-vnwxg" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.799707 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcnnq\" (UniqueName: \"kubernetes.io/projected/a800c1f4-8f07-4d13-a112-e3e4073c0241-kube-api-access-hcnnq\") pod \"machine-config-operator-74547568cd-sv4jj\" (UID: \"a800c1f4-8f07-4d13-a112-e3e4073c0241\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sv4jj" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.799740 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ce6956ee-96b9-4fea-af68-90540d249ba8-socket-dir\") pod \"csi-hostpathplugin-45lhd\" (UID: \"ce6956ee-96b9-4fea-af68-90540d249ba8\") " pod="hostpath-provisioner/csi-hostpathplugin-45lhd" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.799758 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/be377252-9683-4490-960c-300c3c2007c4-apiservice-cert\") pod \"packageserver-d55dfcdfc-z7bll\" (UID: \"be377252-9683-4490-960c-300c3c2007c4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z7bll" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.799798 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-vfbvk\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.799827 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/81e66e5c-cae0-4eb7-87fb-273620ab5331-cert\") pod \"ingress-canary-hs5dk\" (UID: \"81e66e5c-cae0-4eb7-87fb-273620ab5331\") " pod="openshift-ingress-canary/ingress-canary-hs5dk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.799852 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1546cd3a-91cf-4e78-91e3-46ee4b8fbb43-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-8t8vm\" (UID: \"1546cd3a-91cf-4e78-91e3-46ee4b8fbb43\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8t8vm" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.799868 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc45d0df-2f2d-449e-9f66-3d6cb79d1821-serving-cert\") pod \"etcd-operator-b45778765-wdpjf\" (UID: \"fc45d0df-2f2d-449e-9f66-3d6cb79d1821\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wdpjf" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.799885 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdfkn\" (UniqueName: \"kubernetes.io/projected/e0b5c785-906c-4d00-93b3-36111e5d38da-kube-api-access-gdfkn\") pod \"kube-storage-version-migrator-operator-b67b599dd-8l455\" (UID: \"e0b5c785-906c-4d00-93b3-36111e5d38da\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8l455" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.799901 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-vfbvk\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.799918 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whm4d\" (UniqueName: \"kubernetes.io/projected/e3687198-a53d-4a80-baec-bafd1c5ef14d-kube-api-access-whm4d\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.799936 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f86fa65-b079-4331-b597-63a6c9995849-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cb48c\" (UID: \"2f86fa65-b079-4331-b597-63a6c9995849\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cb48c" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.799960 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/8f5feb72-96ca-4e4d-9566-0757c7dd93f9-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-vtd5z\" (UID: \"8f5feb72-96ca-4e4d-9566-0757c7dd93f9\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vtd5z" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.799975 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e8351820-e7b6-4477-9944-5875c76514ce-metrics-tls\") pod \"dns-default-64cqc\" (UID: \"e8351820-e7b6-4477-9944-5875c76514ce\") " pod="openshift-dns/dns-default-64cqc" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.799992 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5ftl\" (UniqueName: \"kubernetes.io/projected/f7b03b61-99a7-4db3-be68-e2ad2149f44b-kube-api-access-g5ftl\") pod \"oauth-openshift-558db77b4-vfbvk\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.800006 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fc45d0df-2f2d-449e-9f66-3d6cb79d1821-etcd-client\") pod \"etcd-operator-b45778765-wdpjf\" (UID: \"fc45d0df-2f2d-449e-9f66-3d6cb79d1821\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wdpjf" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.800025 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a41bab7a-a90f-40a6-ba4f-b48ee6603fb5-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-9tpfm\" (UID: \"a41bab7a-a90f-40a6-ba4f-b48ee6603fb5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tpfm" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.800043 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/93fe04d3-c430-4ed6-ae04-312cf0613834-config-volume\") pod \"collect-profiles-29522220-v5twm\" (UID: \"93fe04d3-c430-4ed6-ae04-312cf0613834\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522220-v5twm" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.800058 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6hp6\" (UniqueName: \"kubernetes.io/projected/2d2365ff-c132-4977-b394-c7bdb5f9c02d-kube-api-access-w6hp6\") pod \"service-ca-9c57cc56f-brmrn\" (UID: \"2d2365ff-c132-4977-b394-c7bdb5f9c02d\") " pod="openshift-service-ca/service-ca-9c57cc56f-brmrn" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.800087 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7jlh\" (UniqueName: \"kubernetes.io/projected/4c087448-28c8-4708-897c-9eaf07323b72-kube-api-access-m7jlh\") pod \"multus-admission-controller-857f4d67dd-6hmm5\" (UID: \"4c087448-28c8-4708-897c-9eaf07323b72\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6hmm5" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.800102 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mshpw\" (UniqueName: \"kubernetes.io/projected/a41bab7a-a90f-40a6-ba4f-b48ee6603fb5-kube-api-access-mshpw\") pod \"machine-config-controller-84d6567774-9tpfm\" (UID: \"a41bab7a-a90f-40a6-ba4f-b48ee6603fb5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tpfm" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.800118 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0b5c785-906c-4d00-93b3-36111e5d38da-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-8l455\" (UID: \"e0b5c785-906c-4d00-93b3-36111e5d38da\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8l455" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.800136 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfggw\" (UniqueName: \"kubernetes.io/projected/e8351820-e7b6-4477-9944-5875c76514ce-kube-api-access-sfggw\") pod \"dns-default-64cqc\" (UID: \"e8351820-e7b6-4477-9944-5875c76514ce\") " pod="openshift-dns/dns-default-64cqc" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.800181 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e3687198-a53d-4a80-baec-bafd1c5ef14d-trusted-ca\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.800198 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1546cd3a-91cf-4e78-91e3-46ee4b8fbb43-config\") pod \"kube-controller-manager-operator-78b949d7b-8t8vm\" (UID: \"1546cd3a-91cf-4e78-91e3-46ee4b8fbb43\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8t8vm" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.800212 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ce6956ee-96b9-4fea-af68-90540d249ba8-mountpoint-dir\") pod \"csi-hostpathplugin-45lhd\" (UID: \"ce6956ee-96b9-4fea-af68-90540d249ba8\") " pod="hostpath-provisioner/csi-hostpathplugin-45lhd" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.801470 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/fc45d0df-2f2d-449e-9f66-3d6cb79d1821-etcd-service-ca\") pod \"etcd-operator-b45778765-wdpjf\" (UID: \"fc45d0df-2f2d-449e-9f66-3d6cb79d1821\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wdpjf" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.802008 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e3687198-a53d-4a80-baec-bafd1c5ef14d-registry-certificates\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.803400 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a800c1f4-8f07-4d13-a112-e3e4073c0241-auth-proxy-config\") pod \"machine-config-operator-74547568cd-sv4jj\" (UID: \"a800c1f4-8f07-4d13-a112-e3e4073c0241\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sv4jj" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.803658 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f7b03b61-99a7-4db3-be68-e2ad2149f44b-audit-policies\") pod \"oauth-openshift-558db77b4-vfbvk\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:44 crc kubenswrapper[4955]: E0217 13:06:44.804279 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:45.304262282 +0000 UTC m=+143.826991815 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.806028 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e3687198-a53d-4a80-baec-bafd1c5ef14d-ca-trust-extracted\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.807028 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-8dx57"] Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.807277 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0b5c785-906c-4d00-93b3-36111e5d38da-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-8l455\" (UID: \"e0b5c785-906c-4d00-93b3-36111e5d38da\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8l455" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.809297 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/84d8965a-d181-4f61-921f-1802347f6606-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-9c9zp\" (UID: \"84d8965a-d181-4f61-921f-1802347f6606\") " pod="openshift-marketplace/marketplace-operator-79b997595-9c9zp" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.809563 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/93fe04d3-c430-4ed6-ae04-312cf0613834-secret-volume\") pod \"collect-profiles-29522220-v5twm\" (UID: \"93fe04d3-c430-4ed6-ae04-312cf0613834\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522220-v5twm" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.809660 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c46d5d3e-92cd-415f-a04d-59756dd57637-srv-cert\") pod \"olm-operator-6b444d44fb-f7ccb\" (UID: \"c46d5d3e-92cd-415f-a04d-59756dd57637\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f7ccb" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.809826 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1546cd3a-91cf-4e78-91e3-46ee4b8fbb43-config\") pod \"kube-controller-manager-operator-78b949d7b-8t8vm\" (UID: \"1546cd3a-91cf-4e78-91e3-46ee4b8fbb43\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8t8vm" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.810118 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4c087448-28c8-4708-897c-9eaf07323b72-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-6hmm5\" (UID: \"4c087448-28c8-4708-897c-9eaf07323b72\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6hmm5" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.810200 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cea310be-8fea-43cf-a000-69b614f260e4-service-ca-bundle\") pod \"router-default-5444994796-m7gfn\" (UID: \"cea310be-8fea-43cf-a000-69b614f260e4\") " pod="openshift-ingress/router-default-5444994796-m7gfn" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.810234 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-vfbvk\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.810252 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-vfbvk\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.810409 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-vfbvk\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.810445 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a41bab7a-a90f-40a6-ba4f-b48ee6603fb5-proxy-tls\") pod \"machine-config-controller-84d6567774-9tpfm\" (UID: \"a41bab7a-a90f-40a6-ba4f-b48ee6603fb5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tpfm" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.810684 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-vfbvk\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.810962 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/89d41921-f56f-4e2e-8ef0-2d981933ad68-profile-collector-cert\") pod \"catalog-operator-68c6474976-4hhxl\" (UID: \"89d41921-f56f-4e2e-8ef0-2d981933ad68\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4hhxl" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.811140 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txj28\" (UniqueName: \"kubernetes.io/projected/91cf9f0e-e498-41af-be13-fec987e4196c-kube-api-access-txj28\") pod \"service-ca-operator-777779d784-zc8xw\" (UID: \"91cf9f0e-e498-41af-be13-fec987e4196c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zc8xw" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.811202 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/784c9c2c-64a9-4887-b52c-0949da50de62-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-qbxwp\" (UID: \"784c9c2c-64a9-4887-b52c-0949da50de62\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qbxwp" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.811223 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cea310be-8fea-43cf-a000-69b614f260e4-metrics-certs\") pod \"router-default-5444994796-m7gfn\" (UID: \"cea310be-8fea-43cf-a000-69b614f260e4\") " pod="openshift-ingress/router-default-5444994796-m7gfn" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.811248 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a800c1f4-8f07-4d13-a112-e3e4073c0241-images\") pod \"machine-config-operator-74547568cd-sv4jj\" (UID: \"a800c1f4-8f07-4d13-a112-e3e4073c0241\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sv4jj" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.811498 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/8259766a-777c-4103-a05b-091adc07ad32-certs\") pod \"machine-config-server-5s2bx\" (UID: \"8259766a-777c-4103-a05b-091adc07ad32\") " pod="openshift-machine-config-operator/machine-config-server-5s2bx" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.811669 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/be377252-9683-4490-960c-300c3c2007c4-webhook-cert\") pod \"packageserver-d55dfcdfc-z7bll\" (UID: \"be377252-9683-4490-960c-300c3c2007c4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z7bll" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.811810 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8gk4\" (UniqueName: \"kubernetes.io/projected/81e66e5c-cae0-4eb7-87fb-273620ab5331-kube-api-access-b8gk4\") pod \"ingress-canary-hs5dk\" (UID: \"81e66e5c-cae0-4eb7-87fb-273620ab5331\") " pod="openshift-ingress-canary/ingress-canary-hs5dk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.811964 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ce6956ee-96b9-4fea-af68-90540d249ba8-plugins-dir\") pod \"csi-hostpathplugin-45lhd\" (UID: \"ce6956ee-96b9-4fea-af68-90540d249ba8\") " pod="hostpath-provisioner/csi-hostpathplugin-45lhd" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.812035 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-vfbvk\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.812104 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc45d0df-2f2d-449e-9f66-3d6cb79d1821-config\") pod \"etcd-operator-b45778765-wdpjf\" (UID: \"fc45d0df-2f2d-449e-9f66-3d6cb79d1821\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wdpjf" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.812172 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsgxb\" (UniqueName: \"kubernetes.io/projected/be377252-9683-4490-960c-300c3c2007c4-kube-api-access-gsgxb\") pod \"packageserver-d55dfcdfc-z7bll\" (UID: \"be377252-9683-4490-960c-300c3c2007c4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z7bll" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.812262 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-vfbvk\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.812340 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2vvx\" (UniqueName: \"kubernetes.io/projected/89d41921-f56f-4e2e-8ef0-2d981933ad68-kube-api-access-s2vvx\") pod \"catalog-operator-68c6474976-4hhxl\" (UID: \"89d41921-f56f-4e2e-8ef0-2d981933ad68\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4hhxl" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.812411 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2r45q\" (UniqueName: \"kubernetes.io/projected/84d8965a-d181-4f61-921f-1802347f6606-kube-api-access-2r45q\") pod \"marketplace-operator-79b997595-9c9zp\" (UID: \"84d8965a-d181-4f61-921f-1802347f6606\") " pod="openshift-marketplace/marketplace-operator-79b997595-9c9zp" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.812482 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f86fa65-b079-4331-b597-63a6c9995849-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cb48c\" (UID: \"2f86fa65-b079-4331-b597-63a6c9995849\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cb48c" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.812549 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/fc45d0df-2f2d-449e-9f66-3d6cb79d1821-etcd-ca\") pod \"etcd-operator-b45778765-wdpjf\" (UID: \"fc45d0df-2f2d-449e-9f66-3d6cb79d1821\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wdpjf" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.812593 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a800c1f4-8f07-4d13-a112-e3e4073c0241-images\") pod \"machine-config-operator-74547568cd-sv4jj\" (UID: \"a800c1f4-8f07-4d13-a112-e3e4073c0241\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sv4jj" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.811675 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e3687198-a53d-4a80-baec-bafd1c5ef14d-trusted-ca\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.812629 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91cf9f0e-e498-41af-be13-fec987e4196c-config\") pod \"service-ca-operator-777779d784-zc8xw\" (UID: \"91cf9f0e-e498-41af-be13-fec987e4196c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zc8xw" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.812799 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/be377252-9683-4490-960c-300c3c2007c4-tmpfs\") pod \"packageserver-d55dfcdfc-z7bll\" (UID: \"be377252-9683-4490-960c-300c3c2007c4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z7bll" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.812839 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc45d0df-2f2d-449e-9f66-3d6cb79d1821-config\") pod \"etcd-operator-b45778765-wdpjf\" (UID: \"fc45d0df-2f2d-449e-9f66-3d6cb79d1821\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wdpjf" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.812943 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c46d5d3e-92cd-415f-a04d-59756dd57637-profile-collector-cert\") pod \"olm-operator-6b444d44fb-f7ccb\" (UID: \"c46d5d3e-92cd-415f-a04d-59756dd57637\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f7ccb" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.812992 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4wkq\" (UniqueName: \"kubernetes.io/projected/cea310be-8fea-43cf-a000-69b614f260e4-kube-api-access-w4wkq\") pod \"router-default-5444994796-m7gfn\" (UID: \"cea310be-8fea-43cf-a000-69b614f260e4\") " pod="openshift-ingress/router-default-5444994796-m7gfn" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.813434 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-vfbvk\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.813515 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/89d41921-f56f-4e2e-8ef0-2d981933ad68-srv-cert\") pod \"catalog-operator-68c6474976-4hhxl\" (UID: \"89d41921-f56f-4e2e-8ef0-2d981933ad68\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4hhxl" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.813574 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/2d2365ff-c132-4977-b394-c7bdb5f9c02d-signing-key\") pod \"service-ca-9c57cc56f-brmrn\" (UID: \"2d2365ff-c132-4977-b394-c7bdb5f9c02d\") " pod="openshift-service-ca/service-ca-9c57cc56f-brmrn" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.813604 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/784c9c2c-64a9-4887-b52c-0949da50de62-config\") pod \"kube-apiserver-operator-766d6c64bb-qbxwp\" (UID: \"784c9c2c-64a9-4887-b52c-0949da50de62\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qbxwp" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.813649 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/cea310be-8fea-43cf-a000-69b614f260e4-stats-auth\") pod \"router-default-5444994796-m7gfn\" (UID: \"cea310be-8fea-43cf-a000-69b614f260e4\") " pod="openshift-ingress/router-default-5444994796-m7gfn" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.813675 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e3687198-a53d-4a80-baec-bafd1c5ef14d-installation-pull-secrets\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.813704 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f86fa65-b079-4331-b597-63a6c9995849-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cb48c\" (UID: \"2f86fa65-b079-4331-b597-63a6c9995849\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cb48c" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.814318 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a800c1f4-8f07-4d13-a112-e3e4073c0241-proxy-tls\") pod \"machine-config-operator-74547568cd-sv4jj\" (UID: \"a800c1f4-8f07-4d13-a112-e3e4073c0241\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sv4jj" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.814596 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/fc45d0df-2f2d-449e-9f66-3d6cb79d1821-etcd-ca\") pod \"etcd-operator-b45778765-wdpjf\" (UID: \"fc45d0df-2f2d-449e-9f66-3d6cb79d1821\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wdpjf" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.814662 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e3687198-a53d-4a80-baec-bafd1c5ef14d-registry-tls\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.814698 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-vfbvk\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.814732 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f7b03b61-99a7-4db3-be68-e2ad2149f44b-audit-dir\") pod \"oauth-openshift-558db77b4-vfbvk\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.814757 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/8259766a-777c-4103-a05b-091adc07ad32-node-bootstrap-token\") pod \"machine-config-server-5s2bx\" (UID: \"8259766a-777c-4103-a05b-091adc07ad32\") " pod="openshift-machine-config-operator/machine-config-server-5s2bx" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.814797 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dw9dp\" (UniqueName: \"kubernetes.io/projected/8259766a-777c-4103-a05b-091adc07ad32-kube-api-access-dw9dp\") pod \"machine-config-server-5s2bx\" (UID: \"8259766a-777c-4103-a05b-091adc07ad32\") " pod="openshift-machine-config-operator/machine-config-server-5s2bx" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.814823 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p785l\" (UniqueName: \"kubernetes.io/projected/ce6956ee-96b9-4fea-af68-90540d249ba8-kube-api-access-p785l\") pod \"csi-hostpathplugin-45lhd\" (UID: \"ce6956ee-96b9-4fea-af68-90540d249ba8\") " pod="hostpath-provisioner/csi-hostpathplugin-45lhd" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.814870 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/784c9c2c-64a9-4887-b52c-0949da50de62-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-qbxwp\" (UID: \"784c9c2c-64a9-4887-b52c-0949da50de62\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qbxwp" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.814889 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/784c9c2c-64a9-4887-b52c-0949da50de62-config\") pod \"kube-apiserver-operator-766d6c64bb-qbxwp\" (UID: \"784c9c2c-64a9-4887-b52c-0949da50de62\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qbxwp" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.814900 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3a2b80bf-ead8-48e2-95fe-837c9ce39f61-metrics-tls\") pod \"dns-operator-744455d44c-vnwxg\" (UID: \"3a2b80bf-ead8-48e2-95fe-837c9ce39f61\") " pod="openshift-dns-operator/dns-operator-744455d44c-vnwxg" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.814972 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e3687198-a53d-4a80-baec-bafd1c5ef14d-bound-sa-token\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.814995 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/91cf9f0e-e498-41af-be13-fec987e4196c-serving-cert\") pod \"service-ca-operator-777779d784-zc8xw\" (UID: \"91cf9f0e-e498-41af-be13-fec987e4196c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zc8xw" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.815216 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f7b03b61-99a7-4db3-be68-e2ad2149f44b-audit-dir\") pod \"oauth-openshift-558db77b4-vfbvk\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.820201 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-vfbvk\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.821016 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-vfbvk\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.821652 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-vfbvk\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.822300 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc45d0df-2f2d-449e-9f66-3d6cb79d1821-serving-cert\") pod \"etcd-operator-b45778765-wdpjf\" (UID: \"fc45d0df-2f2d-449e-9f66-3d6cb79d1821\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wdpjf" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.822875 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-vfbvk\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.823094 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-vfbvk\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.823229 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/8f5feb72-96ca-4e4d-9566-0757c7dd93f9-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-vtd5z\" (UID: \"8f5feb72-96ca-4e4d-9566-0757c7dd93f9\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vtd5z" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.823450 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/784c9c2c-64a9-4887-b52c-0949da50de62-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-qbxwp\" (UID: \"784c9c2c-64a9-4887-b52c-0949da50de62\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qbxwp" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.824411 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f86fa65-b079-4331-b597-63a6c9995849-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cb48c\" (UID: \"2f86fa65-b079-4331-b597-63a6c9995849\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cb48c" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.824462 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0b5c785-906c-4d00-93b3-36111e5d38da-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-8l455\" (UID: \"e0b5c785-906c-4d00-93b3-36111e5d38da\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8l455" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.825248 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/89d41921-f56f-4e2e-8ef0-2d981933ad68-profile-collector-cert\") pod \"catalog-operator-68c6474976-4hhxl\" (UID: \"89d41921-f56f-4e2e-8ef0-2d981933ad68\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4hhxl" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.825269 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e3687198-a53d-4a80-baec-bafd1c5ef14d-installation-pull-secrets\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.825694 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fc45d0df-2f2d-449e-9f66-3d6cb79d1821-etcd-client\") pod \"etcd-operator-b45778765-wdpjf\" (UID: \"fc45d0df-2f2d-449e-9f66-3d6cb79d1821\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wdpjf" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.826254 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/89d41921-f56f-4e2e-8ef0-2d981933ad68-srv-cert\") pod \"catalog-operator-68c6474976-4hhxl\" (UID: \"89d41921-f56f-4e2e-8ef0-2d981933ad68\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4hhxl" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.830231 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdfkn\" (UniqueName: \"kubernetes.io/projected/e0b5c785-906c-4d00-93b3-36111e5d38da-kube-api-access-gdfkn\") pod \"kube-storage-version-migrator-operator-b67b599dd-8l455\" (UID: \"e0b5c785-906c-4d00-93b3-36111e5d38da\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8l455" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.832293 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1546cd3a-91cf-4e78-91e3-46ee4b8fbb43-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-8t8vm\" (UID: \"1546cd3a-91cf-4e78-91e3-46ee4b8fbb43\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8t8vm" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.832709 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-vfbvk\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.833915 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-vfbvk\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.835575 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-vfbvk\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.838118 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vjrvb"] Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.840129 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-vfbvk\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.840462 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e3687198-a53d-4a80-baec-bafd1c5ef14d-registry-tls\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.840688 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3a2b80bf-ead8-48e2-95fe-837c9ce39f61-metrics-tls\") pod \"dns-operator-744455d44c-vnwxg\" (UID: \"3a2b80bf-ead8-48e2-95fe-837c9ce39f61\") " pod="openshift-dns-operator/dns-operator-744455d44c-vnwxg" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.849663 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9vhk\" (UniqueName: \"kubernetes.io/projected/fc45d0df-2f2d-449e-9f66-3d6cb79d1821-kube-api-access-j9vhk\") pod \"etcd-operator-b45778765-wdpjf\" (UID: \"fc45d0df-2f2d-449e-9f66-3d6cb79d1821\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wdpjf" Feb 17 13:06:44 crc kubenswrapper[4955]: W0217 13:06:44.856725 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode690bd42_0ab1_4b3a_802b_4511117b6fe0.slice/crio-c0e9980ef890eb038f5e55419af044fc0ace9c568ae670bb96a43c23a77fb651 WatchSource:0}: Error finding container c0e9980ef890eb038f5e55419af044fc0ace9c568ae670bb96a43c23a77fb651: Status 404 returned error can't find the container with id c0e9980ef890eb038f5e55419af044fc0ace9c568ae670bb96a43c23a77fb651 Feb 17 13:06:44 crc kubenswrapper[4955]: W0217 13:06:44.858294 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod01aa5d18_e942_4a3d_8af5_1e41b503f85a.slice/crio-5f71f61bd80208a6b369ded0de6d8a7cbc12dc7acac4d3cbb0774e9575041ef5 WatchSource:0}: Error finding container 5f71f61bd80208a6b369ded0de6d8a7cbc12dc7acac4d3cbb0774e9575041ef5: Status 404 returned error can't find the container with id 5f71f61bd80208a6b369ded0de6d8a7cbc12dc7acac4d3cbb0774e9575041ef5 Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.883072 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcnnq\" (UniqueName: \"kubernetes.io/projected/a800c1f4-8f07-4d13-a112-e3e4073c0241-kube-api-access-hcnnq\") pod \"machine-config-operator-74547568cd-sv4jj\" (UID: \"a800c1f4-8f07-4d13-a112-e3e4073c0241\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sv4jj" Feb 17 13:06:44 crc kubenswrapper[4955]: W0217 13:06:44.886946 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d7d0f56_00a3_4eae_8ff2_a35e1f20a377.slice/crio-ec4718ef12c797e56430ac067d9659d8cf3629cdcd0b6f18ca27df3fcdccd40d WatchSource:0}: Error finding container ec4718ef12c797e56430ac067d9659d8cf3629cdcd0b6f18ca27df3fcdccd40d: Status 404 returned error can't find the container with id ec4718ef12c797e56430ac067d9659d8cf3629cdcd0b6f18ca27df3fcdccd40d Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.896722 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-wdpjf" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.903373 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t5v8v"] Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.912619 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8l455" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.923280 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.923455 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4c087448-28c8-4708-897c-9eaf07323b72-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-6hmm5\" (UID: \"4c087448-28c8-4708-897c-9eaf07323b72\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6hmm5" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.923481 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cea310be-8fea-43cf-a000-69b614f260e4-service-ca-bundle\") pod \"router-default-5444994796-m7gfn\" (UID: \"cea310be-8fea-43cf-a000-69b614f260e4\") " pod="openshift-ingress/router-default-5444994796-m7gfn" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.923504 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a41bab7a-a90f-40a6-ba4f-b48ee6603fb5-proxy-tls\") pod \"machine-config-controller-84d6567774-9tpfm\" (UID: \"a41bab7a-a90f-40a6-ba4f-b48ee6603fb5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tpfm" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.923522 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txj28\" (UniqueName: \"kubernetes.io/projected/91cf9f0e-e498-41af-be13-fec987e4196c-kube-api-access-txj28\") pod \"service-ca-operator-777779d784-zc8xw\" (UID: \"91cf9f0e-e498-41af-be13-fec987e4196c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zc8xw" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.923545 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cea310be-8fea-43cf-a000-69b614f260e4-metrics-certs\") pod \"router-default-5444994796-m7gfn\" (UID: \"cea310be-8fea-43cf-a000-69b614f260e4\") " pod="openshift-ingress/router-default-5444994796-m7gfn" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.923561 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/8259766a-777c-4103-a05b-091adc07ad32-certs\") pod \"machine-config-server-5s2bx\" (UID: \"8259766a-777c-4103-a05b-091adc07ad32\") " pod="openshift-machine-config-operator/machine-config-server-5s2bx" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.923581 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/be377252-9683-4490-960c-300c3c2007c4-webhook-cert\") pod \"packageserver-d55dfcdfc-z7bll\" (UID: \"be377252-9683-4490-960c-300c3c2007c4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z7bll" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.923600 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8gk4\" (UniqueName: \"kubernetes.io/projected/81e66e5c-cae0-4eb7-87fb-273620ab5331-kube-api-access-b8gk4\") pod \"ingress-canary-hs5dk\" (UID: \"81e66e5c-cae0-4eb7-87fb-273620ab5331\") " pod="openshift-ingress-canary/ingress-canary-hs5dk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.923618 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ce6956ee-96b9-4fea-af68-90540d249ba8-plugins-dir\") pod \"csi-hostpathplugin-45lhd\" (UID: \"ce6956ee-96b9-4fea-af68-90540d249ba8\") " pod="hostpath-provisioner/csi-hostpathplugin-45lhd" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.923634 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsgxb\" (UniqueName: \"kubernetes.io/projected/be377252-9683-4490-960c-300c3c2007c4-kube-api-access-gsgxb\") pod \"packageserver-d55dfcdfc-z7bll\" (UID: \"be377252-9683-4490-960c-300c3c2007c4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z7bll" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.923660 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2r45q\" (UniqueName: \"kubernetes.io/projected/84d8965a-d181-4f61-921f-1802347f6606-kube-api-access-2r45q\") pod \"marketplace-operator-79b997595-9c9zp\" (UID: \"84d8965a-d181-4f61-921f-1802347f6606\") " pod="openshift-marketplace/marketplace-operator-79b997595-9c9zp" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.923679 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91cf9f0e-e498-41af-be13-fec987e4196c-config\") pod \"service-ca-operator-777779d784-zc8xw\" (UID: \"91cf9f0e-e498-41af-be13-fec987e4196c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zc8xw" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.923694 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/be377252-9683-4490-960c-300c3c2007c4-tmpfs\") pod \"packageserver-d55dfcdfc-z7bll\" (UID: \"be377252-9683-4490-960c-300c3c2007c4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z7bll" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.923709 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c46d5d3e-92cd-415f-a04d-59756dd57637-profile-collector-cert\") pod \"olm-operator-6b444d44fb-f7ccb\" (UID: \"c46d5d3e-92cd-415f-a04d-59756dd57637\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f7ccb" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.923723 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4wkq\" (UniqueName: \"kubernetes.io/projected/cea310be-8fea-43cf-a000-69b614f260e4-kube-api-access-w4wkq\") pod \"router-default-5444994796-m7gfn\" (UID: \"cea310be-8fea-43cf-a000-69b614f260e4\") " pod="openshift-ingress/router-default-5444994796-m7gfn" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.923773 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/2d2365ff-c132-4977-b394-c7bdb5f9c02d-signing-key\") pod \"service-ca-9c57cc56f-brmrn\" (UID: \"2d2365ff-c132-4977-b394-c7bdb5f9c02d\") " pod="openshift-service-ca/service-ca-9c57cc56f-brmrn" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.923814 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/cea310be-8fea-43cf-a000-69b614f260e4-stats-auth\") pod \"router-default-5444994796-m7gfn\" (UID: \"cea310be-8fea-43cf-a000-69b614f260e4\") " pod="openshift-ingress/router-default-5444994796-m7gfn" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.923834 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/8259766a-777c-4103-a05b-091adc07ad32-node-bootstrap-token\") pod \"machine-config-server-5s2bx\" (UID: \"8259766a-777c-4103-a05b-091adc07ad32\") " pod="openshift-machine-config-operator/machine-config-server-5s2bx" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.923848 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dw9dp\" (UniqueName: \"kubernetes.io/projected/8259766a-777c-4103-a05b-091adc07ad32-kube-api-access-dw9dp\") pod \"machine-config-server-5s2bx\" (UID: \"8259766a-777c-4103-a05b-091adc07ad32\") " pod="openshift-machine-config-operator/machine-config-server-5s2bx" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.923864 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p785l\" (UniqueName: \"kubernetes.io/projected/ce6956ee-96b9-4fea-af68-90540d249ba8-kube-api-access-p785l\") pod \"csi-hostpathplugin-45lhd\" (UID: \"ce6956ee-96b9-4fea-af68-90540d249ba8\") " pod="hostpath-provisioner/csi-hostpathplugin-45lhd" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.923886 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/91cf9f0e-e498-41af-be13-fec987e4196c-serving-cert\") pod \"service-ca-operator-777779d784-zc8xw\" (UID: \"91cf9f0e-e498-41af-be13-fec987e4196c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zc8xw" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.923903 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccx4h\" (UniqueName: \"kubernetes.io/projected/c46d5d3e-92cd-415f-a04d-59756dd57637-kube-api-access-ccx4h\") pod \"olm-operator-6b444d44fb-f7ccb\" (UID: \"c46d5d3e-92cd-415f-a04d-59756dd57637\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f7ccb" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.923919 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/cea310be-8fea-43cf-a000-69b614f260e4-default-certificate\") pod \"router-default-5444994796-m7gfn\" (UID: \"cea310be-8fea-43cf-a000-69b614f260e4\") " pod="openshift-ingress/router-default-5444994796-m7gfn" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.923952 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e8351820-e7b6-4477-9944-5875c76514ce-config-volume\") pod \"dns-default-64cqc\" (UID: \"e8351820-e7b6-4477-9944-5875c76514ce\") " pod="openshift-dns/dns-default-64cqc" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.923968 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/2d2365ff-c132-4977-b394-c7bdb5f9c02d-signing-cabundle\") pod \"service-ca-9c57cc56f-brmrn\" (UID: \"2d2365ff-c132-4977-b394-c7bdb5f9c02d\") " pod="openshift-service-ca/service-ca-9c57cc56f-brmrn" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.923992 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98xl5\" (UniqueName: \"kubernetes.io/projected/93fe04d3-c430-4ed6-ae04-312cf0613834-kube-api-access-98xl5\") pod \"collect-profiles-29522220-v5twm\" (UID: \"93fe04d3-c430-4ed6-ae04-312cf0613834\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522220-v5twm" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.924009 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/84d8965a-d181-4f61-921f-1802347f6606-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-9c9zp\" (UID: \"84d8965a-d181-4f61-921f-1802347f6606\") " pod="openshift-marketplace/marketplace-operator-79b997595-9c9zp" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.924024 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ce6956ee-96b9-4fea-af68-90540d249ba8-csi-data-dir\") pod \"csi-hostpathplugin-45lhd\" (UID: \"ce6956ee-96b9-4fea-af68-90540d249ba8\") " pod="hostpath-provisioner/csi-hostpathplugin-45lhd" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.924039 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ce6956ee-96b9-4fea-af68-90540d249ba8-registration-dir\") pod \"csi-hostpathplugin-45lhd\" (UID: \"ce6956ee-96b9-4fea-af68-90540d249ba8\") " pod="hostpath-provisioner/csi-hostpathplugin-45lhd" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.924064 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ce6956ee-96b9-4fea-af68-90540d249ba8-socket-dir\") pod \"csi-hostpathplugin-45lhd\" (UID: \"ce6956ee-96b9-4fea-af68-90540d249ba8\") " pod="hostpath-provisioner/csi-hostpathplugin-45lhd" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.924080 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/be377252-9683-4490-960c-300c3c2007c4-apiservice-cert\") pod \"packageserver-d55dfcdfc-z7bll\" (UID: \"be377252-9683-4490-960c-300c3c2007c4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z7bll" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.924106 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/81e66e5c-cae0-4eb7-87fb-273620ab5331-cert\") pod \"ingress-canary-hs5dk\" (UID: \"81e66e5c-cae0-4eb7-87fb-273620ab5331\") " pod="openshift-ingress-canary/ingress-canary-hs5dk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.924150 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e8351820-e7b6-4477-9944-5875c76514ce-metrics-tls\") pod \"dns-default-64cqc\" (UID: \"e8351820-e7b6-4477-9944-5875c76514ce\") " pod="openshift-dns/dns-default-64cqc" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.924172 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a41bab7a-a90f-40a6-ba4f-b48ee6603fb5-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-9tpfm\" (UID: \"a41bab7a-a90f-40a6-ba4f-b48ee6603fb5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tpfm" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.924189 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/93fe04d3-c430-4ed6-ae04-312cf0613834-config-volume\") pod \"collect-profiles-29522220-v5twm\" (UID: \"93fe04d3-c430-4ed6-ae04-312cf0613834\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522220-v5twm" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.924205 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6hp6\" (UniqueName: \"kubernetes.io/projected/2d2365ff-c132-4977-b394-c7bdb5f9c02d-kube-api-access-w6hp6\") pod \"service-ca-9c57cc56f-brmrn\" (UID: \"2d2365ff-c132-4977-b394-c7bdb5f9c02d\") " pod="openshift-service-ca/service-ca-9c57cc56f-brmrn" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.924223 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7jlh\" (UniqueName: \"kubernetes.io/projected/4c087448-28c8-4708-897c-9eaf07323b72-kube-api-access-m7jlh\") pod \"multus-admission-controller-857f4d67dd-6hmm5\" (UID: \"4c087448-28c8-4708-897c-9eaf07323b72\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6hmm5" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.924237 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mshpw\" (UniqueName: \"kubernetes.io/projected/a41bab7a-a90f-40a6-ba4f-b48ee6603fb5-kube-api-access-mshpw\") pod \"machine-config-controller-84d6567774-9tpfm\" (UID: \"a41bab7a-a90f-40a6-ba4f-b48ee6603fb5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tpfm" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.924254 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfggw\" (UniqueName: \"kubernetes.io/projected/e8351820-e7b6-4477-9944-5875c76514ce-kube-api-access-sfggw\") pod \"dns-default-64cqc\" (UID: \"e8351820-e7b6-4477-9944-5875c76514ce\") " pod="openshift-dns/dns-default-64cqc" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.924277 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ce6956ee-96b9-4fea-af68-90540d249ba8-mountpoint-dir\") pod \"csi-hostpathplugin-45lhd\" (UID: \"ce6956ee-96b9-4fea-af68-90540d249ba8\") " pod="hostpath-provisioner/csi-hostpathplugin-45lhd" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.924294 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/84d8965a-d181-4f61-921f-1802347f6606-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-9c9zp\" (UID: \"84d8965a-d181-4f61-921f-1802347f6606\") " pod="openshift-marketplace/marketplace-operator-79b997595-9c9zp" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.924310 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/93fe04d3-c430-4ed6-ae04-312cf0613834-secret-volume\") pod \"collect-profiles-29522220-v5twm\" (UID: \"93fe04d3-c430-4ed6-ae04-312cf0613834\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522220-v5twm" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.924339 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c46d5d3e-92cd-415f-a04d-59756dd57637-srv-cert\") pod \"olm-operator-6b444d44fb-f7ccb\" (UID: \"c46d5d3e-92cd-415f-a04d-59756dd57637\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f7ccb" Feb 17 13:06:44 crc kubenswrapper[4955]: E0217 13:06:44.925212 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:45.425196144 +0000 UTC m=+143.947925677 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.927518 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91cf9f0e-e498-41af-be13-fec987e4196c-config\") pod \"service-ca-operator-777779d784-zc8xw\" (UID: \"91cf9f0e-e498-41af-be13-fec987e4196c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zc8xw" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.928056 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ce6956ee-96b9-4fea-af68-90540d249ba8-registration-dir\") pod \"csi-hostpathplugin-45lhd\" (UID: \"ce6956ee-96b9-4fea-af68-90540d249ba8\") " pod="hostpath-provisioner/csi-hostpathplugin-45lhd" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.928148 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ce6956ee-96b9-4fea-af68-90540d249ba8-csi-data-dir\") pod \"csi-hostpathplugin-45lhd\" (UID: \"ce6956ee-96b9-4fea-af68-90540d249ba8\") " pod="hostpath-provisioner/csi-hostpathplugin-45lhd" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.930692 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/be377252-9683-4490-960c-300c3c2007c4-tmpfs\") pod \"packageserver-d55dfcdfc-z7bll\" (UID: \"be377252-9683-4490-960c-300c3c2007c4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z7bll" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.935570 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/8259766a-777c-4103-a05b-091adc07ad32-node-bootstrap-token\") pod \"machine-config-server-5s2bx\" (UID: \"8259766a-777c-4103-a05b-091adc07ad32\") " pod="openshift-machine-config-operator/machine-config-server-5s2bx" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.936130 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ce6956ee-96b9-4fea-af68-90540d249ba8-socket-dir\") pod \"csi-hostpathplugin-45lhd\" (UID: \"ce6956ee-96b9-4fea-af68-90540d249ba8\") " pod="hostpath-provisioner/csi-hostpathplugin-45lhd" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.937015 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/81e66e5c-cae0-4eb7-87fb-273620ab5331-cert\") pod \"ingress-canary-hs5dk\" (UID: \"81e66e5c-cae0-4eb7-87fb-273620ab5331\") " pod="openshift-ingress-canary/ingress-canary-hs5dk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.937479 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c46d5d3e-92cd-415f-a04d-59756dd57637-profile-collector-cert\") pod \"olm-operator-6b444d44fb-f7ccb\" (UID: \"c46d5d3e-92cd-415f-a04d-59756dd57637\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f7ccb" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.938480 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/84d8965a-d181-4f61-921f-1802347f6606-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-9c9zp\" (UID: \"84d8965a-d181-4f61-921f-1802347f6606\") " pod="openshift-marketplace/marketplace-operator-79b997595-9c9zp" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.938929 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ce6956ee-96b9-4fea-af68-90540d249ba8-mountpoint-dir\") pod \"csi-hostpathplugin-45lhd\" (UID: \"ce6956ee-96b9-4fea-af68-90540d249ba8\") " pod="hostpath-provisioner/csi-hostpathplugin-45lhd" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.939292 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e8351820-e7b6-4477-9944-5875c76514ce-config-volume\") pod \"dns-default-64cqc\" (UID: \"e8351820-e7b6-4477-9944-5875c76514ce\") " pod="openshift-dns/dns-default-64cqc" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.939538 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/93fe04d3-c430-4ed6-ae04-312cf0613834-config-volume\") pod \"collect-profiles-29522220-v5twm\" (UID: \"93fe04d3-c430-4ed6-ae04-312cf0613834\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522220-v5twm" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.940545 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a41bab7a-a90f-40a6-ba4f-b48ee6603fb5-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-9tpfm\" (UID: \"a41bab7a-a90f-40a6-ba4f-b48ee6603fb5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tpfm" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.941253 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/84d8965a-d181-4f61-921f-1802347f6606-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-9c9zp\" (UID: \"84d8965a-d181-4f61-921f-1802347f6606\") " pod="openshift-marketplace/marketplace-operator-79b997595-9c9zp" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.941363 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/2d2365ff-c132-4977-b394-c7bdb5f9c02d-signing-cabundle\") pod \"service-ca-9c57cc56f-brmrn\" (UID: \"2d2365ff-c132-4977-b394-c7bdb5f9c02d\") " pod="openshift-service-ca/service-ca-9c57cc56f-brmrn" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.945302 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/91cf9f0e-e498-41af-be13-fec987e4196c-serving-cert\") pod \"service-ca-operator-777779d784-zc8xw\" (UID: \"91cf9f0e-e498-41af-be13-fec987e4196c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zc8xw" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.946476 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5ftl\" (UniqueName: \"kubernetes.io/projected/f7b03b61-99a7-4db3-be68-e2ad2149f44b-kube-api-access-g5ftl\") pod \"oauth-openshift-558db77b4-vfbvk\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.949599 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9rzm\" (UniqueName: \"kubernetes.io/projected/8f5feb72-96ca-4e4d-9566-0757c7dd93f9-kube-api-access-g9rzm\") pod \"control-plane-machine-set-operator-78cbb6b69f-vtd5z\" (UID: \"8f5feb72-96ca-4e4d-9566-0757c7dd93f9\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vtd5z" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.952884 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/be377252-9683-4490-960c-300c3c2007c4-apiservice-cert\") pod \"packageserver-d55dfcdfc-z7bll\" (UID: \"be377252-9683-4490-960c-300c3c2007c4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z7bll" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.957385 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ce6956ee-96b9-4fea-af68-90540d249ba8-plugins-dir\") pod \"csi-hostpathplugin-45lhd\" (UID: \"ce6956ee-96b9-4fea-af68-90540d249ba8\") " pod="hostpath-provisioner/csi-hostpathplugin-45lhd" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.957389 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/cea310be-8fea-43cf-a000-69b614f260e4-default-certificate\") pod \"router-default-5444994796-m7gfn\" (UID: \"cea310be-8fea-43cf-a000-69b614f260e4\") " pod="openshift-ingress/router-default-5444994796-m7gfn" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.958005 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cea310be-8fea-43cf-a000-69b614f260e4-service-ca-bundle\") pod \"router-default-5444994796-m7gfn\" (UID: \"cea310be-8fea-43cf-a000-69b614f260e4\") " pod="openshift-ingress/router-default-5444994796-m7gfn" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.964656 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sv4jj" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.965992 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/cea310be-8fea-43cf-a000-69b614f260e4-stats-auth\") pod \"router-default-5444994796-m7gfn\" (UID: \"cea310be-8fea-43cf-a000-69b614f260e4\") " pod="openshift-ingress/router-default-5444994796-m7gfn" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.966611 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4c087448-28c8-4708-897c-9eaf07323b72-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-6hmm5\" (UID: \"4c087448-28c8-4708-897c-9eaf07323b72\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6hmm5" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.967184 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/8259766a-777c-4103-a05b-091adc07ad32-certs\") pod \"machine-config-server-5s2bx\" (UID: \"8259766a-777c-4103-a05b-091adc07ad32\") " pod="openshift-machine-config-operator/machine-config-server-5s2bx" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.967202 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnjvv\" (UniqueName: \"kubernetes.io/projected/3a2b80bf-ead8-48e2-95fe-837c9ce39f61-kube-api-access-fnjvv\") pod \"dns-operator-744455d44c-vnwxg\" (UID: \"3a2b80bf-ead8-48e2-95fe-837c9ce39f61\") " pod="openshift-dns-operator/dns-operator-744455d44c-vnwxg" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.967356 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c46d5d3e-92cd-415f-a04d-59756dd57637-srv-cert\") pod \"olm-operator-6b444d44fb-f7ccb\" (UID: \"c46d5d3e-92cd-415f-a04d-59756dd57637\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f7ccb" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.968151 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e8351820-e7b6-4477-9944-5875c76514ce-metrics-tls\") pod \"dns-default-64cqc\" (UID: \"e8351820-e7b6-4477-9944-5875c76514ce\") " pod="openshift-dns/dns-default-64cqc" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.969383 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cea310be-8fea-43cf-a000-69b614f260e4-metrics-certs\") pod \"router-default-5444994796-m7gfn\" (UID: \"cea310be-8fea-43cf-a000-69b614f260e4\") " pod="openshift-ingress/router-default-5444994796-m7gfn" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.969479 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a41bab7a-a90f-40a6-ba4f-b48ee6603fb5-proxy-tls\") pod \"machine-config-controller-84d6567774-9tpfm\" (UID: \"a41bab7a-a90f-40a6-ba4f-b48ee6603fb5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tpfm" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.971132 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/2d2365ff-c132-4977-b394-c7bdb5f9c02d-signing-key\") pod \"service-ca-9c57cc56f-brmrn\" (UID: \"2d2365ff-c132-4977-b394-c7bdb5f9c02d\") " pod="openshift-service-ca/service-ca-9c57cc56f-brmrn" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.976427 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/be377252-9683-4490-960c-300c3c2007c4-webhook-cert\") pod \"packageserver-d55dfcdfc-z7bll\" (UID: \"be377252-9683-4490-960c-300c3c2007c4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z7bll" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.978135 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/93fe04d3-c430-4ed6-ae04-312cf0613834-secret-volume\") pod \"collect-profiles-29522220-v5twm\" (UID: \"93fe04d3-c430-4ed6-ae04-312cf0613834\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522220-v5twm" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.978180 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-qdg4k"] Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.986634 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whm4d\" (UniqueName: \"kubernetes.io/projected/e3687198-a53d-4a80-baec-bafd1c5ef14d-kube-api-access-whm4d\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:44 crc kubenswrapper[4955]: I0217 13:06:44.993288 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1546cd3a-91cf-4e78-91e3-46ee4b8fbb43-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-8t8vm\" (UID: \"1546cd3a-91cf-4e78-91e3-46ee4b8fbb43\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8t8vm" Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.011570 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f86fa65-b079-4331-b597-63a6c9995849-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cb48c\" (UID: \"2f86fa65-b079-4331-b597-63a6c9995849\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cb48c" Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.028323 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:45 crc kubenswrapper[4955]: E0217 13:06:45.029010 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:45.528995523 +0000 UTC m=+144.051725066 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.030527 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/784c9c2c-64a9-4887-b52c-0949da50de62-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-qbxwp\" (UID: \"784c9c2c-64a9-4887-b52c-0949da50de62\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qbxwp" Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.046461 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2vvx\" (UniqueName: \"kubernetes.io/projected/89d41921-f56f-4e2e-8ef0-2d981933ad68-kube-api-access-s2vvx\") pod \"catalog-operator-68c6474976-4hhxl\" (UID: \"89d41921-f56f-4e2e-8ef0-2d981933ad68\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4hhxl" Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.069579 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e3687198-a53d-4a80-baec-bafd1c5ef14d-bound-sa-token\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.091078 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p72m9" event={"ID":"792abe2b-bcce-4af6-828b-7a6726b87976","Type":"ContainerStarted","Data":"0492063cd3ce40a4c7cee85e79ef3b579b10f893eea28c38414681cc4dca2614"} Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.091137 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p72m9" event={"ID":"792abe2b-bcce-4af6-828b-7a6726b87976","Type":"ContainerStarted","Data":"1d0533aca38831c03b8109f678124800d2dbadfedfe6b62c957336ed37541980"} Feb 17 13:06:45 crc kubenswrapper[4955]: W0217 13:06:45.093640 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb3dbb323_561b_4d26_886c_bef115c0dca0.slice/crio-bfe339864977c2077909438ee0723968a61082eb60859359848ae2269ff377f8 WatchSource:0}: Error finding container bfe339864977c2077909438ee0723968a61082eb60859359848ae2269ff377f8: Status 404 returned error can't find the container with id bfe339864977c2077909438ee0723968a61082eb60859359848ae2269ff377f8 Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.097390 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8dx57" event={"ID":"922b208d-fc86-4243-86f1-eacc2797118e","Type":"ContainerStarted","Data":"ad6a3be691fea0f78415bedcc3c98d03dab85a2510df4ab7ab8f980f53a26faa"} Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.115263 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsgxb\" (UniqueName: \"kubernetes.io/projected/be377252-9683-4490-960c-300c3c2007c4-kube-api-access-gsgxb\") pod \"packageserver-d55dfcdfc-z7bll\" (UID: \"be377252-9683-4490-960c-300c3c2007c4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z7bll" Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.129003 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.129039 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4wkq\" (UniqueName: \"kubernetes.io/projected/cea310be-8fea-43cf-a000-69b614f260e4-kube-api-access-w4wkq\") pod \"router-default-5444994796-m7gfn\" (UID: \"cea310be-8fea-43cf-a000-69b614f260e4\") " pod="openshift-ingress/router-default-5444994796-m7gfn" Feb 17 13:06:45 crc kubenswrapper[4955]: E0217 13:06:45.129845 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:45.629828126 +0000 UTC m=+144.152557669 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.136760 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5264p"] Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.137525 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-fpwvv" event={"ID":"6b4e02e4-f80e-44a5-a2a7-f75feaf2c1ce","Type":"ContainerStarted","Data":"6fc682caa7b70e84ed70075fb04ce351f93ecdc5f275c19b7970a2414275fa26"} Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.137555 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-fpwvv" event={"ID":"6b4e02e4-f80e-44a5-a2a7-f75feaf2c1ce","Type":"ContainerStarted","Data":"31e7b0d77f918dead130ab47d2aef16ee53da98f0516f80b1462f65e3858070e"} Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.139303 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-nskw2" event={"ID":"9d7d0f56-00a3-4eae-8ff2-a35e1f20a377","Type":"ContainerStarted","Data":"ec4718ef12c797e56430ac067d9659d8cf3629cdcd0b6f18ca27df3fcdccd40d"} Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.140349 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-qz68q" event={"ID":"01aa5d18-e942-4a3d-8af5-1e41b503f85a","Type":"ContainerStarted","Data":"5f71f61bd80208a6b369ded0de6d8a7cbc12dc7acac4d3cbb0774e9575041ef5"} Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.143009 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vjrvb" event={"ID":"93264a11-973b-435d-a51f-999be787ed0a","Type":"ContainerStarted","Data":"34ef9a5d247bc988ef13bbd9cbb1cbe578fc834e751319fd98e0d1c00eb2dc23"} Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.146807 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccx4h\" (UniqueName: \"kubernetes.io/projected/c46d5d3e-92cd-415f-a04d-59756dd57637-kube-api-access-ccx4h\") pod \"olm-operator-6b444d44fb-f7ccb\" (UID: \"c46d5d3e-92cd-415f-a04d-59756dd57637\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f7ccb" Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.162241 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-4xvzn"] Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.162628 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-d5sv6" event={"ID":"847ef69d-8cd1-4831-8250-05f2cf789d88","Type":"ContainerStarted","Data":"35a853c0936f3f4a3f128582bf179119bc4dafec1d1e562291c47da503550123"} Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.162677 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-d5sv6" event={"ID":"847ef69d-8cd1-4831-8250-05f2cf789d88","Type":"ContainerStarted","Data":"8c94ccba8fcd945dbbbf07ec2101aec27738e56304d739f18816c3241cd2b3ad"} Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.170593 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2r45q\" (UniqueName: \"kubernetes.io/projected/84d8965a-d181-4f61-921f-1802347f6606-kube-api-access-2r45q\") pod \"marketplace-operator-79b997595-9c9zp\" (UID: \"84d8965a-d181-4f61-921f-1802347f6606\") " pod="openshift-marketplace/marketplace-operator-79b997595-9c9zp" Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.176427 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.176907 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-982dt" event={"ID":"de868b10-12d3-4ec1-87d6-f4fcb4994aa3","Type":"ContainerStarted","Data":"ae9b414125f54c6d7fc42cc330fd282c622b531c5645b2a163d35c3ded89b570"} Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.176942 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-982dt" event={"ID":"de868b10-12d3-4ec1-87d6-f4fcb4994aa3","Type":"ContainerStarted","Data":"39369ed3022aa760d0382b697b725ed673a837a273b1932405f1b71961123c9b"} Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.180667 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-npb7n" event={"ID":"e690bd42-0ab1-4b3a-802b-4511117b6fe0","Type":"ContainerStarted","Data":"c0e9980ef890eb038f5e55419af044fc0ace9c568ae670bb96a43c23a77fb651"} Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.180978 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qbxwp" Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.186052 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-trrhn" event={"ID":"32ea5b76-e8e5-4906-bc55-99d5ec780ff2","Type":"ContainerStarted","Data":"08906d556aa78f139c86352c20de5ae166084c72fcef01bc0329fd92a9e4948a"} Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.187945 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8t8vm" Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.196489 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7jlh\" (UniqueName: \"kubernetes.io/projected/4c087448-28c8-4708-897c-9eaf07323b72-kube-api-access-m7jlh\") pod \"multus-admission-controller-857f4d67dd-6hmm5\" (UID: \"4c087448-28c8-4708-897c-9eaf07323b72\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6hmm5" Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.208734 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6hp6\" (UniqueName: \"kubernetes.io/projected/2d2365ff-c132-4977-b394-c7bdb5f9c02d-kube-api-access-w6hp6\") pod \"service-ca-9c57cc56f-brmrn\" (UID: \"2d2365ff-c132-4977-b394-c7bdb5f9c02d\") " pod="openshift-service-ca/service-ca-9c57cc56f-brmrn" Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.216369 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cb48c" Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.222606 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4hhxl" Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.232196 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:45 crc kubenswrapper[4955]: E0217 13:06:45.232762 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:45.732739061 +0000 UTC m=+144.255468614 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.234323 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mshpw\" (UniqueName: \"kubernetes.io/projected/a41bab7a-a90f-40a6-ba4f-b48ee6603fb5-kube-api-access-mshpw\") pod \"machine-config-controller-84d6567774-9tpfm\" (UID: \"a41bab7a-a90f-40a6-ba4f-b48ee6603fb5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tpfm" Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.238380 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mvfqt" event={"ID":"26690efc-79a1-4bc4-bf0a-78b46fb95223","Type":"ContainerStarted","Data":"f6c39053562c7b592965e208d4d5b6f8970bab0ede6b2dcd3ad6cc11a7c41e6c"} Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.238430 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mvfqt" event={"ID":"26690efc-79a1-4bc4-bf0a-78b46fb95223","Type":"ContainerStarted","Data":"e36144cdc20c3e914968c98e53a8a625135126c3bfe48945bcb55a2e260ad8e5"} Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.245457 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-vnwxg" Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.247078 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5x7zc" event={"ID":"58bab2f1-e131-46cb-b705-6bee97eb8452","Type":"ContainerStarted","Data":"0d4c8764bd6735cee843dc34a3a2304b432cccd7e81a6b1df4a329eb8550cfa3"} Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.250937 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfggw\" (UniqueName: \"kubernetes.io/projected/e8351820-e7b6-4477-9944-5875c76514ce-kube-api-access-sfggw\") pod \"dns-default-64cqc\" (UID: \"e8351820-e7b6-4477-9944-5875c76514ce\") " pod="openshift-dns/dns-default-64cqc" Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.257689 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vtd5z" Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.278102 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-m7gfn" Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.278432 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98xl5\" (UniqueName: \"kubernetes.io/projected/93fe04d3-c430-4ed6-ae04-312cf0613834-kube-api-access-98xl5\") pod \"collect-profiles-29522220-v5twm\" (UID: \"93fe04d3-c430-4ed6-ae04-312cf0613834\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522220-v5twm" Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.289959 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-wdpjf"] Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.290029 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f7ccb" Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.293025 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-prtk8" event={"ID":"7db0eeb4-4ffe-4375-a854-0f47efce6a26","Type":"ContainerStarted","Data":"fe421b07e4e9c29655c274a8cefc6a5ff5d1505159119a1ac18382323bc08690"} Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.293087 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-prtk8" event={"ID":"7db0eeb4-4ffe-4375-a854-0f47efce6a26","Type":"ContainerStarted","Data":"b56a5c634f4e287c54fee13d4c6d793ec5b8e65999090f699d2f6746bc6a79bb"} Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.294198 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-prtk8" Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.296992 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8gk4\" (UniqueName: \"kubernetes.io/projected/81e66e5c-cae0-4eb7-87fb-273620ab5331-kube-api-access-b8gk4\") pod \"ingress-canary-hs5dk\" (UID: \"81e66e5c-cae0-4eb7-87fb-273620ab5331\") " pod="openshift-ingress-canary/ingress-canary-hs5dk" Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.299579 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tpfm" Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.300027 4955 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-prtk8 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.300063 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-prtk8" podUID="7db0eeb4-4ffe-4375-a854-0f47efce6a26" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.300195 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-rxsb4" event={"ID":"44192b09-31c7-4e49-b472-5ff0a67d30f5","Type":"ContainerStarted","Data":"cea6c94d0ab43cd6fd80cfd49bf75e78f7b410465779ba38cdfac64107c387a4"} Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.300235 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-rxsb4" event={"ID":"44192b09-31c7-4e49-b472-5ff0a67d30f5","Type":"ContainerStarted","Data":"1401cc5130f7a83a135b176df21f718ed3f9da1d2ca49816cb4fe1ecf0d43f98"} Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.300618 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-rxsb4" Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.306346 4955 patch_prober.go:28] interesting pod/console-operator-58897d9998-rxsb4 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/readyz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.306404 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-rxsb4" podUID="44192b09-31c7-4e49-b472-5ff0a67d30f5" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/readyz\": dial tcp 10.217.0.26:8443: connect: connection refused" Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.308892 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-brmrn" Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.311048 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p785l\" (UniqueName: \"kubernetes.io/projected/ce6956ee-96b9-4fea-af68-90540d249ba8-kube-api-access-p785l\") pod \"csi-hostpathplugin-45lhd\" (UID: \"ce6956ee-96b9-4fea-af68-90540d249ba8\") " pod="hostpath-provisioner/csi-hostpathplugin-45lhd" Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.315209 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29522220-v5twm" Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.327471 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txj28\" (UniqueName: \"kubernetes.io/projected/91cf9f0e-e498-41af-be13-fec987e4196c-kube-api-access-txj28\") pod \"service-ca-operator-777779d784-zc8xw\" (UID: \"91cf9f0e-e498-41af-be13-fec987e4196c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zc8xw" Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.330952 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-9c9zp" Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.333221 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:45 crc kubenswrapper[4955]: E0217 13:06:45.334200 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:45.834179111 +0000 UTC m=+144.356908654 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.342976 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-6hmm5" Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.351741 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dw9dp\" (UniqueName: \"kubernetes.io/projected/8259766a-777c-4103-a05b-091adc07ad32-kube-api-access-dw9dp\") pod \"machine-config-server-5s2bx\" (UID: \"8259766a-777c-4103-a05b-091adc07ad32\") " pod="openshift-machine-config-operator/machine-config-server-5s2bx" Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.366480 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-64cqc" Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.369317 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-5s2bx" Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.369473 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8l455"] Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.373993 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z7bll" Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.389039 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-45lhd" Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.396565 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-hs5dk" Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.434524 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:45 crc kubenswrapper[4955]: E0217 13:06:45.435022 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:45.934997364 +0000 UTC m=+144.457726907 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:45 crc kubenswrapper[4955]: W0217 13:06:45.487030 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode0b5c785_906c_4d00_93b3_36111e5d38da.slice/crio-ddaa12a61b3c374fb763ae5227e660613ae01bad26aff08b659ec5869c6624f5 WatchSource:0}: Error finding container ddaa12a61b3c374fb763ae5227e660613ae01bad26aff08b659ec5869c6624f5: Status 404 returned error can't find the container with id ddaa12a61b3c374fb763ae5227e660613ae01bad26aff08b659ec5869c6624f5 Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.535341 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:45 crc kubenswrapper[4955]: E0217 13:06:45.535529 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:46.035493476 +0000 UTC m=+144.558223019 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.535686 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:45 crc kubenswrapper[4955]: E0217 13:06:45.536330 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:46.036321651 +0000 UTC m=+144.559051194 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.590646 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-zc8xw" Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.632237 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-vfbvk"] Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.637235 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:45 crc kubenswrapper[4955]: E0217 13:06:45.637528 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:46.137480263 +0000 UTC m=+144.660209806 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.637754 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:45 crc kubenswrapper[4955]: E0217 13:06:45.638219 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:46.13805985 +0000 UTC m=+144.660789393 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.652153 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-sv4jj"] Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.713116 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qbxwp"] Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.738814 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:45 crc kubenswrapper[4955]: E0217 13:06:45.738926 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:46.238906543 +0000 UTC m=+144.761636086 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.739361 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:45 crc kubenswrapper[4955]: E0217 13:06:45.739883 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:46.239870962 +0000 UTC m=+144.762600505 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.840264 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:45 crc kubenswrapper[4955]: E0217 13:06:45.840569 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:46.34053273 +0000 UTC m=+144.863262283 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.942299 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:45 crc kubenswrapper[4955]: E0217 13:06:45.942708 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:46.442692292 +0000 UTC m=+144.965421835 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.972093 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-6hmm5"] Feb 17 13:06:45 crc kubenswrapper[4955]: I0217 13:06:45.987002 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8t8vm"] Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.004386 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cb48c"] Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.043976 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:46 crc kubenswrapper[4955]: E0217 13:06:46.044862 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:46.544814023 +0000 UTC m=+145.067543566 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.050240 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:46 crc kubenswrapper[4955]: E0217 13:06:46.050996 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:46.550980197 +0000 UTC m=+145.073709750 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.093318 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4hhxl"] Feb 17 13:06:46 crc kubenswrapper[4955]: W0217 13:06:46.100035 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1546cd3a_91cf_4e78_91e3_46ee4b8fbb43.slice/crio-108fcd14da98367ca71fa4b5f50ac479059cd0e7f2f5e3559dcf3a1bdebc1e04 WatchSource:0}: Error finding container 108fcd14da98367ca71fa4b5f50ac479059cd0e7f2f5e3559dcf3a1bdebc1e04: Status 404 returned error can't find the container with id 108fcd14da98367ca71fa4b5f50ac479059cd0e7f2f5e3559dcf3a1bdebc1e04 Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.141491 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-brmrn"] Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.152892 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.161595 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-vnwxg"] Feb 17 13:06:46 crc kubenswrapper[4955]: E0217 13:06:46.169572 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:46.669527897 +0000 UTC m=+145.192257450 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.221335 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-9tpfm"] Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.255082 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:46 crc kubenswrapper[4955]: E0217 13:06:46.255492 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:46.755472226 +0000 UTC m=+145.278201769 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.261228 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-45lhd"] Feb 17 13:06:46 crc kubenswrapper[4955]: W0217 13:06:46.261338 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda41bab7a_a90f_40a6_ba4f_b48ee6603fb5.slice/crio-b6997bbe0d128c808e280333092207755e4e9ef7a4757b753fd104f0a771f51e WatchSource:0}: Error finding container b6997bbe0d128c808e280333092207755e4e9ef7a4757b753fd104f0a771f51e: Status 404 returned error can't find the container with id b6997bbe0d128c808e280333092207755e4e9ef7a4757b753fd104f0a771f51e Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.328153 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qdg4k" event={"ID":"b3dbb323-561b-4d26-886c-bef115c0dca0","Type":"ContainerStarted","Data":"3dbbe5c1610e3efab7fc4476f5c6ff2a556d70f3df2e3b3c2abbdebfa127e244"} Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.328204 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qdg4k" event={"ID":"b3dbb323-561b-4d26-886c-bef115c0dca0","Type":"ContainerStarted","Data":"bfe339864977c2077909438ee0723968a61082eb60859359848ae2269ff377f8"} Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.331444 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8t8vm" event={"ID":"1546cd3a-91cf-4e78-91e3-46ee4b8fbb43","Type":"ContainerStarted","Data":"108fcd14da98367ca71fa4b5f50ac479059cd0e7f2f5e3559dcf3a1bdebc1e04"} Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.333725 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-trrhn" event={"ID":"32ea5b76-e8e5-4906-bc55-99d5ec780ff2","Type":"ContainerStarted","Data":"cc3917f4c922095ef859cc70871d72ee960aae7344cb604f284832b5d076b35a"} Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.333925 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-trrhn" Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.335717 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-qz68q" event={"ID":"01aa5d18-e942-4a3d-8af5-1e41b503f85a","Type":"ContainerStarted","Data":"b6bc7e1ec3d5cf7078643e21a9c5d078fe43ad7194d6a31f74e705b2f82998cf"} Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.336576 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-qz68q" Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.345503 4955 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-trrhn container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.345546 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-trrhn" podUID="32ea5b76-e8e5-4906-bc55-99d5ec780ff2" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.345627 4955 patch_prober.go:28] interesting pod/downloads-7954f5f757-qz68q container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.345642 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-qz68q" podUID="01aa5d18-e942-4a3d-8af5-1e41b503f85a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.37:8080/\": dial tcp 10.217.0.37:8080: connect: connection refused" Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.346307 4955 generic.go:334] "Generic (PLEG): container finished" podID="9d7d0f56-00a3-4eae-8ff2-a35e1f20a377" containerID="71de3751fdb71044535daa2ace50c4d6caf2d25d5d3285f218f0c26039c6ad5c" exitCode=0 Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.346370 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-nskw2" event={"ID":"9d7d0f56-00a3-4eae-8ff2-a35e1f20a377","Type":"ContainerDied","Data":"71de3751fdb71044535daa2ace50c4d6caf2d25d5d3285f218f0c26039c6ad5c"} Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.355617 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-brmrn" event={"ID":"2d2365ff-c132-4977-b394-c7bdb5f9c02d","Type":"ContainerStarted","Data":"8589bf019b6abc9a97358bb58fdeb3f9a10fb9318877992cfd18e86726a56d4f"} Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.355902 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:46 crc kubenswrapper[4955]: E0217 13:06:46.357641 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:46.857618398 +0000 UTC m=+145.380347941 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.363062 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-m7gfn" event={"ID":"cea310be-8fea-43cf-a000-69b614f260e4","Type":"ContainerStarted","Data":"6857bbaf67886608a729b0c11ff9c17f6778b7d0dce9e6ca0e8dfd0719f86b87"} Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.389998 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4xvzn" event={"ID":"ff950861-7bca-4bfa-9857-cd906f67d5cb","Type":"ContainerStarted","Data":"0985e8445d66fc9274bc0bac879dddeb559b346e7a8fe1d75c005cfeb2a0770e"} Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.390046 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4xvzn" event={"ID":"ff950861-7bca-4bfa-9857-cd906f67d5cb","Type":"ContainerStarted","Data":"0b985a073dde3b8c1267c5a7f3fbef27bd2a165e1db82899c22e4287b9c82c24"} Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.416416 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-982dt" podStartSLOduration=123.416396268 podStartE2EDuration="2m3.416396268s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:46.415049918 +0000 UTC m=+144.937779461" watchObservedRunningTime="2026-02-17 13:06:46.416396268 +0000 UTC m=+144.939125811" Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.422054 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p72m9" podStartSLOduration=123.422041476 podStartE2EDuration="2m3.422041476s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:46.358993509 +0000 UTC m=+144.881723062" watchObservedRunningTime="2026-02-17 13:06:46.422041476 +0000 UTC m=+144.944771019" Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.429002 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-vnwxg" event={"ID":"3a2b80bf-ead8-48e2-95fe-837c9ce39f61","Type":"ContainerStarted","Data":"289b083885b97c7c82e0ee34661e16f81c0cce0c0ae8a5c28e49bb49c0c252fb"} Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.436112 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cb48c" event={"ID":"2f86fa65-b079-4331-b597-63a6c9995849","Type":"ContainerStarted","Data":"21017aebe250f2c2ab2aadacbb38f8f141e9f49c190a00b70baa8ec3d522676d"} Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.442803 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-npb7n" event={"ID":"e690bd42-0ab1-4b3a-802b-4511117b6fe0","Type":"ContainerStarted","Data":"541ff6f58738d8ceca7382578b38728152686198ff183271bc41d9f2d35e933f"} Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.447181 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-5s2bx" event={"ID":"8259766a-777c-4103-a05b-091adc07ad32","Type":"ContainerStarted","Data":"afe962df6bb9aebae30297a6907d7bfd6326134774e05066943ebf044e0a997a"} Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.453240 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-6hmm5" event={"ID":"4c087448-28c8-4708-897c-9eaf07323b72","Type":"ContainerStarted","Data":"1521bf4103d4842b291d153c2b84d868352d56c6e797465a3997b03af28fdc9d"} Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.458246 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:46 crc kubenswrapper[4955]: E0217 13:06:46.461255 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:46.961231813 +0000 UTC m=+145.483961426 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.469845 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t5v8v" event={"ID":"f84f960e-7741-4d13-bd6f-ec2072cd4d14","Type":"ContainerStarted","Data":"269231c2ddf7af8bec09f0ab0803e4d3de23290421e3311e9d070f4354e4c8e5"} Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.470748 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t5v8v" event={"ID":"f84f960e-7741-4d13-bd6f-ec2072cd4d14","Type":"ContainerStarted","Data":"58f547fcd5380cd47af06c63ca8998be58ecd1979f7fda40d2ce424151dc31e0"} Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.479596 4955 generic.go:334] "Generic (PLEG): container finished" podID="922b208d-fc86-4243-86f1-eacc2797118e" containerID="76d9a8d7245f85a36c78dd445d51c26e245bc8c83938eb470d49bd0c677362c1" exitCode=0 Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.479658 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8dx57" event={"ID":"922b208d-fc86-4243-86f1-eacc2797118e","Type":"ContainerDied","Data":"76d9a8d7245f85a36c78dd445d51c26e245bc8c83938eb470d49bd0c677362c1"} Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.485757 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-wdpjf" event={"ID":"fc45d0df-2f2d-449e-9f66-3d6cb79d1821","Type":"ContainerStarted","Data":"de72bfbfae1442cc841f0d2ebe7af769510a9e0ad28c2e06cd2e8a95bcbff822"} Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.486064 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-wdpjf" event={"ID":"fc45d0df-2f2d-449e-9f66-3d6cb79d1821","Type":"ContainerStarted","Data":"dfb30f49010935fda8547ea92715b6f3f6dba36b8fab7dae4fca819252d511a2"} Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.487340 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" event={"ID":"f7b03b61-99a7-4db3-be68-e2ad2149f44b","Type":"ContainerStarted","Data":"200b6ace981980dbd7db743ebc8b8bf7b7ddc033d8a0d9365e8fe8c1b91fb251"} Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.497463 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-zc8xw"] Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.501396 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-45lhd" event={"ID":"ce6956ee-96b9-4fea-af68-90540d249ba8","Type":"ContainerStarted","Data":"aca798cf2b3be2b6482d80db5c96b48f35da1bde9f499d69f581814fb4681c02"} Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.501752 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vtd5z"] Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.522766 4955 generic.go:334] "Generic (PLEG): container finished" podID="58bab2f1-e131-46cb-b705-6bee97eb8452" containerID="12cc66270ed21192ddb4b82571c283a6c5dc9df947353ee5adf2c41f06ebb868" exitCode=0 Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.522934 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5x7zc" event={"ID":"58bab2f1-e131-46cb-b705-6bee97eb8452","Type":"ContainerDied","Data":"12cc66270ed21192ddb4b82571c283a6c5dc9df947353ee5adf2c41f06ebb868"} Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.534638 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8l455" event={"ID":"e0b5c785-906c-4d00-93b3-36111e5d38da","Type":"ContainerStarted","Data":"2f8127d1a40bfa4763db666d3cee9626e7937352c5d8096f2acba971fa988c64"} Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.534688 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8l455" event={"ID":"e0b5c785-906c-4d00-93b3-36111e5d38da","Type":"ContainerStarted","Data":"ddaa12a61b3c374fb763ae5227e660613ae01bad26aff08b659ec5869c6624f5"} Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.554826 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9c9zp"] Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.559152 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:46 crc kubenswrapper[4955]: E0217 13:06:46.559412 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:47.059394226 +0000 UTC m=+145.582123769 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.561327 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-64cqc"] Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.561438 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:46 crc kubenswrapper[4955]: E0217 13:06:46.562625 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:47.062603532 +0000 UTC m=+145.585333075 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.567083 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f7ccb"] Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.567598 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z7bll"] Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.572688 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sv4jj" event={"ID":"a800c1f4-8f07-4d13-a112-e3e4073c0241","Type":"ContainerStarted","Data":"c5f8f7a4abd48cdcf7ebf744e1efcc9df2bad8f3ca83bad651ca817a1bf9767c"} Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.572745 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sv4jj" event={"ID":"a800c1f4-8f07-4d13-a112-e3e4073c0241","Type":"ContainerStarted","Data":"4f0bc27bb83453692916bb855522bd62d3cbce19f901db0fed3fcca1e3bf7d60"} Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.574543 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4hhxl" event={"ID":"89d41921-f56f-4e2e-8ef0-2d981933ad68","Type":"ContainerStarted","Data":"f8944e02ff3b1794a3e949c17d308578d7160de8103293335d49669a0a2170c0"} Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.577912 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tpfm" event={"ID":"a41bab7a-a90f-40a6-ba4f-b48ee6603fb5","Type":"ContainerStarted","Data":"b6997bbe0d128c808e280333092207755e4e9ef7a4757b753fd104f0a771f51e"} Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.586891 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vjrvb" event={"ID":"93264a11-973b-435d-a51f-999be787ed0a","Type":"ContainerStarted","Data":"e398af0ba22a7e4dcdd2d921048f5d32f481cd18e7e14cde02a9bb6fc42ed827"} Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.597919 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qbxwp" event={"ID":"784c9c2c-64a9-4887-b52c-0949da50de62","Type":"ContainerStarted","Data":"529f946ab5f6a80248eafaf4b7738fd1d0df749368cb2b453e0c087b0e5269c4"} Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.608757 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5264p" event={"ID":"3955ee14-c2d4-434e-a9b5-baf95abcbd02","Type":"ContainerStarted","Data":"6617b8c52748853761da4dbd09339e86de28361751026a8c9e8f12a940a3503c"} Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.608839 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5264p" event={"ID":"3955ee14-c2d4-434e-a9b5-baf95abcbd02","Type":"ContainerStarted","Data":"c94819836b1a578f9277b1ccbc094ca14347813c2d1e23fed357e025a150e338"} Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.622628 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-fpwvv" podStartSLOduration=123.622596188 podStartE2EDuration="2m3.622596188s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:46.591188103 +0000 UTC m=+145.113917646" watchObservedRunningTime="2026-02-17 13:06:46.622596188 +0000 UTC m=+145.145325731" Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.634468 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29522220-v5twm"] Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.638331 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-rxsb4" podStartSLOduration=123.638311106 podStartE2EDuration="2m3.638311106s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:46.627761722 +0000 UTC m=+145.150491265" watchObservedRunningTime="2026-02-17 13:06:46.638311106 +0000 UTC m=+145.161040649" Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.664432 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:46 crc kubenswrapper[4955]: E0217 13:06:46.665765 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:47.165727133 +0000 UTC m=+145.688456676 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.669523 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-hs5dk"] Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.687285 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mvfqt" event={"ID":"26690efc-79a1-4bc4-bf0a-78b46fb95223","Type":"ContainerStarted","Data":"1452c1c0699233d47d52e87b307f8f0aea8a3ab7ba7c22449be2ccc5ccc90657"} Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.700332 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-d5sv6" event={"ID":"847ef69d-8cd1-4831-8250-05f2cf789d88","Type":"ContainerStarted","Data":"567c52dad664e0475626e91d4f17b347365da1149922267c41830bc852976e30"} Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.701089 4955 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-prtk8 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.701132 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-prtk8" podUID="7db0eeb4-4ffe-4375-a854-0f47efce6a26" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Feb 17 13:06:46 crc kubenswrapper[4955]: W0217 13:06:46.701636 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbe377252_9683_4490_960c_300c3c2007c4.slice/crio-5437ec716ecc13fe67695e1ff8169699de5a57e09cb46c2e353a502415ea405c WatchSource:0}: Error finding container 5437ec716ecc13fe67695e1ff8169699de5a57e09cb46c2e353a502415ea405c: Status 404 returned error can't find the container with id 5437ec716ecc13fe67695e1ff8169699de5a57e09cb46c2e353a502415ea405c Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.776106 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:46 crc kubenswrapper[4955]: E0217 13:06:46.777463 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:47.27744827 +0000 UTC m=+145.800177813 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.867639 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-prtk8" podStartSLOduration=123.867616645 podStartE2EDuration="2m3.867616645s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:46.865390778 +0000 UTC m=+145.388120341" watchObservedRunningTime="2026-02-17 13:06:46.867616645 +0000 UTC m=+145.390346178" Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.876943 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:46 crc kubenswrapper[4955]: E0217 13:06:46.877141 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:47.377117658 +0000 UTC m=+145.899847201 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.877256 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:46 crc kubenswrapper[4955]: E0217 13:06:46.877596 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:47.377586301 +0000 UTC m=+145.900315834 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:46 crc kubenswrapper[4955]: I0217 13:06:46.978642 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:46 crc kubenswrapper[4955]: E0217 13:06:46.979091 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:47.479071364 +0000 UTC m=+146.001800907 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.063908 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mvfqt" podStartSLOduration=125.063882949 podStartE2EDuration="2m5.063882949s" podCreationTimestamp="2026-02-17 13:04:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:47.061935561 +0000 UTC m=+145.584665124" watchObservedRunningTime="2026-02-17 13:06:47.063882949 +0000 UTC m=+145.586612492" Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.080426 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:47 crc kubenswrapper[4955]: E0217 13:06:47.080979 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:47.580962108 +0000 UTC m=+146.103691651 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.136415 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-d5sv6" podStartSLOduration=124.136384528 podStartE2EDuration="2m4.136384528s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:47.111953811 +0000 UTC m=+145.634683354" watchObservedRunningTime="2026-02-17 13:06:47.136384528 +0000 UTC m=+145.659114081" Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.146717 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-npb7n" podStartSLOduration=124.146691215 podStartE2EDuration="2m4.146691215s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:47.145344375 +0000 UTC m=+145.668073958" watchObservedRunningTime="2026-02-17 13:06:47.146691215 +0000 UTC m=+145.669420758" Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.150857 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-rxsb4" Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.184499 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.184507 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-trrhn" podStartSLOduration=123.18448701 podStartE2EDuration="2m3.18448701s" podCreationTimestamp="2026-02-17 13:04:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:47.184035337 +0000 UTC m=+145.706764890" watchObservedRunningTime="2026-02-17 13:06:47.18448701 +0000 UTC m=+145.707216553" Feb 17 13:06:47 crc kubenswrapper[4955]: E0217 13:06:47.184912 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:47.684893143 +0000 UTC m=+146.207622686 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.285846 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:47 crc kubenswrapper[4955]: E0217 13:06:47.286294 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:47.786279042 +0000 UTC m=+146.309008585 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.335605 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vjrvb" podStartSLOduration=124.33557949 podStartE2EDuration="2m4.33557949s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:47.328350984 +0000 UTC m=+145.851080547" watchObservedRunningTime="2026-02-17 13:06:47.33557949 +0000 UTC m=+145.858309033" Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.389672 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:47 crc kubenswrapper[4955]: E0217 13:06:47.390153 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:47.890136524 +0000 UTC m=+146.412866067 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.432715 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8l455" podStartSLOduration=124.43264418 podStartE2EDuration="2m4.43264418s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:47.427680672 +0000 UTC m=+145.950410215" watchObservedRunningTime="2026-02-17 13:06:47.43264418 +0000 UTC m=+145.955373723" Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.433747 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-qz68q" podStartSLOduration=124.433740673 podStartE2EDuration="2m4.433740673s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:47.403403879 +0000 UTC m=+145.926133422" watchObservedRunningTime="2026-02-17 13:06:47.433740673 +0000 UTC m=+145.956470226" Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.490971 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:47 crc kubenswrapper[4955]: E0217 13:06:47.491359 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:47.991346358 +0000 UTC m=+146.514075901 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.592126 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:47 crc kubenswrapper[4955]: E0217 13:06:47.592511 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:48.0924911 +0000 UTC m=+146.615220643 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.693552 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:47 crc kubenswrapper[4955]: E0217 13:06:47.694453 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:48.194433186 +0000 UTC m=+146.717162739 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.727564 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f7ccb" event={"ID":"c46d5d3e-92cd-415f-a04d-59756dd57637","Type":"ContainerStarted","Data":"7a4608e735500c50d26b336b7e43462f47d014d070bc0146118ce53d26be5653"} Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.730225 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tpfm" event={"ID":"a41bab7a-a90f-40a6-ba4f-b48ee6603fb5","Type":"ContainerStarted","Data":"b459b9053be096682bf2f488f7f338bd10614fa1f8bd0477ef1465eacc7ef057"} Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.734573 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4xvzn" event={"ID":"ff950861-7bca-4bfa-9857-cd906f67d5cb","Type":"ContainerStarted","Data":"6a3b91a8db5f346e11ffac93713b6fa1545d91b25b02ceb76be1d0a1b122bb9f"} Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.736159 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cb48c" event={"ID":"2f86fa65-b079-4331-b597-63a6c9995849","Type":"ContainerStarted","Data":"09d202ac6a33dc3c37f5ee79fb725cea92383cd9b990ad546c03e36d50692cfb"} Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.744208 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t5v8v" event={"ID":"f84f960e-7741-4d13-bd6f-ec2072cd4d14","Type":"ContainerStarted","Data":"5ac2917f285eaa7248b89417ead28b93ef26f1206721ce5e8b81b04216df737b"} Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.746691 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29522220-v5twm" event={"ID":"93fe04d3-c430-4ed6-ae04-312cf0613834","Type":"ContainerStarted","Data":"7283f9671ff1028392d197ada4636103257b37048f7bd0019d1e4c6209f7e59c"} Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.747817 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-vnwxg" event={"ID":"3a2b80bf-ead8-48e2-95fe-837c9ce39f61","Type":"ContainerStarted","Data":"7588a7e2ae877d913cf5faff37de7cfa14bb3afa780c55de3ee69ff1b3bd53b4"} Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.752337 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-zc8xw" event={"ID":"91cf9f0e-e498-41af-be13-fec987e4196c","Type":"ContainerStarted","Data":"e1dfde88a2b71e32fb65becf3d5ed16ee5b886b7e4a563926a04589694371e81"} Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.763802 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cb48c" podStartSLOduration=124.7637628 podStartE2EDuration="2m4.7637628s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:47.76275653 +0000 UTC m=+146.285486073" watchObservedRunningTime="2026-02-17 13:06:47.7637628 +0000 UTC m=+146.286492343" Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.764626 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" event={"ID":"f7b03b61-99a7-4db3-be68-e2ad2149f44b","Type":"ContainerStarted","Data":"cc506ca7959b82fc864ec0e4adc9ad768b07457682de68c5eae57837298f4c1b"} Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.766165 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.772147 4955 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-vfbvk container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.21:6443/healthz\": dial tcp 10.217.0.21:6443: connect: connection refused" start-of-body= Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.772269 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" podUID="f7b03b61-99a7-4db3-be68-e2ad2149f44b" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.21:6443/healthz\": dial tcp 10.217.0.21:6443: connect: connection refused" Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.788658 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z7bll" event={"ID":"be377252-9683-4490-960c-300c3c2007c4","Type":"ContainerStarted","Data":"8d8c728963fcd1a9370361f5b4db6bac0c4efb4e73e448157ba568cf8cab132f"} Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.788726 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z7bll" event={"ID":"be377252-9683-4490-960c-300c3c2007c4","Type":"ContainerStarted","Data":"5437ec716ecc13fe67695e1ff8169699de5a57e09cb46c2e353a502415ea405c"} Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.794561 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:47 crc kubenswrapper[4955]: E0217 13:06:47.795200 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:48.295106924 +0000 UTC m=+146.817836467 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.795991 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-brmrn" event={"ID":"2d2365ff-c132-4977-b394-c7bdb5f9c02d","Type":"ContainerStarted","Data":"f90a6455ed0ad18f08444a89256f49a53cf177de62b099d18d1838c9eb156459"} Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.802835 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" podStartSLOduration=124.802823764 podStartE2EDuration="2m4.802823764s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:47.799703621 +0000 UTC m=+146.322433164" watchObservedRunningTime="2026-02-17 13:06:47.802823764 +0000 UTC m=+146.325553307" Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.807686 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-64cqc" event={"ID":"e8351820-e7b6-4477-9944-5875c76514ce","Type":"ContainerStarted","Data":"29ad2f09c0648d60952ee8c98c54203f278d65cc8c38cf6f8378705b0f825bf6"} Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.815409 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4hhxl" event={"ID":"89d41921-f56f-4e2e-8ef0-2d981933ad68","Type":"ContainerStarted","Data":"d856583b435f84c07b4c26845a70910e2c2e20ba56f24eb259286215b9c082b2"} Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.816548 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4hhxl" Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.820545 4955 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-4hhxl container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" start-of-body= Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.820587 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4hhxl" podUID="89d41921-f56f-4e2e-8ef0-2d981933ad68" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.828973 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vtd5z" event={"ID":"8f5feb72-96ca-4e4d-9566-0757c7dd93f9","Type":"ContainerStarted","Data":"65af0794e599ac274a61ee831072cff54dded093ba3ce99fe2a08c12b738ad21"} Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.829388 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vtd5z" event={"ID":"8f5feb72-96ca-4e4d-9566-0757c7dd93f9","Type":"ContainerStarted","Data":"ad699fcebce76c6ef151fa20f8734f66ccd548881a5b80155a8ec649706bb607"} Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.834599 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-nskw2" event={"ID":"9d7d0f56-00a3-4eae-8ff2-a35e1f20a377","Type":"ContainerStarted","Data":"ba278d6f08b258e0098222e18b271d5f90e52bcc641aa9d25aeab46957f04a4f"} Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.835333 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-nskw2" Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.874063 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5264p" event={"ID":"3955ee14-c2d4-434e-a9b5-baf95abcbd02","Type":"ContainerStarted","Data":"9cf9d81d01389caefa83acc047aac90b2596dc0296c8f0acc49cfa5939dec90f"} Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.874752 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5264p" Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.893454 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-m7gfn" event={"ID":"cea310be-8fea-43cf-a000-69b614f260e4","Type":"ContainerStarted","Data":"72a919e6fc8dd3235b20b39ddda0e2275f7e98b4cd2950b6719f3b5ff497e5d0"} Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.896659 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:47 crc kubenswrapper[4955]: E0217 13:06:47.899321 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:48.399304907 +0000 UTC m=+146.922034450 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.912336 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-hs5dk" event={"ID":"81e66e5c-cae0-4eb7-87fb-273620ab5331","Type":"ContainerStarted","Data":"a92e05378c4f7ac577472b8fb3819a73c7c65fecb1345b37c60e04f00d3bfb93"} Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.925493 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4hhxl" podStartSLOduration=124.925476076 podStartE2EDuration="2m4.925476076s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:47.849426801 +0000 UTC m=+146.372156344" watchObservedRunningTime="2026-02-17 13:06:47.925476076 +0000 UTC m=+146.448205619" Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.926153 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-nskw2" podStartSLOduration=124.926147736 podStartE2EDuration="2m4.926147736s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:47.923682663 +0000 UTC m=+146.446412206" watchObservedRunningTime="2026-02-17 13:06:47.926147736 +0000 UTC m=+146.448877279" Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.945259 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-6hmm5" event={"ID":"4c087448-28c8-4708-897c-9eaf07323b72","Type":"ContainerStarted","Data":"66ead0837747764ba200b19659c3eeb90ef88b0f7d39806c6dde9393a29a5d12"} Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.952712 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5264p" podStartSLOduration=124.952692456 podStartE2EDuration="2m4.952692456s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:47.947464261 +0000 UTC m=+146.470193804" watchObservedRunningTime="2026-02-17 13:06:47.952692456 +0000 UTC m=+146.475421999" Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.959178 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8dx57" event={"ID":"922b208d-fc86-4243-86f1-eacc2797118e","Type":"ContainerStarted","Data":"a873b8e91ed8d70d7a4d2658d963ce68fb1fd0ed308c6785040db5aa769db9e7"} Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.969830 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sv4jj" event={"ID":"a800c1f4-8f07-4d13-a112-e3e4073c0241","Type":"ContainerStarted","Data":"9328ea3aa0294816061c3e71c4441d292579d12217635f499f7a69deb9fe8cf6"} Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.994086 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-5s2bx" event={"ID":"8259766a-777c-4103-a05b-091adc07ad32","Type":"ContainerStarted","Data":"b10f4829e614d27244043e341b4ee474feb0f33538933964aa243757ca97c20c"} Feb 17 13:06:47 crc kubenswrapper[4955]: I0217 13:06:47.998331 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:47 crc kubenswrapper[4955]: E0217 13:06:47.999742 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:48.499716887 +0000 UTC m=+147.022446500 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:48 crc kubenswrapper[4955]: I0217 13:06:48.018387 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qbxwp" event={"ID":"784c9c2c-64a9-4887-b52c-0949da50de62","Type":"ContainerStarted","Data":"44b76c3898d5f2dec35fbf89d4ebad85060a173244385d0e27c3e58b7435e322"} Feb 17 13:06:48 crc kubenswrapper[4955]: I0217 13:06:48.029171 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-m7gfn" podStartSLOduration=125.029151603 podStartE2EDuration="2m5.029151603s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:47.989047579 +0000 UTC m=+146.511777122" watchObservedRunningTime="2026-02-17 13:06:48.029151603 +0000 UTC m=+146.551881146" Feb 17 13:06:48 crc kubenswrapper[4955]: I0217 13:06:48.054070 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9c9zp" event={"ID":"84d8965a-d181-4f61-921f-1802347f6606","Type":"ContainerStarted","Data":"c1840dd14017cbc9ef34d4d29a76e441fab814a35992f6aaa4a0948828651f3e"} Feb 17 13:06:48 crc kubenswrapper[4955]: I0217 13:06:48.064549 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qdg4k" event={"ID":"b3dbb323-561b-4d26-886c-bef115c0dca0","Type":"ContainerStarted","Data":"f8690b331738914c89b67b1768fe2cf62c0b63c6e6dd539f63a7e10274b199af"} Feb 17 13:06:48 crc kubenswrapper[4955]: I0217 13:06:48.067981 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8t8vm" event={"ID":"1546cd3a-91cf-4e78-91e3-46ee4b8fbb43","Type":"ContainerStarted","Data":"ddec318221c1496c4d52bd48129a266bb7ff0ba9bade9ef71d01e4d747254199"} Feb 17 13:06:48 crc kubenswrapper[4955]: I0217 13:06:48.078658 4955 patch_prober.go:28] interesting pod/downloads-7954f5f757-qz68q container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Feb 17 13:06:48 crc kubenswrapper[4955]: I0217 13:06:48.078702 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-qz68q" podUID="01aa5d18-e942-4a3d-8af5-1e41b503f85a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.37:8080/\": dial tcp 10.217.0.37:8080: connect: connection refused" Feb 17 13:06:48 crc kubenswrapper[4955]: I0217 13:06:48.084909 4955 csr.go:261] certificate signing request csr-gvxlq is approved, waiting to be issued Feb 17 13:06:48 crc kubenswrapper[4955]: I0217 13:06:48.102835 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:48 crc kubenswrapper[4955]: E0217 13:06:48.103708 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:48.603694323 +0000 UTC m=+147.126423856 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:48 crc kubenswrapper[4955]: I0217 13:06:48.113259 4955 csr.go:257] certificate signing request csr-gvxlq is issued Feb 17 13:06:48 crc kubenswrapper[4955]: I0217 13:06:48.127489 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sv4jj" podStartSLOduration=125.127460211 podStartE2EDuration="2m5.127460211s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:48.063607119 +0000 UTC m=+146.586336652" watchObservedRunningTime="2026-02-17 13:06:48.127460211 +0000 UTC m=+146.650189754" Feb 17 13:06:48 crc kubenswrapper[4955]: I0217 13:06:48.177812 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-trrhn" Feb 17 13:06:48 crc kubenswrapper[4955]: I0217 13:06:48.203291 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-prtk8" Feb 17 13:06:48 crc kubenswrapper[4955]: I0217 13:06:48.204078 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:48 crc kubenswrapper[4955]: E0217 13:06:48.205954 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:48.705931707 +0000 UTC m=+147.228661250 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:48 crc kubenswrapper[4955]: I0217 13:06:48.211597 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-5s2bx" podStartSLOduration=6.211563905 podStartE2EDuration="6.211563905s" podCreationTimestamp="2026-02-17 13:06:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:48.178350796 +0000 UTC m=+146.701080349" watchObservedRunningTime="2026-02-17 13:06:48.211563905 +0000 UTC m=+146.734293548" Feb 17 13:06:48 crc kubenswrapper[4955]: I0217 13:06:48.216461 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qbxwp" podStartSLOduration=125.216270975 podStartE2EDuration="2m5.216270975s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:48.132040467 +0000 UTC m=+146.654770010" watchObservedRunningTime="2026-02-17 13:06:48.216270975 +0000 UTC m=+146.739000528" Feb 17 13:06:48 crc kubenswrapper[4955]: I0217 13:06:48.242013 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8t8vm" podStartSLOduration=125.241982881 podStartE2EDuration="2m5.241982881s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:48.207658999 +0000 UTC m=+146.730388542" watchObservedRunningTime="2026-02-17 13:06:48.241982881 +0000 UTC m=+146.764712424" Feb 17 13:06:48 crc kubenswrapper[4955]: I0217 13:06:48.262358 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-wdpjf" podStartSLOduration=125.262343647 podStartE2EDuration="2m5.262343647s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:48.261377978 +0000 UTC m=+146.784107521" watchObservedRunningTime="2026-02-17 13:06:48.262343647 +0000 UTC m=+146.785073190" Feb 17 13:06:48 crc kubenswrapper[4955]: I0217 13:06:48.281643 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-m7gfn" Feb 17 13:06:48 crc kubenswrapper[4955]: I0217 13:06:48.281838 4955 patch_prober.go:28] interesting pod/router-default-5444994796-m7gfn container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Feb 17 13:06:48 crc kubenswrapper[4955]: I0217 13:06:48.281881 4955 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-m7gfn" podUID="cea310be-8fea-43cf-a000-69b614f260e4" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Feb 17 13:06:48 crc kubenswrapper[4955]: I0217 13:06:48.306362 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:48 crc kubenswrapper[4955]: E0217 13:06:48.307240 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:48.807225324 +0000 UTC m=+147.329954857 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:48 crc kubenswrapper[4955]: I0217 13:06:48.336045 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qdg4k" podStartSLOduration=125.336016991 podStartE2EDuration="2m5.336016991s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:48.335700012 +0000 UTC m=+146.858429585" watchObservedRunningTime="2026-02-17 13:06:48.336016991 +0000 UTC m=+146.858746544" Feb 17 13:06:48 crc kubenswrapper[4955]: I0217 13:06:48.418697 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:48 crc kubenswrapper[4955]: E0217 13:06:48.419677 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:48.919652732 +0000 UTC m=+147.442382285 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:48 crc kubenswrapper[4955]: I0217 13:06:48.521396 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:48 crc kubenswrapper[4955]: E0217 13:06:48.521931 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:49.021913537 +0000 UTC m=+147.544643080 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:48 crc kubenswrapper[4955]: I0217 13:06:48.623043 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:48 crc kubenswrapper[4955]: E0217 13:06:48.623963 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:49.123940024 +0000 UTC m=+147.646669567 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:48 crc kubenswrapper[4955]: I0217 13:06:48.725600 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:48 crc kubenswrapper[4955]: E0217 13:06:48.726192 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:49.226170738 +0000 UTC m=+147.748900441 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:48 crc kubenswrapper[4955]: I0217 13:06:48.827548 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:48 crc kubenswrapper[4955]: E0217 13:06:48.827995 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:49.32797404 +0000 UTC m=+147.850703583 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:48 crc kubenswrapper[4955]: I0217 13:06:48.930129 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:48 crc kubenswrapper[4955]: E0217 13:06:48.931023 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:49.431005068 +0000 UTC m=+147.953734611 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.031609 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:49 crc kubenswrapper[4955]: E0217 13:06:49.032049 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:49.532026946 +0000 UTC m=+148.054756489 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.032502 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:49 crc kubenswrapper[4955]: E0217 13:06:49.033037 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:49.533029186 +0000 UTC m=+148.055758719 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.114635 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-17 13:01:48 +0000 UTC, rotation deadline is 2026-11-13 07:34:04.145022965 +0000 UTC Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.115144 4955 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6450h27m15.029884774s for next certificate rotation Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.133347 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:49 crc kubenswrapper[4955]: E0217 13:06:49.133490 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:49.633459186 +0000 UTC m=+148.156188729 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.133602 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:49 crc kubenswrapper[4955]: E0217 13:06:49.134136 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:49.634121256 +0000 UTC m=+148.156850799 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.163321 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-64cqc" event={"ID":"e8351820-e7b6-4477-9944-5875c76514ce","Type":"ContainerStarted","Data":"97d5e984bba928f09636d6f1d82a4ed6eff101edcfc79521bc6227591e3b4e27"} Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.165854 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tpfm" event={"ID":"a41bab7a-a90f-40a6-ba4f-b48ee6603fb5","Type":"ContainerStarted","Data":"5d565ce032239fdf0be0b6083f13e9684c9db3b17fc98590b8947b020b9c1d66"} Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.170255 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-vnwxg" event={"ID":"3a2b80bf-ead8-48e2-95fe-837c9ce39f61","Type":"ContainerStarted","Data":"c9ee2a862f6830f5928c085620051d4b7d6e54e9e27225b5994006191a62240d"} Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.179770 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-hs5dk" event={"ID":"81e66e5c-cae0-4eb7-87fb-273620ab5331","Type":"ContainerStarted","Data":"28c1baec0e07679bb5203292685c5cde42d3045b1add49260b10c0868be52df5"} Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.183154 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29522220-v5twm" event={"ID":"93fe04d3-c430-4ed6-ae04-312cf0613834","Type":"ContainerStarted","Data":"a32124e8d3b973e4e5f3f6e8a4c8205bc70e8fba11e9335172e8be548ef9d4cb"} Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.188015 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5x7zc" event={"ID":"58bab2f1-e131-46cb-b705-6bee97eb8452","Type":"ContainerStarted","Data":"b22ac1aae5ea9903d93a433b57961310221c1a29342ed00f68701ac4ebb556cf"} Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.191514 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8dx57" event={"ID":"922b208d-fc86-4243-86f1-eacc2797118e","Type":"ContainerStarted","Data":"ddd4c66e3e684e2a243fca189cc61ff3ffe8f7896faf1ef7a698c340be2f08c0"} Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.194066 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9c9zp" event={"ID":"84d8965a-d181-4f61-921f-1802347f6606","Type":"ContainerStarted","Data":"11d49f6fc9e716b23a59cefaaaff05e16ef8d468c92c03e402a3224d7476fe31"} Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.194266 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-9c9zp" Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.196349 4955 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-9c9zp container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.196516 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-9c9zp" podUID="84d8965a-d181-4f61-921f-1802347f6606" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.199251 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-zc8xw" event={"ID":"91cf9f0e-e498-41af-be13-fec987e4196c","Type":"ContainerStarted","Data":"25551ec1c25cd41c0b05caddf9723942c408d591b48c8fd546143069817dd002"} Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.204193 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-6hmm5" event={"ID":"4c087448-28c8-4708-897c-9eaf07323b72","Type":"ContainerStarted","Data":"a504c001370eede0b67c89c34380d31df69a37803e27f2054e9320c55a87c735"} Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.207855 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f7ccb" event={"ID":"c46d5d3e-92cd-415f-a04d-59756dd57637","Type":"ContainerStarted","Data":"e9e43d328a047e6d204be0ab6fa2941381627568e96370f1b02b1c40b2cb7311"} Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.208459 4955 patch_prober.go:28] interesting pod/downloads-7954f5f757-qz68q container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.208516 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-qz68q" podUID="01aa5d18-e942-4a3d-8af5-1e41b503f85a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.37:8080/\": dial tcp 10.217.0.37:8080: connect: connection refused" Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.208666 4955 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-4hhxl container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" start-of-body= Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.208730 4955 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-vfbvk container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.21:6443/healthz\": dial tcp 10.217.0.21:6443: connect: connection refused" start-of-body= Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.208766 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" podUID="f7b03b61-99a7-4db3-be68-e2ad2149f44b" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.21:6443/healthz\": dial tcp 10.217.0.21:6443: connect: connection refused" Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.208727 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4hhxl" podUID="89d41921-f56f-4e2e-8ef0-2d981933ad68" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.210379 4955 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-f7ccb container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.210414 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f7ccb" podUID="c46d5d3e-92cd-415f-a04d-59756dd57637" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.210723 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f7ccb" Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.234915 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:49 crc kubenswrapper[4955]: E0217 13:06:49.235394 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:49.735371781 +0000 UTC m=+148.258101324 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.237696 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tpfm" podStartSLOduration=126.23766981 podStartE2EDuration="2m6.23766981s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:49.226517367 +0000 UTC m=+147.749246920" watchObservedRunningTime="2026-02-17 13:06:49.23766981 +0000 UTC m=+147.760399353" Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.281093 4955 patch_prober.go:28] interesting pod/router-default-5444994796-m7gfn container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.281482 4955 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-m7gfn" podUID="cea310be-8fea-43cf-a000-69b614f260e4" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.290802 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-zc8xw" podStartSLOduration=125.290761651 podStartE2EDuration="2m5.290761651s" podCreationTimestamp="2026-02-17 13:04:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:49.287485523 +0000 UTC m=+147.810215066" watchObservedRunningTime="2026-02-17 13:06:49.290761651 +0000 UTC m=+147.813491194" Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.336611 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:49 crc kubenswrapper[4955]: E0217 13:06:49.375256 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:49.875232966 +0000 UTC m=+148.397962509 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.410032 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-8dx57" Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.410491 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-8dx57" Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.412666 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z7bll" podStartSLOduration=126.41264767 podStartE2EDuration="2m6.41264767s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:49.339042828 +0000 UTC m=+147.861772371" watchObservedRunningTime="2026-02-17 13:06:49.41264767 +0000 UTC m=+147.935377213" Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.414000 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4xvzn" podStartSLOduration=126.41399507 podStartE2EDuration="2m6.41399507s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:49.41097305 +0000 UTC m=+147.933702593" watchObservedRunningTime="2026-02-17 13:06:49.41399507 +0000 UTC m=+147.936724613" Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.418231 4955 patch_prober.go:28] interesting pod/apiserver-76f77b778f-8dx57 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.11:8443/livez\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.418279 4955 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-8dx57" podUID="922b208d-fc86-4243-86f1-eacc2797118e" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.11:8443/livez\": dial tcp 10.217.0.11:8443: connect: connection refused" Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.438457 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:49 crc kubenswrapper[4955]: E0217 13:06:49.438772 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:49.938747857 +0000 UTC m=+148.461477400 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.438958 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:49 crc kubenswrapper[4955]: E0217 13:06:49.439422 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:49.939412237 +0000 UTC m=+148.462141780 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.442802 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vtd5z" podStartSLOduration=126.442767657 podStartE2EDuration="2m6.442767657s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:49.441976694 +0000 UTC m=+147.964706237" watchObservedRunningTime="2026-02-17 13:06:49.442767657 +0000 UTC m=+147.965497200" Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.539120 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-9c9zp" podStartSLOduration=126.539099516 podStartE2EDuration="2m6.539099516s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:49.538104366 +0000 UTC m=+148.060833909" watchObservedRunningTime="2026-02-17 13:06:49.539099516 +0000 UTC m=+148.061829069" Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.540011 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:49 crc kubenswrapper[4955]: E0217 13:06:49.540467 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:50.040450476 +0000 UTC m=+148.563180019 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.589190 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-hs5dk" podStartSLOduration=7.589173797 podStartE2EDuration="7.589173797s" podCreationTimestamp="2026-02-17 13:06:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:49.588218538 +0000 UTC m=+148.110948081" watchObservedRunningTime="2026-02-17 13:06:49.589173797 +0000 UTC m=+148.111903340" Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.641690 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:49 crc kubenswrapper[4955]: E0217 13:06:49.642204 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:50.142185076 +0000 UTC m=+148.664914629 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.693811 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5x7zc" podStartSLOduration=126.693791792 podStartE2EDuration="2m6.693791792s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:49.651003608 +0000 UTC m=+148.173733151" watchObservedRunningTime="2026-02-17 13:06:49.693791792 +0000 UTC m=+148.216521335" Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.742584 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:49 crc kubenswrapper[4955]: E0217 13:06:49.743070 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:50.243049519 +0000 UTC m=+148.765779062 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.746423 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-8dx57" podStartSLOduration=126.746403309 podStartE2EDuration="2m6.746403309s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:49.694769681 +0000 UTC m=+148.217499224" watchObservedRunningTime="2026-02-17 13:06:49.746403309 +0000 UTC m=+148.269132852" Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.774286 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f7ccb" podStartSLOduration=126.774268099 podStartE2EDuration="2m6.774268099s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:49.773121505 +0000 UTC m=+148.295851048" watchObservedRunningTime="2026-02-17 13:06:49.774268099 +0000 UTC m=+148.296997642" Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.774768 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-6hmm5" podStartSLOduration=126.774761904 podStartE2EDuration="2m6.774761904s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:49.746825712 +0000 UTC m=+148.269555255" watchObservedRunningTime="2026-02-17 13:06:49.774761904 +0000 UTC m=+148.297491447" Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.843820 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:49 crc kubenswrapper[4955]: E0217 13:06:49.844189 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:50.344173391 +0000 UTC m=+148.866902924 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.855705 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29522220-v5twm" podStartSLOduration=126.855686683 podStartE2EDuration="2m6.855686683s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:49.829860204 +0000 UTC m=+148.352589747" watchObservedRunningTime="2026-02-17 13:06:49.855686683 +0000 UTC m=+148.378416226" Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.920463 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t5v8v" podStartSLOduration=126.920445762 podStartE2EDuration="2m6.920445762s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:49.856226019 +0000 UTC m=+148.378955562" watchObservedRunningTime="2026-02-17 13:06:49.920445762 +0000 UTC m=+148.443175305" Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.921279 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-brmrn" podStartSLOduration=125.921273176 podStartE2EDuration="2m5.921273176s" podCreationTimestamp="2026-02-17 13:04:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:49.918348539 +0000 UTC m=+148.441078082" watchObservedRunningTime="2026-02-17 13:06:49.921273176 +0000 UTC m=+148.444002719" Feb 17 13:06:49 crc kubenswrapper[4955]: I0217 13:06:49.945285 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:49 crc kubenswrapper[4955]: E0217 13:06:49.946263 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:50.44622765 +0000 UTC m=+148.968957193 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:50 crc kubenswrapper[4955]: I0217 13:06:50.047400 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:50 crc kubenswrapper[4955]: E0217 13:06:50.047870 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:50.547853696 +0000 UTC m=+149.070583239 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:50 crc kubenswrapper[4955]: I0217 13:06:50.148995 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:50 crc kubenswrapper[4955]: E0217 13:06:50.149427 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:50.64940994 +0000 UTC m=+149.172139483 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:50 crc kubenswrapper[4955]: I0217 13:06:50.214705 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-64cqc" event={"ID":"e8351820-e7b6-4477-9944-5875c76514ce","Type":"ContainerStarted","Data":"e83eb412187f87a64e70d352735b5cf6fd9963dd5e2d736361c234e1fbc871fd"} Feb 17 13:06:50 crc kubenswrapper[4955]: I0217 13:06:50.214861 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-64cqc" Feb 17 13:06:50 crc kubenswrapper[4955]: I0217 13:06:50.217268 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-45lhd" event={"ID":"ce6956ee-96b9-4fea-af68-90540d249ba8","Type":"ContainerStarted","Data":"bf25c321a6deb9fb481f91827155ee79de95d7435f260d49aee022d5864eb69c"} Feb 17 13:06:50 crc kubenswrapper[4955]: I0217 13:06:50.217742 4955 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-9c9zp container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Feb 17 13:06:50 crc kubenswrapper[4955]: I0217 13:06:50.217793 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-9c9zp" podUID="84d8965a-d181-4f61-921f-1802347f6606" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" Feb 17 13:06:50 crc kubenswrapper[4955]: I0217 13:06:50.217929 4955 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-f7ccb container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Feb 17 13:06:50 crc kubenswrapper[4955]: I0217 13:06:50.217967 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f7ccb" podUID="c46d5d3e-92cd-415f-a04d-59756dd57637" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" Feb 17 13:06:50 crc kubenswrapper[4955]: I0217 13:06:50.226595 4955 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-nskw2 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Feb 17 13:06:50 crc kubenswrapper[4955]: I0217 13:06:50.226652 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-nskw2" podUID="9d7d0f56-00a3-4eae-8ff2-a35e1f20a377" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" Feb 17 13:06:50 crc kubenswrapper[4955]: I0217 13:06:50.233233 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4hhxl" Feb 17 13:06:50 crc kubenswrapper[4955]: I0217 13:06:50.248226 4955 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-nskw2 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Feb 17 13:06:50 crc kubenswrapper[4955]: I0217 13:06:50.248469 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-nskw2" podUID="9d7d0f56-00a3-4eae-8ff2-a35e1f20a377" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" Feb 17 13:06:50 crc kubenswrapper[4955]: I0217 13:06:50.248633 4955 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-nskw2 container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Feb 17 13:06:50 crc kubenswrapper[4955]: I0217 13:06:50.248677 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-nskw2" podUID="9d7d0f56-00a3-4eae-8ff2-a35e1f20a377" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" Feb 17 13:06:50 crc kubenswrapper[4955]: I0217 13:06:50.250425 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:50 crc kubenswrapper[4955]: E0217 13:06:50.250739 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:50.750727917 +0000 UTC m=+149.273457460 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:50 crc kubenswrapper[4955]: I0217 13:06:50.261124 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-vnwxg" podStartSLOduration=127.261074435 podStartE2EDuration="2m7.261074435s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:49.953819536 +0000 UTC m=+148.476549109" watchObservedRunningTime="2026-02-17 13:06:50.261074435 +0000 UTC m=+148.783803978" Feb 17 13:06:50 crc kubenswrapper[4955]: I0217 13:06:50.261508 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-64cqc" podStartSLOduration=8.261502698 podStartE2EDuration="8.261502698s" podCreationTimestamp="2026-02-17 13:06:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:50.254608193 +0000 UTC m=+148.777337736" watchObservedRunningTime="2026-02-17 13:06:50.261502698 +0000 UTC m=+148.784232241" Feb 17 13:06:50 crc kubenswrapper[4955]: I0217 13:06:50.283451 4955 patch_prober.go:28] interesting pod/router-default-5444994796-m7gfn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 17 13:06:50 crc kubenswrapper[4955]: [-]has-synced failed: reason withheld Feb 17 13:06:50 crc kubenswrapper[4955]: [+]process-running ok Feb 17 13:06:50 crc kubenswrapper[4955]: healthz check failed Feb 17 13:06:50 crc kubenswrapper[4955]: I0217 13:06:50.283730 4955 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-m7gfn" podUID="cea310be-8fea-43cf-a000-69b614f260e4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 17 13:06:50 crc kubenswrapper[4955]: I0217 13:06:50.351886 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:50 crc kubenswrapper[4955]: E0217 13:06:50.352542 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:50.852521528 +0000 UTC m=+149.375251071 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:50 crc kubenswrapper[4955]: I0217 13:06:50.358311 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:50 crc kubenswrapper[4955]: E0217 13:06:50.359119 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:50.859103554 +0000 UTC m=+149.381833097 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:50 crc kubenswrapper[4955]: I0217 13:06:50.462373 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:50 crc kubenswrapper[4955]: E0217 13:06:50.462832 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:50.962814203 +0000 UTC m=+149.485543746 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:50 crc kubenswrapper[4955]: I0217 13:06:50.569277 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:50 crc kubenswrapper[4955]: E0217 13:06:50.569765 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:51.069741297 +0000 UTC m=+149.592470840 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:50 crc kubenswrapper[4955]: I0217 13:06:50.671252 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:50 crc kubenswrapper[4955]: E0217 13:06:50.671588 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:51.171571189 +0000 UTC m=+149.694300732 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:50 crc kubenswrapper[4955]: I0217 13:06:50.772878 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:50 crc kubenswrapper[4955]: E0217 13:06:50.773685 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:51.273669259 +0000 UTC m=+149.796398802 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:50 crc kubenswrapper[4955]: I0217 13:06:50.875225 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:50 crc kubenswrapper[4955]: E0217 13:06:50.876002 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:51.375981776 +0000 UTC m=+149.898711319 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:50 crc kubenswrapper[4955]: I0217 13:06:50.976920 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:50 crc kubenswrapper[4955]: E0217 13:06:50.977394 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:51.477376095 +0000 UTC m=+150.000105628 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:51 crc kubenswrapper[4955]: I0217 13:06:51.078105 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:51 crc kubenswrapper[4955]: E0217 13:06:51.078285 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:51.578241519 +0000 UTC m=+150.100971062 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:51 crc kubenswrapper[4955]: I0217 13:06:51.078692 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:51 crc kubenswrapper[4955]: E0217 13:06:51.079203 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:51.579174367 +0000 UTC m=+150.101903940 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:51 crc kubenswrapper[4955]: I0217 13:06:51.179560 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:51 crc kubenswrapper[4955]: E0217 13:06:51.179763 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:51.679730491 +0000 UTC m=+150.202460034 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:51 crc kubenswrapper[4955]: I0217 13:06:51.179901 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:06:51 crc kubenswrapper[4955]: I0217 13:06:51.179948 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:06:51 crc kubenswrapper[4955]: I0217 13:06:51.179972 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:06:51 crc kubenswrapper[4955]: I0217 13:06:51.180054 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:06:51 crc kubenswrapper[4955]: I0217 13:06:51.180101 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:51 crc kubenswrapper[4955]: E0217 13:06:51.180496 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:51.680482594 +0000 UTC m=+150.203212137 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:51 crc kubenswrapper[4955]: I0217 13:06:51.189003 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:06:51 crc kubenswrapper[4955]: I0217 13:06:51.189112 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:06:51 crc kubenswrapper[4955]: I0217 13:06:51.192481 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:06:51 crc kubenswrapper[4955]: I0217 13:06:51.201713 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:06:51 crc kubenswrapper[4955]: I0217 13:06:51.218066 4955 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-vfbvk container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.21:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 17 13:06:51 crc kubenswrapper[4955]: I0217 13:06:51.219391 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" podUID="f7b03b61-99a7-4db3-be68-e2ad2149f44b" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.21:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 17 13:06:51 crc kubenswrapper[4955]: I0217 13:06:51.249152 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:06:51 crc kubenswrapper[4955]: I0217 13:06:51.290301 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 17 13:06:51 crc kubenswrapper[4955]: I0217 13:06:51.291052 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:51 crc kubenswrapper[4955]: I0217 13:06:51.291102 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 17 13:06:51 crc kubenswrapper[4955]: E0217 13:06:51.291386 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:51.791366675 +0000 UTC m=+150.314096218 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:51 crc kubenswrapper[4955]: I0217 13:06:51.294919 4955 patch_prober.go:28] interesting pod/router-default-5444994796-m7gfn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 17 13:06:51 crc kubenswrapper[4955]: [-]has-synced failed: reason withheld Feb 17 13:06:51 crc kubenswrapper[4955]: [+]process-running ok Feb 17 13:06:51 crc kubenswrapper[4955]: healthz check failed Feb 17 13:06:51 crc kubenswrapper[4955]: I0217 13:06:51.294957 4955 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-m7gfn" podUID="cea310be-8fea-43cf-a000-69b614f260e4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 17 13:06:51 crc kubenswrapper[4955]: I0217 13:06:51.392151 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:51 crc kubenswrapper[4955]: E0217 13:06:51.393643 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:51.8936228 +0000 UTC m=+150.416352413 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:51 crc kubenswrapper[4955]: I0217 13:06:51.494041 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:51 crc kubenswrapper[4955]: E0217 13:06:51.494749 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:51.994732281 +0000 UTC m=+150.517461824 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:51 crc kubenswrapper[4955]: I0217 13:06:51.595674 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:51 crc kubenswrapper[4955]: E0217 13:06:51.598611 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:52.098596014 +0000 UTC m=+150.621325557 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:51 crc kubenswrapper[4955]: I0217 13:06:51.696562 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:51 crc kubenswrapper[4955]: E0217 13:06:51.696942 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:52.196926022 +0000 UTC m=+150.719655565 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:51 crc kubenswrapper[4955]: I0217 13:06:51.805627 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:51 crc kubenswrapper[4955]: E0217 13:06:51.806038 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:52.306023641 +0000 UTC m=+150.828753184 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:51 crc kubenswrapper[4955]: I0217 13:06:51.907383 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:51 crc kubenswrapper[4955]: E0217 13:06:51.907716 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:52.407700029 +0000 UTC m=+150.930429572 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.012546 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:52 crc kubenswrapper[4955]: E0217 13:06:52.012879 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:52.51286496 +0000 UTC m=+151.035594503 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:52 crc kubenswrapper[4955]: W0217 13:06:52.019895 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-9f80f1809804a655c67a305b53b91c083c400633e26bed245492a9a2ee7f04af WatchSource:0}: Error finding container 9f80f1809804a655c67a305b53b91c083c400633e26bed245492a9a2ee7f04af: Status 404 returned error can't find the container with id 9f80f1809804a655c67a305b53b91c083c400633e26bed245492a9a2ee7f04af Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.116874 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:52 crc kubenswrapper[4955]: E0217 13:06:52.117226 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:52.617210448 +0000 UTC m=+151.139939991 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.169998 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pf44z"] Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.175347 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pf44z" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.209754 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.220711 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:52 crc kubenswrapper[4955]: E0217 13:06:52.221119 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:52.72110551 +0000 UTC m=+151.243835053 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.296053 4955 patch_prober.go:28] interesting pod/router-default-5444994796-m7gfn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 17 13:06:52 crc kubenswrapper[4955]: [-]has-synced failed: reason withheld Feb 17 13:06:52 crc kubenswrapper[4955]: [+]process-running ok Feb 17 13:06:52 crc kubenswrapper[4955]: healthz check failed Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.296140 4955 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-m7gfn" podUID="cea310be-8fea-43cf-a000-69b614f260e4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.296332 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pf44z"] Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.310412 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"9f80f1809804a655c67a305b53b91c083c400633e26bed245492a9a2ee7f04af"} Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.328405 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.328760 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b1996f2-ac59-43dd-a18b-baeabe4f5da3-utilities\") pod \"community-operators-pf44z\" (UID: \"3b1996f2-ac59-43dd-a18b-baeabe4f5da3\") " pod="openshift-marketplace/community-operators-pf44z" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.330816 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6xlxd"] Feb 17 13:06:52 crc kubenswrapper[4955]: E0217 13:06:52.333003 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:52.832963901 +0000 UTC m=+151.355693444 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.333073 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b1996f2-ac59-43dd-a18b-baeabe4f5da3-catalog-content\") pod \"community-operators-pf44z\" (UID: \"3b1996f2-ac59-43dd-a18b-baeabe4f5da3\") " pod="openshift-marketplace/community-operators-pf44z" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.333575 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vldh5\" (UniqueName: \"kubernetes.io/projected/3b1996f2-ac59-43dd-a18b-baeabe4f5da3-kube-api-access-vldh5\") pod \"community-operators-pf44z\" (UID: \"3b1996f2-ac59-43dd-a18b-baeabe4f5da3\") " pod="openshift-marketplace/community-operators-pf44z" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.342833 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6xlxd" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.378532 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.392145 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-45lhd" event={"ID":"ce6956ee-96b9-4fea-af68-90540d249ba8","Type":"ContainerStarted","Data":"57289742f1ca88a6bb29c0c2b572fed1ad7b8ad37ccd71f815a435f0895c0649"} Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.392205 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-45lhd" event={"ID":"ce6956ee-96b9-4fea-af68-90540d249ba8","Type":"ContainerStarted","Data":"5b8961388991824e03bdc74a7557b4a6e3f2db8f5a19d15897f837393ebec43e"} Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.392596 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6xlxd"] Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.436919 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vj64\" (UniqueName: \"kubernetes.io/projected/6d73442c-9d91-4b5b-b0dd-efc137a55959-kube-api-access-7vj64\") pod \"certified-operators-6xlxd\" (UID: \"6d73442c-9d91-4b5b-b0dd-efc137a55959\") " pod="openshift-marketplace/certified-operators-6xlxd" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.436975 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b1996f2-ac59-43dd-a18b-baeabe4f5da3-catalog-content\") pod \"community-operators-pf44z\" (UID: \"3b1996f2-ac59-43dd-a18b-baeabe4f5da3\") " pod="openshift-marketplace/community-operators-pf44z" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.436995 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vldh5\" (UniqueName: \"kubernetes.io/projected/3b1996f2-ac59-43dd-a18b-baeabe4f5da3-kube-api-access-vldh5\") pod \"community-operators-pf44z\" (UID: \"3b1996f2-ac59-43dd-a18b-baeabe4f5da3\") " pod="openshift-marketplace/community-operators-pf44z" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.437067 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d73442c-9d91-4b5b-b0dd-efc137a55959-utilities\") pod \"certified-operators-6xlxd\" (UID: \"6d73442c-9d91-4b5b-b0dd-efc137a55959\") " pod="openshift-marketplace/certified-operators-6xlxd" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.437100 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.437135 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b1996f2-ac59-43dd-a18b-baeabe4f5da3-utilities\") pod \"community-operators-pf44z\" (UID: \"3b1996f2-ac59-43dd-a18b-baeabe4f5da3\") " pod="openshift-marketplace/community-operators-pf44z" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.437156 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d73442c-9d91-4b5b-b0dd-efc137a55959-catalog-content\") pod \"certified-operators-6xlxd\" (UID: \"6d73442c-9d91-4b5b-b0dd-efc137a55959\") " pod="openshift-marketplace/certified-operators-6xlxd" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.437525 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b1996f2-ac59-43dd-a18b-baeabe4f5da3-catalog-content\") pod \"community-operators-pf44z\" (UID: \"3b1996f2-ac59-43dd-a18b-baeabe4f5da3\") " pod="openshift-marketplace/community-operators-pf44z" Feb 17 13:06:52 crc kubenswrapper[4955]: E0217 13:06:52.438138 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:52.938124903 +0000 UTC m=+151.460854446 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.443925 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"965beeba472f6c00e62cdabf190dde903ded58debece611f102800983c69708d"} Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.444576 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.450244 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b1996f2-ac59-43dd-a18b-baeabe4f5da3-utilities\") pod \"community-operators-pf44z\" (UID: \"3b1996f2-ac59-43dd-a18b-baeabe4f5da3\") " pod="openshift-marketplace/community-operators-pf44z" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.496637 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vldh5\" (UniqueName: \"kubernetes.io/projected/3b1996f2-ac59-43dd-a18b-baeabe4f5da3-kube-api-access-vldh5\") pod \"community-operators-pf44z\" (UID: \"3b1996f2-ac59-43dd-a18b-baeabe4f5da3\") " pod="openshift-marketplace/community-operators-pf44z" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.537465 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.537878 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d73442c-9d91-4b5b-b0dd-efc137a55959-catalog-content\") pod \"certified-operators-6xlxd\" (UID: \"6d73442c-9d91-4b5b-b0dd-efc137a55959\") " pod="openshift-marketplace/certified-operators-6xlxd" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.537977 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vj64\" (UniqueName: \"kubernetes.io/projected/6d73442c-9d91-4b5b-b0dd-efc137a55959-kube-api-access-7vj64\") pod \"certified-operators-6xlxd\" (UID: \"6d73442c-9d91-4b5b-b0dd-efc137a55959\") " pod="openshift-marketplace/certified-operators-6xlxd" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.538145 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d73442c-9d91-4b5b-b0dd-efc137a55959-utilities\") pod \"certified-operators-6xlxd\" (UID: \"6d73442c-9d91-4b5b-b0dd-efc137a55959\") " pod="openshift-marketplace/certified-operators-6xlxd" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.538603 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d73442c-9d91-4b5b-b0dd-efc137a55959-utilities\") pod \"certified-operators-6xlxd\" (UID: \"6d73442c-9d91-4b5b-b0dd-efc137a55959\") " pod="openshift-marketplace/certified-operators-6xlxd" Feb 17 13:06:52 crc kubenswrapper[4955]: E0217 13:06:52.538734 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:53.038689627 +0000 UTC m=+151.561419170 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.538964 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d73442c-9d91-4b5b-b0dd-efc137a55959-catalog-content\") pod \"certified-operators-6xlxd\" (UID: \"6d73442c-9d91-4b5b-b0dd-efc137a55959\") " pod="openshift-marketplace/certified-operators-6xlxd" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.544653 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-f8t7f"] Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.545945 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f8t7f" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.549216 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pf44z" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.563861 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f8t7f"] Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.645306 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1ff2d84-9dcc-43f3-821f-989fbeec8e93-utilities\") pod \"community-operators-f8t7f\" (UID: \"d1ff2d84-9dcc-43f3-821f-989fbeec8e93\") " pod="openshift-marketplace/community-operators-f8t7f" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.645388 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.645427 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbt48\" (UniqueName: \"kubernetes.io/projected/d1ff2d84-9dcc-43f3-821f-989fbeec8e93-kube-api-access-pbt48\") pod \"community-operators-f8t7f\" (UID: \"d1ff2d84-9dcc-43f3-821f-989fbeec8e93\") " pod="openshift-marketplace/community-operators-f8t7f" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.645452 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1ff2d84-9dcc-43f3-821f-989fbeec8e93-catalog-content\") pod \"community-operators-f8t7f\" (UID: \"d1ff2d84-9dcc-43f3-821f-989fbeec8e93\") " pod="openshift-marketplace/community-operators-f8t7f" Feb 17 13:06:52 crc kubenswrapper[4955]: E0217 13:06:52.645862 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:53.145847619 +0000 UTC m=+151.668577162 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.647263 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5kwdn"] Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.658254 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5kwdn" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.659756 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vj64\" (UniqueName: \"kubernetes.io/projected/6d73442c-9d91-4b5b-b0dd-efc137a55959-kube-api-access-7vj64\") pod \"certified-operators-6xlxd\" (UID: \"6d73442c-9d91-4b5b-b0dd-efc137a55959\") " pod="openshift-marketplace/certified-operators-6xlxd" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.691282 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5kwdn"] Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.747274 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.747572 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a9eeae4-44ec-4f00-a49f-2af0b006e5cb-catalog-content\") pod \"certified-operators-5kwdn\" (UID: \"7a9eeae4-44ec-4f00-a49f-2af0b006e5cb\") " pod="openshift-marketplace/certified-operators-5kwdn" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.747613 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbt48\" (UniqueName: \"kubernetes.io/projected/d1ff2d84-9dcc-43f3-821f-989fbeec8e93-kube-api-access-pbt48\") pod \"community-operators-f8t7f\" (UID: \"d1ff2d84-9dcc-43f3-821f-989fbeec8e93\") " pod="openshift-marketplace/community-operators-f8t7f" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.747635 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4r6tg\" (UniqueName: \"kubernetes.io/projected/7a9eeae4-44ec-4f00-a49f-2af0b006e5cb-kube-api-access-4r6tg\") pod \"certified-operators-5kwdn\" (UID: \"7a9eeae4-44ec-4f00-a49f-2af0b006e5cb\") " pod="openshift-marketplace/certified-operators-5kwdn" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.747657 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1ff2d84-9dcc-43f3-821f-989fbeec8e93-catalog-content\") pod \"community-operators-f8t7f\" (UID: \"d1ff2d84-9dcc-43f3-821f-989fbeec8e93\") " pod="openshift-marketplace/community-operators-f8t7f" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.747698 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1ff2d84-9dcc-43f3-821f-989fbeec8e93-utilities\") pod \"community-operators-f8t7f\" (UID: \"d1ff2d84-9dcc-43f3-821f-989fbeec8e93\") " pod="openshift-marketplace/community-operators-f8t7f" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.747738 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a9eeae4-44ec-4f00-a49f-2af0b006e5cb-utilities\") pod \"certified-operators-5kwdn\" (UID: \"7a9eeae4-44ec-4f00-a49f-2af0b006e5cb\") " pod="openshift-marketplace/certified-operators-5kwdn" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.748206 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1ff2d84-9dcc-43f3-821f-989fbeec8e93-catalog-content\") pod \"community-operators-f8t7f\" (UID: \"d1ff2d84-9dcc-43f3-821f-989fbeec8e93\") " pod="openshift-marketplace/community-operators-f8t7f" Feb 17 13:06:52 crc kubenswrapper[4955]: E0217 13:06:52.748264 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:53.248234927 +0000 UTC m=+151.770964470 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.748438 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1ff2d84-9dcc-43f3-821f-989fbeec8e93-utilities\") pod \"community-operators-f8t7f\" (UID: \"d1ff2d84-9dcc-43f3-821f-989fbeec8e93\") " pod="openshift-marketplace/community-operators-f8t7f" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.754616 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6xlxd" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.772922 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbt48\" (UniqueName: \"kubernetes.io/projected/d1ff2d84-9dcc-43f3-821f-989fbeec8e93-kube-api-access-pbt48\") pod \"community-operators-f8t7f\" (UID: \"d1ff2d84-9dcc-43f3-821f-989fbeec8e93\") " pod="openshift-marketplace/community-operators-f8t7f" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.850890 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a9eeae4-44ec-4f00-a49f-2af0b006e5cb-utilities\") pod \"certified-operators-5kwdn\" (UID: \"7a9eeae4-44ec-4f00-a49f-2af0b006e5cb\") " pod="openshift-marketplace/certified-operators-5kwdn" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.851191 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.851242 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a9eeae4-44ec-4f00-a49f-2af0b006e5cb-catalog-content\") pod \"certified-operators-5kwdn\" (UID: \"7a9eeae4-44ec-4f00-a49f-2af0b006e5cb\") " pod="openshift-marketplace/certified-operators-5kwdn" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.851271 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4r6tg\" (UniqueName: \"kubernetes.io/projected/7a9eeae4-44ec-4f00-a49f-2af0b006e5cb-kube-api-access-4r6tg\") pod \"certified-operators-5kwdn\" (UID: \"7a9eeae4-44ec-4f00-a49f-2af0b006e5cb\") " pod="openshift-marketplace/certified-operators-5kwdn" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.852174 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a9eeae4-44ec-4f00-a49f-2af0b006e5cb-utilities\") pod \"certified-operators-5kwdn\" (UID: \"7a9eeae4-44ec-4f00-a49f-2af0b006e5cb\") " pod="openshift-marketplace/certified-operators-5kwdn" Feb 17 13:06:52 crc kubenswrapper[4955]: E0217 13:06:52.852466 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:53.352453651 +0000 UTC m=+151.875183194 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.853578 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a9eeae4-44ec-4f00-a49f-2af0b006e5cb-catalog-content\") pod \"certified-operators-5kwdn\" (UID: \"7a9eeae4-44ec-4f00-a49f-2af0b006e5cb\") " pod="openshift-marketplace/certified-operators-5kwdn" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.877066 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f8t7f" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.877517 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4r6tg\" (UniqueName: \"kubernetes.io/projected/7a9eeae4-44ec-4f00-a49f-2af0b006e5cb-kube-api-access-4r6tg\") pod \"certified-operators-5kwdn\" (UID: \"7a9eeae4-44ec-4f00-a49f-2af0b006e5cb\") " pod="openshift-marketplace/certified-operators-5kwdn" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.898636 4955 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.935312 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pf44z"] Feb 17 13:06:52 crc kubenswrapper[4955]: I0217 13:06:52.954443 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:52 crc kubenswrapper[4955]: E0217 13:06:52.954861 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:53.45484433 +0000 UTC m=+151.977573863 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:52 crc kubenswrapper[4955]: W0217 13:06:52.961131 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b1996f2_ac59_43dd_a18b_baeabe4f5da3.slice/crio-072f6579e746217ed7519c11635ff385805ef526c07510840caa7c5b45b62b71 WatchSource:0}: Error finding container 072f6579e746217ed7519c11635ff385805ef526c07510840caa7c5b45b62b71: Status 404 returned error can't find the container with id 072f6579e746217ed7519c11635ff385805ef526c07510840caa7c5b45b62b71 Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:52.998391 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5kwdn" Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.056598 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:53 crc kubenswrapper[4955]: E0217 13:06:53.057024 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:53.557011022 +0000 UTC m=+152.079740565 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.158466 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:53 crc kubenswrapper[4955]: E0217 13:06:53.158916 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:53.658894146 +0000 UTC m=+152.181623689 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.256644 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-nskw2" Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.259431 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:53 crc kubenswrapper[4955]: E0217 13:06:53.259729 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:53.759716839 +0000 UTC m=+152.282446382 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.274088 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.274745 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.280065 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.280333 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.294082 4955 patch_prober.go:28] interesting pod/router-default-5444994796-m7gfn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 17 13:06:53 crc kubenswrapper[4955]: [-]has-synced failed: reason withheld Feb 17 13:06:53 crc kubenswrapper[4955]: [+]process-running ok Feb 17 13:06:53 crc kubenswrapper[4955]: healthz check failed Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.294138 4955 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-m7gfn" podUID="cea310be-8fea-43cf-a000-69b614f260e4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.298165 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.350567 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6xlxd"] Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.362086 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:53 crc kubenswrapper[4955]: E0217 13:06:53.364738 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-17 13:06:53.864704185 +0000 UTC m=+152.387433738 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.365306 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/529893f0-2392-4d15-ae52-7f4d1646de55-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"529893f0-2392-4d15-ae52-7f4d1646de55\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.365346 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/529893f0-2392-4d15-ae52-7f4d1646de55-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"529893f0-2392-4d15-ae52-7f4d1646de55\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.365397 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:53 crc kubenswrapper[4955]: E0217 13:06:53.368584 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-17 13:06:53.86857029 +0000 UTC m=+152.391299823 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8mkhz" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.429421 4955 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-02-17T13:06:52.898667397Z","Handler":null,"Name":""} Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.435700 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f8t7f"] Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.452035 4955 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.452246 4955 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.458489 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pf44z" event={"ID":"3b1996f2-ac59-43dd-a18b-baeabe4f5da3","Type":"ContainerDied","Data":"33a9b4f3a61747aceb04131eff2af6ceea8dbeb951a20741327b4b4b44057854"} Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.458456 4955 generic.go:334] "Generic (PLEG): container finished" podID="3b1996f2-ac59-43dd-a18b-baeabe4f5da3" containerID="33a9b4f3a61747aceb04131eff2af6ceea8dbeb951a20741327b4b4b44057854" exitCode=0 Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.458774 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pf44z" event={"ID":"3b1996f2-ac59-43dd-a18b-baeabe4f5da3","Type":"ContainerStarted","Data":"072f6579e746217ed7519c11635ff385805ef526c07510840caa7c5b45b62b71"} Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.460634 4955 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.462207 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"bf313256e373e7cdb2781db3fb8e530dbe2c116ab0810007f513ccfac1a4c231"} Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.466377 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.466773 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xlxd" event={"ID":"6d73442c-9d91-4b5b-b0dd-efc137a55959","Type":"ContainerStarted","Data":"45fabf4277a99d2180ad6afedd56eb75fdf4de846186c002b6a9127577f0eea3"} Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.467042 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/529893f0-2392-4d15-ae52-7f4d1646de55-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"529893f0-2392-4d15-ae52-7f4d1646de55\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.467091 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/529893f0-2392-4d15-ae52-7f4d1646de55-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"529893f0-2392-4d15-ae52-7f4d1646de55\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.467534 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/529893f0-2392-4d15-ae52-7f4d1646de55-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"529893f0-2392-4d15-ae52-7f4d1646de55\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.473488 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"5cc00b73864fd21a1fab521fef2ed2c10bfada1a1b403fffc76df0d0a57eb4bc"} Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.473531 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"1fa60246beb6f0eb27f9e00222fcc09caba8ea1ec45796739bfa1bf5f26ddedd"} Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.488267 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.489918 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/529893f0-2392-4d15-ae52-7f4d1646de55-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"529893f0-2392-4d15-ae52-7f4d1646de55\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.495217 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-45lhd" event={"ID":"ce6956ee-96b9-4fea-af68-90540d249ba8","Type":"ContainerStarted","Data":"572e2fb8dbe764556057151e9be5b3bb9f36f24531defe65e759a3fc0c05e75e"} Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.497917 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"763929ca0f0978664e1d65d55916e40d0e4f880bf3239cdb2c126b0432506dd3"} Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.567974 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.575280 4955 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.575318 4955 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.622529 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8mkhz\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.646666 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-45lhd" podStartSLOduration=11.64664374 podStartE2EDuration="11.64664374s" podCreationTimestamp="2026-02-17 13:06:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:53.581258573 +0000 UTC m=+152.103988116" watchObservedRunningTime="2026-02-17 13:06:53.64664374 +0000 UTC m=+152.169373284" Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.648988 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5kwdn"] Feb 17 13:06:53 crc kubenswrapper[4955]: W0217 13:06:53.686479 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a9eeae4_44ec_4f00_a49f_2af0b006e5cb.slice/crio-ba52cde8153159687724b3fcc5c1bfca99223b16c5c451ec2386361448aa9241 WatchSource:0}: Error finding container ba52cde8153159687724b3fcc5c1bfca99223b16c5c451ec2386361448aa9241: Status 404 returned error can't find the container with id ba52cde8153159687724b3fcc5c1bfca99223b16c5c451ec2386361448aa9241 Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.693183 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.859433 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:53 crc kubenswrapper[4955]: I0217 13:06:53.884582 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 17 13:06:53 crc kubenswrapper[4955]: W0217 13:06:53.895026 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod529893f0_2392_4d15_ae52_7f4d1646de55.slice/crio-d76ce0fa515f984ffa196532da1e4177d2df827133131fb7252c63ab47cd16a3 WatchSource:0}: Error finding container d76ce0fa515f984ffa196532da1e4177d2df827133131fb7252c63ab47cd16a3: Status 404 returned error can't find the container with id d76ce0fa515f984ffa196532da1e4177d2df827133131fb7252c63ab47cd16a3 Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.020601 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5x7zc" Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.020882 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5x7zc" Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.027767 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5x7zc" Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.052761 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mxmt9"] Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.054468 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mxmt9" Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.059093 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.068305 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mxmt9"] Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.145304 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8mkhz"] Feb 17 13:06:54 crc kubenswrapper[4955]: W0217 13:06:54.158527 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3687198_a53d_4a80_baec_bafd1c5ef14d.slice/crio-c072344e5397278b7930e09c1b5a562c0a54495e3ac858fd306420c45604017c WatchSource:0}: Error finding container c072344e5397278b7930e09c1b5a562c0a54495e3ac858fd306420c45604017c: Status 404 returned error can't find the container with id c072344e5397278b7930e09c1b5a562c0a54495e3ac858fd306420c45604017c Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.173582 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/700722b8-137a-4fcb-83bb-3cd0806c676c-utilities\") pod \"redhat-marketplace-mxmt9\" (UID: \"700722b8-137a-4fcb-83bb-3cd0806c676c\") " pod="openshift-marketplace/redhat-marketplace-mxmt9" Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.173637 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/700722b8-137a-4fcb-83bb-3cd0806c676c-catalog-content\") pod \"redhat-marketplace-mxmt9\" (UID: \"700722b8-137a-4fcb-83bb-3cd0806c676c\") " pod="openshift-marketplace/redhat-marketplace-mxmt9" Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.173710 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnckc\" (UniqueName: \"kubernetes.io/projected/700722b8-137a-4fcb-83bb-3cd0806c676c-kube-api-access-rnckc\") pod \"redhat-marketplace-mxmt9\" (UID: \"700722b8-137a-4fcb-83bb-3cd0806c676c\") " pod="openshift-marketplace/redhat-marketplace-mxmt9" Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.233192 4955 patch_prober.go:28] interesting pod/downloads-7954f5f757-qz68q container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.233268 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-qz68q" podUID="01aa5d18-e942-4a3d-8af5-1e41b503f85a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.37:8080/\": dial tcp 10.217.0.37:8080: connect: connection refused" Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.233393 4955 patch_prober.go:28] interesting pod/downloads-7954f5f757-qz68q container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.37:8080/\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.233458 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-qz68q" podUID="01aa5d18-e942-4a3d-8af5-1e41b503f85a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.37:8080/\": dial tcp 10.217.0.37:8080: connect: connection refused" Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.265227 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.275819 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/700722b8-137a-4fcb-83bb-3cd0806c676c-utilities\") pod \"redhat-marketplace-mxmt9\" (UID: \"700722b8-137a-4fcb-83bb-3cd0806c676c\") " pod="openshift-marketplace/redhat-marketplace-mxmt9" Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.275881 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/700722b8-137a-4fcb-83bb-3cd0806c676c-catalog-content\") pod \"redhat-marketplace-mxmt9\" (UID: \"700722b8-137a-4fcb-83bb-3cd0806c676c\") " pod="openshift-marketplace/redhat-marketplace-mxmt9" Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.275981 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnckc\" (UniqueName: \"kubernetes.io/projected/700722b8-137a-4fcb-83bb-3cd0806c676c-kube-api-access-rnckc\") pod \"redhat-marketplace-mxmt9\" (UID: \"700722b8-137a-4fcb-83bb-3cd0806c676c\") " pod="openshift-marketplace/redhat-marketplace-mxmt9" Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.276655 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/700722b8-137a-4fcb-83bb-3cd0806c676c-utilities\") pod \"redhat-marketplace-mxmt9\" (UID: \"700722b8-137a-4fcb-83bb-3cd0806c676c\") " pod="openshift-marketplace/redhat-marketplace-mxmt9" Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.277113 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/700722b8-137a-4fcb-83bb-3cd0806c676c-catalog-content\") pod \"redhat-marketplace-mxmt9\" (UID: \"700722b8-137a-4fcb-83bb-3cd0806c676c\") " pod="openshift-marketplace/redhat-marketplace-mxmt9" Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.285706 4955 patch_prober.go:28] interesting pod/router-default-5444994796-m7gfn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 17 13:06:54 crc kubenswrapper[4955]: [-]has-synced failed: reason withheld Feb 17 13:06:54 crc kubenswrapper[4955]: [+]process-running ok Feb 17 13:06:54 crc kubenswrapper[4955]: healthz check failed Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.285812 4955 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-m7gfn" podUID="cea310be-8fea-43cf-a000-69b614f260e4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.308054 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnckc\" (UniqueName: \"kubernetes.io/projected/700722b8-137a-4fcb-83bb-3cd0806c676c-kube-api-access-rnckc\") pod \"redhat-marketplace-mxmt9\" (UID: \"700722b8-137a-4fcb-83bb-3cd0806c676c\") " pod="openshift-marketplace/redhat-marketplace-mxmt9" Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.397682 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-8dx57" Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.405698 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-8dx57" Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.410003 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mxmt9" Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.452666 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9mrhv"] Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.453769 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9mrhv" Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.466761 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-npb7n" Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.466820 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-npb7n" Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.467085 4955 patch_prober.go:28] interesting pod/console-f9d7485db-npb7n container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.467138 4955 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-npb7n" podUID="e690bd42-0ab1-4b3a-802b-4511117b6fe0" containerName="console" probeResult="failure" output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.500557 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9mrhv"] Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.541839 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"529893f0-2392-4d15-ae52-7f4d1646de55","Type":"ContainerStarted","Data":"b708605249b124cadd8d861e7ba9c70bbafce276ca209038a0070d209b1bea06"} Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.541909 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"529893f0-2392-4d15-ae52-7f4d1646de55","Type":"ContainerStarted","Data":"d76ce0fa515f984ffa196532da1e4177d2df827133131fb7252c63ab47cd16a3"} Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.549424 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" event={"ID":"e3687198-a53d-4a80-baec-bafd1c5ef14d","Type":"ContainerStarted","Data":"0cb73d5edcc4e737199a46ba7e9ae93d7ebbe1acbe6100bcafb4d31958bd04f4"} Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.549482 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" event={"ID":"e3687198-a53d-4a80-baec-bafd1c5ef14d","Type":"ContainerStarted","Data":"c072344e5397278b7930e09c1b5a562c0a54495e3ac858fd306420c45604017c"} Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.549601 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.552013 4955 generic.go:334] "Generic (PLEG): container finished" podID="7a9eeae4-44ec-4f00-a49f-2af0b006e5cb" containerID="3a20d1087687a07f7f544aab5b1a406c0667b4adc019b4ac5459239b9a668f04" exitCode=0 Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.552094 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5kwdn" event={"ID":"7a9eeae4-44ec-4f00-a49f-2af0b006e5cb","Type":"ContainerDied","Data":"3a20d1087687a07f7f544aab5b1a406c0667b4adc019b4ac5459239b9a668f04"} Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.552137 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5kwdn" event={"ID":"7a9eeae4-44ec-4f00-a49f-2af0b006e5cb","Type":"ContainerStarted","Data":"ba52cde8153159687724b3fcc5c1bfca99223b16c5c451ec2386361448aa9241"} Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.566560 4955 generic.go:334] "Generic (PLEG): container finished" podID="d1ff2d84-9dcc-43f3-821f-989fbeec8e93" containerID="9d29fd01df2f747e6753f6a518a53d94be5d5d392283929731819e995ae462dc" exitCode=0 Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.567075 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f8t7f" event={"ID":"d1ff2d84-9dcc-43f3-821f-989fbeec8e93","Type":"ContainerDied","Data":"9d29fd01df2f747e6753f6a518a53d94be5d5d392283929731819e995ae462dc"} Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.567162 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f8t7f" event={"ID":"d1ff2d84-9dcc-43f3-821f-989fbeec8e93","Type":"ContainerStarted","Data":"4ec34fa41d133037332717c6813a39a4521a4cf4290d290277655f78abed90fc"} Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.574353 4955 generic.go:334] "Generic (PLEG): container finished" podID="6d73442c-9d91-4b5b-b0dd-efc137a55959" containerID="c973f82be5e4407d035f3ecc76b729686a60f36aede147e0901b6827226bc268" exitCode=0 Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.575901 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xlxd" event={"ID":"6d73442c-9d91-4b5b-b0dd-efc137a55959","Type":"ContainerDied","Data":"c973f82be5e4407d035f3ecc76b729686a60f36aede147e0901b6827226bc268"} Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.583599 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=1.583576131 podStartE2EDuration="1.583576131s" podCreationTimestamp="2026-02-17 13:06:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:54.583131627 +0000 UTC m=+153.105861170" watchObservedRunningTime="2026-02-17 13:06:54.583576131 +0000 UTC m=+153.106305684" Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.584245 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b4cdbfa-dd6b-4586-9127-23b2605d94b4-utilities\") pod \"redhat-marketplace-9mrhv\" (UID: \"6b4cdbfa-dd6b-4586-9127-23b2605d94b4\") " pod="openshift-marketplace/redhat-marketplace-9mrhv" Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.584332 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n99kp\" (UniqueName: \"kubernetes.io/projected/6b4cdbfa-dd6b-4586-9127-23b2605d94b4-kube-api-access-n99kp\") pod \"redhat-marketplace-9mrhv\" (UID: \"6b4cdbfa-dd6b-4586-9127-23b2605d94b4\") " pod="openshift-marketplace/redhat-marketplace-9mrhv" Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.584360 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b4cdbfa-dd6b-4586-9127-23b2605d94b4-catalog-content\") pod \"redhat-marketplace-9mrhv\" (UID: \"6b4cdbfa-dd6b-4586-9127-23b2605d94b4\") " pod="openshift-marketplace/redhat-marketplace-9mrhv" Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.599219 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5x7zc" Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.655571 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" podStartSLOduration=131.655549824 podStartE2EDuration="2m11.655549824s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:06:54.653295887 +0000 UTC m=+153.176025430" watchObservedRunningTime="2026-02-17 13:06:54.655549824 +0000 UTC m=+153.178279367" Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.685825 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b4cdbfa-dd6b-4586-9127-23b2605d94b4-utilities\") pod \"redhat-marketplace-9mrhv\" (UID: \"6b4cdbfa-dd6b-4586-9127-23b2605d94b4\") " pod="openshift-marketplace/redhat-marketplace-9mrhv" Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.685899 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n99kp\" (UniqueName: \"kubernetes.io/projected/6b4cdbfa-dd6b-4586-9127-23b2605d94b4-kube-api-access-n99kp\") pod \"redhat-marketplace-9mrhv\" (UID: \"6b4cdbfa-dd6b-4586-9127-23b2605d94b4\") " pod="openshift-marketplace/redhat-marketplace-9mrhv" Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.685918 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b4cdbfa-dd6b-4586-9127-23b2605d94b4-catalog-content\") pod \"redhat-marketplace-9mrhv\" (UID: \"6b4cdbfa-dd6b-4586-9127-23b2605d94b4\") " pod="openshift-marketplace/redhat-marketplace-9mrhv" Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.687173 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b4cdbfa-dd6b-4586-9127-23b2605d94b4-catalog-content\") pod \"redhat-marketplace-9mrhv\" (UID: \"6b4cdbfa-dd6b-4586-9127-23b2605d94b4\") " pod="openshift-marketplace/redhat-marketplace-9mrhv" Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.687690 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b4cdbfa-dd6b-4586-9127-23b2605d94b4-utilities\") pod \"redhat-marketplace-9mrhv\" (UID: \"6b4cdbfa-dd6b-4586-9127-23b2605d94b4\") " pod="openshift-marketplace/redhat-marketplace-9mrhv" Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.721508 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n99kp\" (UniqueName: \"kubernetes.io/projected/6b4cdbfa-dd6b-4586-9127-23b2605d94b4-kube-api-access-n99kp\") pod \"redhat-marketplace-9mrhv\" (UID: \"6b4cdbfa-dd6b-4586-9127-23b2605d94b4\") " pod="openshift-marketplace/redhat-marketplace-9mrhv" Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.772343 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9mrhv" Feb 17 13:06:54 crc kubenswrapper[4955]: I0217 13:06:54.925913 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mxmt9"] Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.115160 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9mrhv"] Feb 17 13:06:55 crc kubenswrapper[4955]: W0217 13:06:55.138008 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6b4cdbfa_dd6b_4586_9127_23b2605d94b4.slice/crio-4af0a8de45aed09f11485d2c269d8ac3002558203f61670af50d44c43a191b73 WatchSource:0}: Error finding container 4af0a8de45aed09f11485d2c269d8ac3002558203f61670af50d44c43a191b73: Status 404 returned error can't find the container with id 4af0a8de45aed09f11485d2c269d8ac3002558203f61670af50d44c43a191b73 Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.189246 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.245334 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7cjtv"] Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.246706 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7cjtv" Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.249737 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.258146 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7cjtv"] Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.281050 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-m7gfn" Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.287984 4955 patch_prober.go:28] interesting pod/router-default-5444994796-m7gfn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 17 13:06:55 crc kubenswrapper[4955]: [-]has-synced failed: reason withheld Feb 17 13:06:55 crc kubenswrapper[4955]: [+]process-running ok Feb 17 13:06:55 crc kubenswrapper[4955]: healthz check failed Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.288051 4955 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-m7gfn" podUID="cea310be-8fea-43cf-a000-69b614f260e4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.308601 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-f7ccb" Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.356015 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-9c9zp" Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.377263 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z7bll" Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.405150 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z7bll" Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.408506 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b3b3288-6b3f-4a48-ae9a-f087715c7e47-catalog-content\") pod \"redhat-operators-7cjtv\" (UID: \"3b3b3288-6b3f-4a48-ae9a-f087715c7e47\") " pod="openshift-marketplace/redhat-operators-7cjtv" Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.408549 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hfx7\" (UniqueName: \"kubernetes.io/projected/3b3b3288-6b3f-4a48-ae9a-f087715c7e47-kube-api-access-2hfx7\") pod \"redhat-operators-7cjtv\" (UID: \"3b3b3288-6b3f-4a48-ae9a-f087715c7e47\") " pod="openshift-marketplace/redhat-operators-7cjtv" Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.408583 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b3b3288-6b3f-4a48-ae9a-f087715c7e47-utilities\") pod \"redhat-operators-7cjtv\" (UID: \"3b3b3288-6b3f-4a48-ae9a-f087715c7e47\") " pod="openshift-marketplace/redhat-operators-7cjtv" Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.509928 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b3b3288-6b3f-4a48-ae9a-f087715c7e47-catalog-content\") pod \"redhat-operators-7cjtv\" (UID: \"3b3b3288-6b3f-4a48-ae9a-f087715c7e47\") " pod="openshift-marketplace/redhat-operators-7cjtv" Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.509968 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hfx7\" (UniqueName: \"kubernetes.io/projected/3b3b3288-6b3f-4a48-ae9a-f087715c7e47-kube-api-access-2hfx7\") pod \"redhat-operators-7cjtv\" (UID: \"3b3b3288-6b3f-4a48-ae9a-f087715c7e47\") " pod="openshift-marketplace/redhat-operators-7cjtv" Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.509998 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b3b3288-6b3f-4a48-ae9a-f087715c7e47-utilities\") pod \"redhat-operators-7cjtv\" (UID: \"3b3b3288-6b3f-4a48-ae9a-f087715c7e47\") " pod="openshift-marketplace/redhat-operators-7cjtv" Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.510510 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b3b3288-6b3f-4a48-ae9a-f087715c7e47-utilities\") pod \"redhat-operators-7cjtv\" (UID: \"3b3b3288-6b3f-4a48-ae9a-f087715c7e47\") " pod="openshift-marketplace/redhat-operators-7cjtv" Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.513989 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b3b3288-6b3f-4a48-ae9a-f087715c7e47-catalog-content\") pod \"redhat-operators-7cjtv\" (UID: \"3b3b3288-6b3f-4a48-ae9a-f087715c7e47\") " pod="openshift-marketplace/redhat-operators-7cjtv" Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.555646 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hfx7\" (UniqueName: \"kubernetes.io/projected/3b3b3288-6b3f-4a48-ae9a-f087715c7e47-kube-api-access-2hfx7\") pod \"redhat-operators-7cjtv\" (UID: \"3b3b3288-6b3f-4a48-ae9a-f087715c7e47\") " pod="openshift-marketplace/redhat-operators-7cjtv" Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.609657 4955 generic.go:334] "Generic (PLEG): container finished" podID="529893f0-2392-4d15-ae52-7f4d1646de55" containerID="b708605249b124cadd8d861e7ba9c70bbafce276ca209038a0070d209b1bea06" exitCode=0 Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.609768 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"529893f0-2392-4d15-ae52-7f4d1646de55","Type":"ContainerDied","Data":"b708605249b124cadd8d861e7ba9c70bbafce276ca209038a0070d209b1bea06"} Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.618376 4955 generic.go:334] "Generic (PLEG): container finished" podID="700722b8-137a-4fcb-83bb-3cd0806c676c" containerID="a876575f7716e64e2b9eb9fa534c5536269b1a2555b420ab2046d4e281b3fe80" exitCode=0 Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.620932 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mxmt9" event={"ID":"700722b8-137a-4fcb-83bb-3cd0806c676c","Type":"ContainerDied","Data":"a876575f7716e64e2b9eb9fa534c5536269b1a2555b420ab2046d4e281b3fe80"} Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.620999 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mxmt9" event={"ID":"700722b8-137a-4fcb-83bb-3cd0806c676c","Type":"ContainerStarted","Data":"7ebd289f4ccb08edcba5ca62d06e44c968ea5e89217e8c8ab97af99902390498"} Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.636934 4955 generic.go:334] "Generic (PLEG): container finished" podID="6b4cdbfa-dd6b-4586-9127-23b2605d94b4" containerID="263a9bfa1557ea4609f633f59559743f457fba14eb0d627cdf702bc80ae8ab93" exitCode=0 Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.638464 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9mrhv" event={"ID":"6b4cdbfa-dd6b-4586-9127-23b2605d94b4","Type":"ContainerDied","Data":"263a9bfa1557ea4609f633f59559743f457fba14eb0d627cdf702bc80ae8ab93"} Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.638506 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9mrhv" event={"ID":"6b4cdbfa-dd6b-4586-9127-23b2605d94b4","Type":"ContainerStarted","Data":"4af0a8de45aed09f11485d2c269d8ac3002558203f61670af50d44c43a191b73"} Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.667050 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7cjtv" Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.670627 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8c8cn"] Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.671950 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8c8cn" Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.680853 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8c8cn"] Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.814330 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bd63c82-a771-47d1-8ba4-a4036d456e0a-catalog-content\") pod \"redhat-operators-8c8cn\" (UID: \"0bd63c82-a771-47d1-8ba4-a4036d456e0a\") " pod="openshift-marketplace/redhat-operators-8c8cn" Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.814401 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5r6zt\" (UniqueName: \"kubernetes.io/projected/0bd63c82-a771-47d1-8ba4-a4036d456e0a-kube-api-access-5r6zt\") pod \"redhat-operators-8c8cn\" (UID: \"0bd63c82-a771-47d1-8ba4-a4036d456e0a\") " pod="openshift-marketplace/redhat-operators-8c8cn" Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.814430 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bd63c82-a771-47d1-8ba4-a4036d456e0a-utilities\") pod \"redhat-operators-8c8cn\" (UID: \"0bd63c82-a771-47d1-8ba4-a4036d456e0a\") " pod="openshift-marketplace/redhat-operators-8c8cn" Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.916563 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5r6zt\" (UniqueName: \"kubernetes.io/projected/0bd63c82-a771-47d1-8ba4-a4036d456e0a-kube-api-access-5r6zt\") pod \"redhat-operators-8c8cn\" (UID: \"0bd63c82-a771-47d1-8ba4-a4036d456e0a\") " pod="openshift-marketplace/redhat-operators-8c8cn" Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.916699 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bd63c82-a771-47d1-8ba4-a4036d456e0a-utilities\") pod \"redhat-operators-8c8cn\" (UID: \"0bd63c82-a771-47d1-8ba4-a4036d456e0a\") " pod="openshift-marketplace/redhat-operators-8c8cn" Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.916861 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bd63c82-a771-47d1-8ba4-a4036d456e0a-catalog-content\") pod \"redhat-operators-8c8cn\" (UID: \"0bd63c82-a771-47d1-8ba4-a4036d456e0a\") " pod="openshift-marketplace/redhat-operators-8c8cn" Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.917761 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bd63c82-a771-47d1-8ba4-a4036d456e0a-catalog-content\") pod \"redhat-operators-8c8cn\" (UID: \"0bd63c82-a771-47d1-8ba4-a4036d456e0a\") " pod="openshift-marketplace/redhat-operators-8c8cn" Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.917843 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bd63c82-a771-47d1-8ba4-a4036d456e0a-utilities\") pod \"redhat-operators-8c8cn\" (UID: \"0bd63c82-a771-47d1-8ba4-a4036d456e0a\") " pod="openshift-marketplace/redhat-operators-8c8cn" Feb 17 13:06:55 crc kubenswrapper[4955]: I0217 13:06:55.945828 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5r6zt\" (UniqueName: \"kubernetes.io/projected/0bd63c82-a771-47d1-8ba4-a4036d456e0a-kube-api-access-5r6zt\") pod \"redhat-operators-8c8cn\" (UID: \"0bd63c82-a771-47d1-8ba4-a4036d456e0a\") " pod="openshift-marketplace/redhat-operators-8c8cn" Feb 17 13:06:56 crc kubenswrapper[4955]: I0217 13:06:56.010327 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8c8cn" Feb 17 13:06:56 crc kubenswrapper[4955]: I0217 13:06:56.286189 4955 patch_prober.go:28] interesting pod/router-default-5444994796-m7gfn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 17 13:06:56 crc kubenswrapper[4955]: [-]has-synced failed: reason withheld Feb 17 13:06:56 crc kubenswrapper[4955]: [+]process-running ok Feb 17 13:06:56 crc kubenswrapper[4955]: healthz check failed Feb 17 13:06:56 crc kubenswrapper[4955]: I0217 13:06:56.286264 4955 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-m7gfn" podUID="cea310be-8fea-43cf-a000-69b614f260e4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 17 13:06:56 crc kubenswrapper[4955]: I0217 13:06:56.300423 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7cjtv"] Feb 17 13:06:56 crc kubenswrapper[4955]: I0217 13:06:56.419586 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8c8cn"] Feb 17 13:06:56 crc kubenswrapper[4955]: I0217 13:06:56.653038 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7cjtv" event={"ID":"3b3b3288-6b3f-4a48-ae9a-f087715c7e47","Type":"ContainerStarted","Data":"848d77e9aa3904615c7e41f94882fc61e4734894e3f404dd43a07d0d4929fd6f"} Feb 17 13:06:56 crc kubenswrapper[4955]: I0217 13:06:56.664045 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8c8cn" event={"ID":"0bd63c82-a771-47d1-8ba4-a4036d456e0a","Type":"ContainerStarted","Data":"168a75ffcf3acea83ca2de33c238e20282a74cfefdcffa122a401b6a9fcf6794"} Feb 17 13:06:56 crc kubenswrapper[4955]: I0217 13:06:56.672606 4955 generic.go:334] "Generic (PLEG): container finished" podID="93fe04d3-c430-4ed6-ae04-312cf0613834" containerID="a32124e8d3b973e4e5f3f6e8a4c8205bc70e8fba11e9335172e8be548ef9d4cb" exitCode=0 Feb 17 13:06:56 crc kubenswrapper[4955]: I0217 13:06:56.674739 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29522220-v5twm" event={"ID":"93fe04d3-c430-4ed6-ae04-312cf0613834","Type":"ContainerDied","Data":"a32124e8d3b973e4e5f3f6e8a4c8205bc70e8fba11e9335172e8be548ef9d4cb"} Feb 17 13:06:56 crc kubenswrapper[4955]: I0217 13:06:56.972948 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 17 13:06:57 crc kubenswrapper[4955]: I0217 13:06:57.036585 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/529893f0-2392-4d15-ae52-7f4d1646de55-kube-api-access\") pod \"529893f0-2392-4d15-ae52-7f4d1646de55\" (UID: \"529893f0-2392-4d15-ae52-7f4d1646de55\") " Feb 17 13:06:57 crc kubenswrapper[4955]: I0217 13:06:57.036668 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/529893f0-2392-4d15-ae52-7f4d1646de55-kubelet-dir\") pod \"529893f0-2392-4d15-ae52-7f4d1646de55\" (UID: \"529893f0-2392-4d15-ae52-7f4d1646de55\") " Feb 17 13:06:57 crc kubenswrapper[4955]: I0217 13:06:57.037121 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/529893f0-2392-4d15-ae52-7f4d1646de55-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "529893f0-2392-4d15-ae52-7f4d1646de55" (UID: "529893f0-2392-4d15-ae52-7f4d1646de55"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 13:06:57 crc kubenswrapper[4955]: I0217 13:06:57.047850 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/529893f0-2392-4d15-ae52-7f4d1646de55-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "529893f0-2392-4d15-ae52-7f4d1646de55" (UID: "529893f0-2392-4d15-ae52-7f4d1646de55"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:06:57 crc kubenswrapper[4955]: I0217 13:06:57.138340 4955 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/529893f0-2392-4d15-ae52-7f4d1646de55-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 17 13:06:57 crc kubenswrapper[4955]: I0217 13:06:57.138374 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/529893f0-2392-4d15-ae52-7f4d1646de55-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 17 13:06:57 crc kubenswrapper[4955]: I0217 13:06:57.300116 4955 patch_prober.go:28] interesting pod/router-default-5444994796-m7gfn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 17 13:06:57 crc kubenswrapper[4955]: [-]has-synced failed: reason withheld Feb 17 13:06:57 crc kubenswrapper[4955]: [+]process-running ok Feb 17 13:06:57 crc kubenswrapper[4955]: healthz check failed Feb 17 13:06:57 crc kubenswrapper[4955]: I0217 13:06:57.300568 4955 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-m7gfn" podUID="cea310be-8fea-43cf-a000-69b614f260e4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 17 13:06:57 crc kubenswrapper[4955]: I0217 13:06:57.381416 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-64cqc" Feb 17 13:06:57 crc kubenswrapper[4955]: I0217 13:06:57.733823 4955 generic.go:334] "Generic (PLEG): container finished" podID="0bd63c82-a771-47d1-8ba4-a4036d456e0a" containerID="ef9ddd759d5fd2c654467d9c2894226dc3b4b44d11432dccc92b98b8a3450ddc" exitCode=0 Feb 17 13:06:57 crc kubenswrapper[4955]: I0217 13:06:57.733891 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8c8cn" event={"ID":"0bd63c82-a771-47d1-8ba4-a4036d456e0a","Type":"ContainerDied","Data":"ef9ddd759d5fd2c654467d9c2894226dc3b4b44d11432dccc92b98b8a3450ddc"} Feb 17 13:06:57 crc kubenswrapper[4955]: I0217 13:06:57.746297 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 17 13:06:57 crc kubenswrapper[4955]: I0217 13:06:57.746293 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"529893f0-2392-4d15-ae52-7f4d1646de55","Type":"ContainerDied","Data":"d76ce0fa515f984ffa196532da1e4177d2df827133131fb7252c63ab47cd16a3"} Feb 17 13:06:57 crc kubenswrapper[4955]: I0217 13:06:57.746468 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d76ce0fa515f984ffa196532da1e4177d2df827133131fb7252c63ab47cd16a3" Feb 17 13:06:57 crc kubenswrapper[4955]: I0217 13:06:57.754339 4955 generic.go:334] "Generic (PLEG): container finished" podID="3b3b3288-6b3f-4a48-ae9a-f087715c7e47" containerID="7d89db3dc6759750846d4e0d9a51140f2f4b1db903871367b498b94a29987160" exitCode=0 Feb 17 13:06:57 crc kubenswrapper[4955]: I0217 13:06:57.754532 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7cjtv" event={"ID":"3b3b3288-6b3f-4a48-ae9a-f087715c7e47","Type":"ContainerDied","Data":"7d89db3dc6759750846d4e0d9a51140f2f4b1db903871367b498b94a29987160"} Feb 17 13:06:58 crc kubenswrapper[4955]: I0217 13:06:58.145423 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29522220-v5twm" Feb 17 13:06:58 crc kubenswrapper[4955]: I0217 13:06:58.170250 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/93fe04d3-c430-4ed6-ae04-312cf0613834-config-volume\") pod \"93fe04d3-c430-4ed6-ae04-312cf0613834\" (UID: \"93fe04d3-c430-4ed6-ae04-312cf0613834\") " Feb 17 13:06:58 crc kubenswrapper[4955]: I0217 13:06:58.170800 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/93fe04d3-c430-4ed6-ae04-312cf0613834-secret-volume\") pod \"93fe04d3-c430-4ed6-ae04-312cf0613834\" (UID: \"93fe04d3-c430-4ed6-ae04-312cf0613834\") " Feb 17 13:06:58 crc kubenswrapper[4955]: I0217 13:06:58.170898 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98xl5\" (UniqueName: \"kubernetes.io/projected/93fe04d3-c430-4ed6-ae04-312cf0613834-kube-api-access-98xl5\") pod \"93fe04d3-c430-4ed6-ae04-312cf0613834\" (UID: \"93fe04d3-c430-4ed6-ae04-312cf0613834\") " Feb 17 13:06:58 crc kubenswrapper[4955]: I0217 13:06:58.172523 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93fe04d3-c430-4ed6-ae04-312cf0613834-config-volume" (OuterVolumeSpecName: "config-volume") pod "93fe04d3-c430-4ed6-ae04-312cf0613834" (UID: "93fe04d3-c430-4ed6-ae04-312cf0613834"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:06:58 crc kubenswrapper[4955]: I0217 13:06:58.190679 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93fe04d3-c430-4ed6-ae04-312cf0613834-kube-api-access-98xl5" (OuterVolumeSpecName: "kube-api-access-98xl5") pod "93fe04d3-c430-4ed6-ae04-312cf0613834" (UID: "93fe04d3-c430-4ed6-ae04-312cf0613834"). InnerVolumeSpecName "kube-api-access-98xl5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:06:58 crc kubenswrapper[4955]: I0217 13:06:58.190832 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93fe04d3-c430-4ed6-ae04-312cf0613834-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "93fe04d3-c430-4ed6-ae04-312cf0613834" (UID: "93fe04d3-c430-4ed6-ae04-312cf0613834"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:06:58 crc kubenswrapper[4955]: I0217 13:06:58.272420 4955 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/93fe04d3-c430-4ed6-ae04-312cf0613834-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 17 13:06:58 crc kubenswrapper[4955]: I0217 13:06:58.272473 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98xl5\" (UniqueName: \"kubernetes.io/projected/93fe04d3-c430-4ed6-ae04-312cf0613834-kube-api-access-98xl5\") on node \"crc\" DevicePath \"\"" Feb 17 13:06:58 crc kubenswrapper[4955]: I0217 13:06:58.272486 4955 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/93fe04d3-c430-4ed6-ae04-312cf0613834-config-volume\") on node \"crc\" DevicePath \"\"" Feb 17 13:06:58 crc kubenswrapper[4955]: I0217 13:06:58.284116 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-m7gfn" Feb 17 13:06:58 crc kubenswrapper[4955]: I0217 13:06:58.286569 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-m7gfn" Feb 17 13:06:58 crc kubenswrapper[4955]: I0217 13:06:58.485544 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 17 13:06:58 crc kubenswrapper[4955]: E0217 13:06:58.485793 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="529893f0-2392-4d15-ae52-7f4d1646de55" containerName="pruner" Feb 17 13:06:58 crc kubenswrapper[4955]: I0217 13:06:58.485807 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="529893f0-2392-4d15-ae52-7f4d1646de55" containerName="pruner" Feb 17 13:06:58 crc kubenswrapper[4955]: E0217 13:06:58.485828 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93fe04d3-c430-4ed6-ae04-312cf0613834" containerName="collect-profiles" Feb 17 13:06:58 crc kubenswrapper[4955]: I0217 13:06:58.485834 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="93fe04d3-c430-4ed6-ae04-312cf0613834" containerName="collect-profiles" Feb 17 13:06:58 crc kubenswrapper[4955]: I0217 13:06:58.485927 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="529893f0-2392-4d15-ae52-7f4d1646de55" containerName="pruner" Feb 17 13:06:58 crc kubenswrapper[4955]: I0217 13:06:58.485945 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="93fe04d3-c430-4ed6-ae04-312cf0613834" containerName="collect-profiles" Feb 17 13:06:58 crc kubenswrapper[4955]: I0217 13:06:58.486352 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 17 13:06:58 crc kubenswrapper[4955]: I0217 13:06:58.489989 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 17 13:06:58 crc kubenswrapper[4955]: I0217 13:06:58.506867 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 17 13:06:58 crc kubenswrapper[4955]: I0217 13:06:58.514525 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 17 13:06:58 crc kubenswrapper[4955]: I0217 13:06:58.578805 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f2cf9078-0c99-4e54-a4a0-c8d779b44c93-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"f2cf9078-0c99-4e54-a4a0-c8d779b44c93\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 17 13:06:58 crc kubenswrapper[4955]: I0217 13:06:58.578940 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f2cf9078-0c99-4e54-a4a0-c8d779b44c93-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"f2cf9078-0c99-4e54-a4a0-c8d779b44c93\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 17 13:06:58 crc kubenswrapper[4955]: I0217 13:06:58.680832 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f2cf9078-0c99-4e54-a4a0-c8d779b44c93-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"f2cf9078-0c99-4e54-a4a0-c8d779b44c93\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 17 13:06:58 crc kubenswrapper[4955]: I0217 13:06:58.683435 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f2cf9078-0c99-4e54-a4a0-c8d779b44c93-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"f2cf9078-0c99-4e54-a4a0-c8d779b44c93\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 17 13:06:58 crc kubenswrapper[4955]: I0217 13:06:58.683656 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f2cf9078-0c99-4e54-a4a0-c8d779b44c93-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"f2cf9078-0c99-4e54-a4a0-c8d779b44c93\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 17 13:06:58 crc kubenswrapper[4955]: I0217 13:06:58.707143 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f2cf9078-0c99-4e54-a4a0-c8d779b44c93-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"f2cf9078-0c99-4e54-a4a0-c8d779b44c93\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 17 13:06:58 crc kubenswrapper[4955]: I0217 13:06:58.787648 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29522220-v5twm" event={"ID":"93fe04d3-c430-4ed6-ae04-312cf0613834","Type":"ContainerDied","Data":"7283f9671ff1028392d197ada4636103257b37048f7bd0019d1e4c6209f7e59c"} Feb 17 13:06:58 crc kubenswrapper[4955]: I0217 13:06:58.787714 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7283f9671ff1028392d197ada4636103257b37048f7bd0019d1e4c6209f7e59c" Feb 17 13:06:58 crc kubenswrapper[4955]: I0217 13:06:58.787852 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29522220-v5twm" Feb 17 13:06:58 crc kubenswrapper[4955]: I0217 13:06:58.819994 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 17 13:06:59 crc kubenswrapper[4955]: I0217 13:06:59.618716 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 17 13:06:59 crc kubenswrapper[4955]: I0217 13:06:59.622826 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:06:59 crc kubenswrapper[4955]: I0217 13:06:59.798117 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"f2cf9078-0c99-4e54-a4a0-c8d779b44c93","Type":"ContainerStarted","Data":"342ebec8da543032f7e9ee7c21a15327ecf1a09f7330d3b90a6f53ee054faa48"} Feb 17 13:07:01 crc kubenswrapper[4955]: I0217 13:07:01.832082 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"f2cf9078-0c99-4e54-a4a0-c8d779b44c93","Type":"ContainerStarted","Data":"d1b6a434e14a8abc797430785d8f9c636a642620c4a158a95a0d0915cd5f4652"} Feb 17 13:07:01 crc kubenswrapper[4955]: I0217 13:07:01.850890 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=3.850872446 podStartE2EDuration="3.850872446s" podCreationTimestamp="2026-02-17 13:06:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:07:01.84896985 +0000 UTC m=+160.371699403" watchObservedRunningTime="2026-02-17 13:07:01.850872446 +0000 UTC m=+160.373601989" Feb 17 13:07:02 crc kubenswrapper[4955]: I0217 13:07:02.859442 4955 generic.go:334] "Generic (PLEG): container finished" podID="f2cf9078-0c99-4e54-a4a0-c8d779b44c93" containerID="d1b6a434e14a8abc797430785d8f9c636a642620c4a158a95a0d0915cd5f4652" exitCode=0 Feb 17 13:07:02 crc kubenswrapper[4955]: I0217 13:07:02.859529 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"f2cf9078-0c99-4e54-a4a0-c8d779b44c93","Type":"ContainerDied","Data":"d1b6a434e14a8abc797430785d8f9c636a642620c4a158a95a0d0915cd5f4652"} Feb 17 13:07:04 crc kubenswrapper[4955]: I0217 13:07:04.247897 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-qz68q" Feb 17 13:07:04 crc kubenswrapper[4955]: I0217 13:07:04.470187 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-npb7n" Feb 17 13:07:04 crc kubenswrapper[4955]: I0217 13:07:04.474930 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-npb7n" Feb 17 13:07:04 crc kubenswrapper[4955]: I0217 13:07:04.575496 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 13:07:04 crc kubenswrapper[4955]: I0217 13:07:04.575587 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 13:07:06 crc kubenswrapper[4955]: I0217 13:07:06.625006 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/77480847-3271-4161-a833-e62b3bb4413e-metrics-certs\") pod \"network-metrics-daemon-2k5pz\" (UID: \"77480847-3271-4161-a833-e62b3bb4413e\") " pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:07:06 crc kubenswrapper[4955]: I0217 13:07:06.634173 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/77480847-3271-4161-a833-e62b3bb4413e-metrics-certs\") pod \"network-metrics-daemon-2k5pz\" (UID: \"77480847-3271-4161-a833-e62b3bb4413e\") " pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:07:06 crc kubenswrapper[4955]: I0217 13:07:06.864830 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2k5pz" Feb 17 13:07:11 crc kubenswrapper[4955]: I0217 13:07:11.621202 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 17 13:07:11 crc kubenswrapper[4955]: I0217 13:07:11.807363 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f2cf9078-0c99-4e54-a4a0-c8d779b44c93-kubelet-dir\") pod \"f2cf9078-0c99-4e54-a4a0-c8d779b44c93\" (UID: \"f2cf9078-0c99-4e54-a4a0-c8d779b44c93\") " Feb 17 13:07:11 crc kubenswrapper[4955]: I0217 13:07:11.807437 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f2cf9078-0c99-4e54-a4a0-c8d779b44c93-kube-api-access\") pod \"f2cf9078-0c99-4e54-a4a0-c8d779b44c93\" (UID: \"f2cf9078-0c99-4e54-a4a0-c8d779b44c93\") " Feb 17 13:07:11 crc kubenswrapper[4955]: I0217 13:07:11.807623 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f2cf9078-0c99-4e54-a4a0-c8d779b44c93-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "f2cf9078-0c99-4e54-a4a0-c8d779b44c93" (UID: "f2cf9078-0c99-4e54-a4a0-c8d779b44c93"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 13:07:11 crc kubenswrapper[4955]: I0217 13:07:11.807826 4955 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f2cf9078-0c99-4e54-a4a0-c8d779b44c93-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 17 13:07:11 crc kubenswrapper[4955]: I0217 13:07:11.816020 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2cf9078-0c99-4e54-a4a0-c8d779b44c93-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "f2cf9078-0c99-4e54-a4a0-c8d779b44c93" (UID: "f2cf9078-0c99-4e54-a4a0-c8d779b44c93"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:07:11 crc kubenswrapper[4955]: I0217 13:07:11.909794 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f2cf9078-0c99-4e54-a4a0-c8d779b44c93-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 17 13:07:11 crc kubenswrapper[4955]: I0217 13:07:11.921433 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"f2cf9078-0c99-4e54-a4a0-c8d779b44c93","Type":"ContainerDied","Data":"342ebec8da543032f7e9ee7c21a15327ecf1a09f7330d3b90a6f53ee054faa48"} Feb 17 13:07:11 crc kubenswrapper[4955]: I0217 13:07:11.921479 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="342ebec8da543032f7e9ee7c21a15327ecf1a09f7330d3b90a6f53ee054faa48" Feb 17 13:07:11 crc kubenswrapper[4955]: I0217 13:07:11.921551 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 17 13:07:13 crc kubenswrapper[4955]: I0217 13:07:13.864866 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:07:24 crc kubenswrapper[4955]: I0217 13:07:24.676506 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5264p" Feb 17 13:07:25 crc kubenswrapper[4955]: E0217 13:07:25.282613 4955 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 17 13:07:25 crc kubenswrapper[4955]: E0217 13:07:25.283170 4955 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n99kp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-9mrhv_openshift-marketplace(6b4cdbfa-dd6b-4586-9127-23b2605d94b4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 17 13:07:25 crc kubenswrapper[4955]: E0217 13:07:25.284560 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-9mrhv" podUID="6b4cdbfa-dd6b-4586-9127-23b2605d94b4" Feb 17 13:07:28 crc kubenswrapper[4955]: E0217 13:07:28.380970 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-9mrhv" podUID="6b4cdbfa-dd6b-4586-9127-23b2605d94b4" Feb 17 13:07:28 crc kubenswrapper[4955]: E0217 13:07:28.536011 4955 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 17 13:07:28 crc kubenswrapper[4955]: E0217 13:07:28.536187 4955 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rnckc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-mxmt9_openshift-marketplace(700722b8-137a-4fcb-83bb-3cd0806c676c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 17 13:07:28 crc kubenswrapper[4955]: E0217 13:07:28.537712 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-mxmt9" podUID="700722b8-137a-4fcb-83bb-3cd0806c676c" Feb 17 13:07:28 crc kubenswrapper[4955]: E0217 13:07:28.617040 4955 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 17 13:07:28 crc kubenswrapper[4955]: E0217 13:07:28.617544 4955 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7vj64,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-6xlxd_openshift-marketplace(6d73442c-9d91-4b5b-b0dd-efc137a55959): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 17 13:07:28 crc kubenswrapper[4955]: E0217 13:07:28.618740 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-6xlxd" podUID="6d73442c-9d91-4b5b-b0dd-efc137a55959" Feb 17 13:07:28 crc kubenswrapper[4955]: I0217 13:07:28.903713 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-2k5pz"] Feb 17 13:07:28 crc kubenswrapper[4955]: W0217 13:07:28.913951 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod77480847_3271_4161_a833_e62b3bb4413e.slice/crio-1cf1cb5bbb59a16da122559224679e367e71ec4f908d46a513164972bf3b11bf WatchSource:0}: Error finding container 1cf1cb5bbb59a16da122559224679e367e71ec4f908d46a513164972bf3b11bf: Status 404 returned error can't find the container with id 1cf1cb5bbb59a16da122559224679e367e71ec4f908d46a513164972bf3b11bf Feb 17 13:07:29 crc kubenswrapper[4955]: I0217 13:07:29.024975 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5kwdn" event={"ID":"7a9eeae4-44ec-4f00-a49f-2af0b006e5cb","Type":"ContainerStarted","Data":"d9b3be727e3c8fe306a4a17f4efb2478e2b8969d98a3083a5371a88d252a7185"} Feb 17 13:07:29 crc kubenswrapper[4955]: I0217 13:07:29.029259 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f8t7f" event={"ID":"d1ff2d84-9dcc-43f3-821f-989fbeec8e93","Type":"ContainerStarted","Data":"9fff0794b0a3f8ac54d4b89ee18e2397711499eadc31e7e20fd1d8ce41a8b500"} Feb 17 13:07:29 crc kubenswrapper[4955]: I0217 13:07:29.037064 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7cjtv" event={"ID":"3b3b3288-6b3f-4a48-ae9a-f087715c7e47","Type":"ContainerStarted","Data":"47b4525f063afdd024c6414642a647009881bf4f0c4cf5d2b1743d1536ab81eb"} Feb 17 13:07:29 crc kubenswrapper[4955]: I0217 13:07:29.040922 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-2k5pz" event={"ID":"77480847-3271-4161-a833-e62b3bb4413e","Type":"ContainerStarted","Data":"1cf1cb5bbb59a16da122559224679e367e71ec4f908d46a513164972bf3b11bf"} Feb 17 13:07:29 crc kubenswrapper[4955]: I0217 13:07:29.043663 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8c8cn" event={"ID":"0bd63c82-a771-47d1-8ba4-a4036d456e0a","Type":"ContainerStarted","Data":"98ad7e49c398a2de9469eab67c57d3fd091fe96d3b384065b4e3679b4dcf6148"} Feb 17 13:07:29 crc kubenswrapper[4955]: I0217 13:07:29.074709 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pf44z" event={"ID":"3b1996f2-ac59-43dd-a18b-baeabe4f5da3","Type":"ContainerStarted","Data":"91aaf6a67d05a4134b039fc74edefcae2f346e9197261e82b421cb391020989e"} Feb 17 13:07:29 crc kubenswrapper[4955]: E0217 13:07:29.078940 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-6xlxd" podUID="6d73442c-9d91-4b5b-b0dd-efc137a55959" Feb 17 13:07:29 crc kubenswrapper[4955]: E0217 13:07:29.079027 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mxmt9" podUID="700722b8-137a-4fcb-83bb-3cd0806c676c" Feb 17 13:07:30 crc kubenswrapper[4955]: I0217 13:07:30.092925 4955 generic.go:334] "Generic (PLEG): container finished" podID="7a9eeae4-44ec-4f00-a49f-2af0b006e5cb" containerID="d9b3be727e3c8fe306a4a17f4efb2478e2b8969d98a3083a5371a88d252a7185" exitCode=0 Feb 17 13:07:30 crc kubenswrapper[4955]: I0217 13:07:30.093074 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5kwdn" event={"ID":"7a9eeae4-44ec-4f00-a49f-2af0b006e5cb","Type":"ContainerDied","Data":"d9b3be727e3c8fe306a4a17f4efb2478e2b8969d98a3083a5371a88d252a7185"} Feb 17 13:07:30 crc kubenswrapper[4955]: I0217 13:07:30.096594 4955 generic.go:334] "Generic (PLEG): container finished" podID="d1ff2d84-9dcc-43f3-821f-989fbeec8e93" containerID="9fff0794b0a3f8ac54d4b89ee18e2397711499eadc31e7e20fd1d8ce41a8b500" exitCode=0 Feb 17 13:07:30 crc kubenswrapper[4955]: I0217 13:07:30.096678 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f8t7f" event={"ID":"d1ff2d84-9dcc-43f3-821f-989fbeec8e93","Type":"ContainerDied","Data":"9fff0794b0a3f8ac54d4b89ee18e2397711499eadc31e7e20fd1d8ce41a8b500"} Feb 17 13:07:30 crc kubenswrapper[4955]: I0217 13:07:30.099083 4955 generic.go:334] "Generic (PLEG): container finished" podID="3b3b3288-6b3f-4a48-ae9a-f087715c7e47" containerID="47b4525f063afdd024c6414642a647009881bf4f0c4cf5d2b1743d1536ab81eb" exitCode=0 Feb 17 13:07:30 crc kubenswrapper[4955]: I0217 13:07:30.099156 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7cjtv" event={"ID":"3b3b3288-6b3f-4a48-ae9a-f087715c7e47","Type":"ContainerDied","Data":"47b4525f063afdd024c6414642a647009881bf4f0c4cf5d2b1743d1536ab81eb"} Feb 17 13:07:30 crc kubenswrapper[4955]: I0217 13:07:30.102335 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-2k5pz" event={"ID":"77480847-3271-4161-a833-e62b3bb4413e","Type":"ContainerStarted","Data":"4c2004e290885977500bee55a128f86c9071ec8c45239a115161a3b97bd5f223"} Feb 17 13:07:30 crc kubenswrapper[4955]: I0217 13:07:30.106967 4955 generic.go:334] "Generic (PLEG): container finished" podID="3b1996f2-ac59-43dd-a18b-baeabe4f5da3" containerID="91aaf6a67d05a4134b039fc74edefcae2f346e9197261e82b421cb391020989e" exitCode=0 Feb 17 13:07:30 crc kubenswrapper[4955]: I0217 13:07:30.107041 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pf44z" event={"ID":"3b1996f2-ac59-43dd-a18b-baeabe4f5da3","Type":"ContainerDied","Data":"91aaf6a67d05a4134b039fc74edefcae2f346e9197261e82b421cb391020989e"} Feb 17 13:07:30 crc kubenswrapper[4955]: I0217 13:07:30.117889 4955 generic.go:334] "Generic (PLEG): container finished" podID="0bd63c82-a771-47d1-8ba4-a4036d456e0a" containerID="98ad7e49c398a2de9469eab67c57d3fd091fe96d3b384065b4e3679b4dcf6148" exitCode=0 Feb 17 13:07:30 crc kubenswrapper[4955]: I0217 13:07:30.117946 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8c8cn" event={"ID":"0bd63c82-a771-47d1-8ba4-a4036d456e0a","Type":"ContainerDied","Data":"98ad7e49c398a2de9469eab67c57d3fd091fe96d3b384065b4e3679b4dcf6148"} Feb 17 13:07:31 crc kubenswrapper[4955]: I0217 13:07:31.125183 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-2k5pz" event={"ID":"77480847-3271-4161-a833-e62b3bb4413e","Type":"ContainerStarted","Data":"eeaee5e93e1621c739e42782dc1280ad24e90f6db876f3946fe1b54fb9321d59"} Feb 17 13:07:31 crc kubenswrapper[4955]: I0217 13:07:31.254528 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 17 13:07:31 crc kubenswrapper[4955]: I0217 13:07:31.271330 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-2k5pz" podStartSLOduration=168.271311187 podStartE2EDuration="2m48.271311187s" podCreationTimestamp="2026-02-17 13:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:07:31.145648005 +0000 UTC m=+189.668377548" watchObservedRunningTime="2026-02-17 13:07:31.271311187 +0000 UTC m=+189.794040730" Feb 17 13:07:32 crc kubenswrapper[4955]: I0217 13:07:32.076544 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 17 13:07:32 crc kubenswrapper[4955]: E0217 13:07:32.077332 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2cf9078-0c99-4e54-a4a0-c8d779b44c93" containerName="pruner" Feb 17 13:07:32 crc kubenswrapper[4955]: I0217 13:07:32.077366 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2cf9078-0c99-4e54-a4a0-c8d779b44c93" containerName="pruner" Feb 17 13:07:32 crc kubenswrapper[4955]: I0217 13:07:32.077497 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2cf9078-0c99-4e54-a4a0-c8d779b44c93" containerName="pruner" Feb 17 13:07:32 crc kubenswrapper[4955]: I0217 13:07:32.078035 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 17 13:07:32 crc kubenswrapper[4955]: I0217 13:07:32.080912 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 17 13:07:32 crc kubenswrapper[4955]: I0217 13:07:32.081311 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 17 13:07:32 crc kubenswrapper[4955]: I0217 13:07:32.088707 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 17 13:07:32 crc kubenswrapper[4955]: I0217 13:07:32.132178 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pf44z" event={"ID":"3b1996f2-ac59-43dd-a18b-baeabe4f5da3","Type":"ContainerStarted","Data":"161b16ccd4cf9bad13d106f561a8b041a67853d47710ef575698bd11f4ce17ea"} Feb 17 13:07:32 crc kubenswrapper[4955]: I0217 13:07:32.134722 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8c8cn" event={"ID":"0bd63c82-a771-47d1-8ba4-a4036d456e0a","Type":"ContainerStarted","Data":"bb7138407c15fec362fd25409711ec8aee6010f94ae67c0743c3555f05c6cdca"} Feb 17 13:07:32 crc kubenswrapper[4955]: I0217 13:07:32.139063 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5kwdn" event={"ID":"7a9eeae4-44ec-4f00-a49f-2af0b006e5cb","Type":"ContainerStarted","Data":"9e5aaeadddbfe4f4e422de498fb4eaec110b68f02e8eacbd1daf7fc78d54d6b4"} Feb 17 13:07:32 crc kubenswrapper[4955]: I0217 13:07:32.141536 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f8t7f" event={"ID":"d1ff2d84-9dcc-43f3-821f-989fbeec8e93","Type":"ContainerStarted","Data":"e2fef23908a1dcdb6e8491f45cee5be3d9b90e13172cb832b30d5007fe216764"} Feb 17 13:07:32 crc kubenswrapper[4955]: I0217 13:07:32.151148 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7cjtv" event={"ID":"3b3b3288-6b3f-4a48-ae9a-f087715c7e47","Type":"ContainerStarted","Data":"ef06c6e7987a2c938fe2dda1695fd2332c2134b6230674e46cd6fd0b7d28c35a"} Feb 17 13:07:32 crc kubenswrapper[4955]: I0217 13:07:32.153493 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pf44z" podStartSLOduration=2.4100508019999998 podStartE2EDuration="40.153471075s" podCreationTimestamp="2026-02-17 13:06:52 +0000 UTC" firstStartedPulling="2026-02-17 13:06:53.460291071 +0000 UTC m=+151.983020614" lastFinishedPulling="2026-02-17 13:07:31.203711344 +0000 UTC m=+189.726440887" observedRunningTime="2026-02-17 13:07:32.150240979 +0000 UTC m=+190.672970522" watchObservedRunningTime="2026-02-17 13:07:32.153471075 +0000 UTC m=+190.676200618" Feb 17 13:07:32 crc kubenswrapper[4955]: I0217 13:07:32.232853 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-f8t7f" podStartSLOduration=3.361300759 podStartE2EDuration="40.232824888s" podCreationTimestamp="2026-02-17 13:06:52 +0000 UTC" firstStartedPulling="2026-02-17 13:06:54.570986426 +0000 UTC m=+153.093715969" lastFinishedPulling="2026-02-17 13:07:31.442510555 +0000 UTC m=+189.965240098" observedRunningTime="2026-02-17 13:07:32.230313534 +0000 UTC m=+190.753043077" watchObservedRunningTime="2026-02-17 13:07:32.232824888 +0000 UTC m=+190.755554431" Feb 17 13:07:32 crc kubenswrapper[4955]: I0217 13:07:32.233484 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5kwdn" podStartSLOduration=3.527439787 podStartE2EDuration="40.233473538s" podCreationTimestamp="2026-02-17 13:06:52 +0000 UTC" firstStartedPulling="2026-02-17 13:06:54.556941508 +0000 UTC m=+153.079671051" lastFinishedPulling="2026-02-17 13:07:31.262975259 +0000 UTC m=+189.785704802" observedRunningTime="2026-02-17 13:07:32.176857202 +0000 UTC m=+190.699586755" watchObservedRunningTime="2026-02-17 13:07:32.233473538 +0000 UTC m=+190.756203081" Feb 17 13:07:32 crc kubenswrapper[4955]: I0217 13:07:32.239329 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ff4e71b6-e88e-4605-9b19-e8b35a4690d1-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"ff4e71b6-e88e-4605-9b19-e8b35a4690d1\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 17 13:07:32 crc kubenswrapper[4955]: I0217 13:07:32.239733 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ff4e71b6-e88e-4605-9b19-e8b35a4690d1-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"ff4e71b6-e88e-4605-9b19-e8b35a4690d1\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 17 13:07:32 crc kubenswrapper[4955]: I0217 13:07:32.273017 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8c8cn" podStartSLOduration=3.443691077 podStartE2EDuration="37.272989304s" podCreationTimestamp="2026-02-17 13:06:55 +0000 UTC" firstStartedPulling="2026-02-17 13:06:57.750754813 +0000 UTC m=+156.273484356" lastFinishedPulling="2026-02-17 13:07:31.58005304 +0000 UTC m=+190.102782583" observedRunningTime="2026-02-17 13:07:32.269865361 +0000 UTC m=+190.792594914" watchObservedRunningTime="2026-02-17 13:07:32.272989304 +0000 UTC m=+190.795718847" Feb 17 13:07:32 crc kubenswrapper[4955]: I0217 13:07:32.297433 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7cjtv" podStartSLOduration=3.560260039 podStartE2EDuration="37.297413782s" podCreationTimestamp="2026-02-17 13:06:55 +0000 UTC" firstStartedPulling="2026-02-17 13:06:57.764402969 +0000 UTC m=+156.287132512" lastFinishedPulling="2026-02-17 13:07:31.501556712 +0000 UTC m=+190.024286255" observedRunningTime="2026-02-17 13:07:32.293714451 +0000 UTC m=+190.816443994" watchObservedRunningTime="2026-02-17 13:07:32.297413782 +0000 UTC m=+190.820143325" Feb 17 13:07:32 crc kubenswrapper[4955]: I0217 13:07:32.342444 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ff4e71b6-e88e-4605-9b19-e8b35a4690d1-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"ff4e71b6-e88e-4605-9b19-e8b35a4690d1\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 17 13:07:32 crc kubenswrapper[4955]: I0217 13:07:32.342592 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ff4e71b6-e88e-4605-9b19-e8b35a4690d1-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"ff4e71b6-e88e-4605-9b19-e8b35a4690d1\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 17 13:07:32 crc kubenswrapper[4955]: I0217 13:07:32.342760 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ff4e71b6-e88e-4605-9b19-e8b35a4690d1-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"ff4e71b6-e88e-4605-9b19-e8b35a4690d1\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 17 13:07:32 crc kubenswrapper[4955]: I0217 13:07:32.365973 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ff4e71b6-e88e-4605-9b19-e8b35a4690d1-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"ff4e71b6-e88e-4605-9b19-e8b35a4690d1\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 17 13:07:32 crc kubenswrapper[4955]: I0217 13:07:32.396958 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 17 13:07:32 crc kubenswrapper[4955]: I0217 13:07:32.550304 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pf44z" Feb 17 13:07:32 crc kubenswrapper[4955]: I0217 13:07:32.550338 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pf44z" Feb 17 13:07:32 crc kubenswrapper[4955]: I0217 13:07:32.638655 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 17 13:07:32 crc kubenswrapper[4955]: W0217 13:07:32.648062 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podff4e71b6_e88e_4605_9b19_e8b35a4690d1.slice/crio-dce3381a0311b8026c8a63153d91a4d18054e2f123637861698339037d659ff5 WatchSource:0}: Error finding container dce3381a0311b8026c8a63153d91a4d18054e2f123637861698339037d659ff5: Status 404 returned error can't find the container with id dce3381a0311b8026c8a63153d91a4d18054e2f123637861698339037d659ff5 Feb 17 13:07:32 crc kubenswrapper[4955]: I0217 13:07:32.878515 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-f8t7f" Feb 17 13:07:32 crc kubenswrapper[4955]: I0217 13:07:32.878570 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-f8t7f" Feb 17 13:07:32 crc kubenswrapper[4955]: I0217 13:07:32.999200 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5kwdn" Feb 17 13:07:32 crc kubenswrapper[4955]: I0217 13:07:32.999263 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5kwdn" Feb 17 13:07:33 crc kubenswrapper[4955]: I0217 13:07:33.158399 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"ff4e71b6-e88e-4605-9b19-e8b35a4690d1","Type":"ContainerStarted","Data":"74179db084d2a36dd95639591a7fba9da49cfca812e1f3d6c6ef35d915f32678"} Feb 17 13:07:33 crc kubenswrapper[4955]: I0217 13:07:33.158437 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"ff4e71b6-e88e-4605-9b19-e8b35a4690d1","Type":"ContainerStarted","Data":"dce3381a0311b8026c8a63153d91a4d18054e2f123637861698339037d659ff5"} Feb 17 13:07:33 crc kubenswrapper[4955]: I0217 13:07:33.182811 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=1.182773236 podStartE2EDuration="1.182773236s" podCreationTimestamp="2026-02-17 13:07:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:07:33.180867959 +0000 UTC m=+191.703597522" watchObservedRunningTime="2026-02-17 13:07:33.182773236 +0000 UTC m=+191.705502779" Feb 17 13:07:33 crc kubenswrapper[4955]: I0217 13:07:33.767697 4955 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-pf44z" podUID="3b1996f2-ac59-43dd-a18b-baeabe4f5da3" containerName="registry-server" probeResult="failure" output=< Feb 17 13:07:33 crc kubenswrapper[4955]: timeout: failed to connect service ":50051" within 1s Feb 17 13:07:33 crc kubenswrapper[4955]: > Feb 17 13:07:33 crc kubenswrapper[4955]: I0217 13:07:33.915723 4955 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-f8t7f" podUID="d1ff2d84-9dcc-43f3-821f-989fbeec8e93" containerName="registry-server" probeResult="failure" output=< Feb 17 13:07:33 crc kubenswrapper[4955]: timeout: failed to connect service ":50051" within 1s Feb 17 13:07:33 crc kubenswrapper[4955]: > Feb 17 13:07:34 crc kubenswrapper[4955]: I0217 13:07:34.038594 4955 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-5kwdn" podUID="7a9eeae4-44ec-4f00-a49f-2af0b006e5cb" containerName="registry-server" probeResult="failure" output=< Feb 17 13:07:34 crc kubenswrapper[4955]: timeout: failed to connect service ":50051" within 1s Feb 17 13:07:34 crc kubenswrapper[4955]: > Feb 17 13:07:34 crc kubenswrapper[4955]: I0217 13:07:34.164254 4955 generic.go:334] "Generic (PLEG): container finished" podID="ff4e71b6-e88e-4605-9b19-e8b35a4690d1" containerID="74179db084d2a36dd95639591a7fba9da49cfca812e1f3d6c6ef35d915f32678" exitCode=0 Feb 17 13:07:34 crc kubenswrapper[4955]: I0217 13:07:34.164367 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"ff4e71b6-e88e-4605-9b19-e8b35a4690d1","Type":"ContainerDied","Data":"74179db084d2a36dd95639591a7fba9da49cfca812e1f3d6c6ef35d915f32678"} Feb 17 13:07:34 crc kubenswrapper[4955]: I0217 13:07:34.574412 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 13:07:34 crc kubenswrapper[4955]: I0217 13:07:34.574747 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 13:07:35 crc kubenswrapper[4955]: I0217 13:07:35.459723 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 17 13:07:35 crc kubenswrapper[4955]: I0217 13:07:35.586382 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ff4e71b6-e88e-4605-9b19-e8b35a4690d1-kubelet-dir\") pod \"ff4e71b6-e88e-4605-9b19-e8b35a4690d1\" (UID: \"ff4e71b6-e88e-4605-9b19-e8b35a4690d1\") " Feb 17 13:07:35 crc kubenswrapper[4955]: I0217 13:07:35.586443 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ff4e71b6-e88e-4605-9b19-e8b35a4690d1-kube-api-access\") pod \"ff4e71b6-e88e-4605-9b19-e8b35a4690d1\" (UID: \"ff4e71b6-e88e-4605-9b19-e8b35a4690d1\") " Feb 17 13:07:35 crc kubenswrapper[4955]: I0217 13:07:35.586535 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ff4e71b6-e88e-4605-9b19-e8b35a4690d1-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "ff4e71b6-e88e-4605-9b19-e8b35a4690d1" (UID: "ff4e71b6-e88e-4605-9b19-e8b35a4690d1"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 13:07:35 crc kubenswrapper[4955]: I0217 13:07:35.586942 4955 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ff4e71b6-e88e-4605-9b19-e8b35a4690d1-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 17 13:07:35 crc kubenswrapper[4955]: I0217 13:07:35.598569 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff4e71b6-e88e-4605-9b19-e8b35a4690d1-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "ff4e71b6-e88e-4605-9b19-e8b35a4690d1" (UID: "ff4e71b6-e88e-4605-9b19-e8b35a4690d1"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:07:35 crc kubenswrapper[4955]: I0217 13:07:35.667772 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7cjtv" Feb 17 13:07:35 crc kubenswrapper[4955]: I0217 13:07:35.667856 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7cjtv" Feb 17 13:07:35 crc kubenswrapper[4955]: I0217 13:07:35.688322 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ff4e71b6-e88e-4605-9b19-e8b35a4690d1-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 17 13:07:36 crc kubenswrapper[4955]: I0217 13:07:36.011042 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8c8cn" Feb 17 13:07:36 crc kubenswrapper[4955]: I0217 13:07:36.011957 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8c8cn" Feb 17 13:07:36 crc kubenswrapper[4955]: I0217 13:07:36.182960 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"ff4e71b6-e88e-4605-9b19-e8b35a4690d1","Type":"ContainerDied","Data":"dce3381a0311b8026c8a63153d91a4d18054e2f123637861698339037d659ff5"} Feb 17 13:07:36 crc kubenswrapper[4955]: I0217 13:07:36.183031 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 17 13:07:36 crc kubenswrapper[4955]: I0217 13:07:36.183040 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dce3381a0311b8026c8a63153d91a4d18054e2f123637861698339037d659ff5" Feb 17 13:07:36 crc kubenswrapper[4955]: I0217 13:07:36.716528 4955 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7cjtv" podUID="3b3b3288-6b3f-4a48-ae9a-f087715c7e47" containerName="registry-server" probeResult="failure" output=< Feb 17 13:07:36 crc kubenswrapper[4955]: timeout: failed to connect service ":50051" within 1s Feb 17 13:07:36 crc kubenswrapper[4955]: > Feb 17 13:07:37 crc kubenswrapper[4955]: I0217 13:07:37.047357 4955 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8c8cn" podUID="0bd63c82-a771-47d1-8ba4-a4036d456e0a" containerName="registry-server" probeResult="failure" output=< Feb 17 13:07:37 crc kubenswrapper[4955]: timeout: failed to connect service ":50051" within 1s Feb 17 13:07:37 crc kubenswrapper[4955]: > Feb 17 13:07:37 crc kubenswrapper[4955]: I0217 13:07:37.270569 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 17 13:07:37 crc kubenswrapper[4955]: E0217 13:07:37.270845 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff4e71b6-e88e-4605-9b19-e8b35a4690d1" containerName="pruner" Feb 17 13:07:37 crc kubenswrapper[4955]: I0217 13:07:37.270864 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff4e71b6-e88e-4605-9b19-e8b35a4690d1" containerName="pruner" Feb 17 13:07:37 crc kubenswrapper[4955]: I0217 13:07:37.270964 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff4e71b6-e88e-4605-9b19-e8b35a4690d1" containerName="pruner" Feb 17 13:07:37 crc kubenswrapper[4955]: I0217 13:07:37.271383 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 17 13:07:37 crc kubenswrapper[4955]: I0217 13:07:37.274732 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 17 13:07:37 crc kubenswrapper[4955]: I0217 13:07:37.282073 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 17 13:07:37 crc kubenswrapper[4955]: I0217 13:07:37.284469 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 17 13:07:37 crc kubenswrapper[4955]: I0217 13:07:37.410494 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/0a1ce3e3-775f-4b8f-8a70-41428dd22021-var-lock\") pod \"installer-9-crc\" (UID: \"0a1ce3e3-775f-4b8f-8a70-41428dd22021\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 17 13:07:37 crc kubenswrapper[4955]: I0217 13:07:37.410563 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0a1ce3e3-775f-4b8f-8a70-41428dd22021-kubelet-dir\") pod \"installer-9-crc\" (UID: \"0a1ce3e3-775f-4b8f-8a70-41428dd22021\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 17 13:07:37 crc kubenswrapper[4955]: I0217 13:07:37.410607 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0a1ce3e3-775f-4b8f-8a70-41428dd22021-kube-api-access\") pod \"installer-9-crc\" (UID: \"0a1ce3e3-775f-4b8f-8a70-41428dd22021\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 17 13:07:37 crc kubenswrapper[4955]: I0217 13:07:37.511587 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/0a1ce3e3-775f-4b8f-8a70-41428dd22021-var-lock\") pod \"installer-9-crc\" (UID: \"0a1ce3e3-775f-4b8f-8a70-41428dd22021\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 17 13:07:37 crc kubenswrapper[4955]: I0217 13:07:37.511655 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0a1ce3e3-775f-4b8f-8a70-41428dd22021-kubelet-dir\") pod \"installer-9-crc\" (UID: \"0a1ce3e3-775f-4b8f-8a70-41428dd22021\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 17 13:07:37 crc kubenswrapper[4955]: I0217 13:07:37.511676 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0a1ce3e3-775f-4b8f-8a70-41428dd22021-kube-api-access\") pod \"installer-9-crc\" (UID: \"0a1ce3e3-775f-4b8f-8a70-41428dd22021\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 17 13:07:37 crc kubenswrapper[4955]: I0217 13:07:37.511740 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/0a1ce3e3-775f-4b8f-8a70-41428dd22021-var-lock\") pod \"installer-9-crc\" (UID: \"0a1ce3e3-775f-4b8f-8a70-41428dd22021\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 17 13:07:37 crc kubenswrapper[4955]: I0217 13:07:37.511829 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0a1ce3e3-775f-4b8f-8a70-41428dd22021-kubelet-dir\") pod \"installer-9-crc\" (UID: \"0a1ce3e3-775f-4b8f-8a70-41428dd22021\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 17 13:07:37 crc kubenswrapper[4955]: I0217 13:07:37.531726 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0a1ce3e3-775f-4b8f-8a70-41428dd22021-kube-api-access\") pod \"installer-9-crc\" (UID: \"0a1ce3e3-775f-4b8f-8a70-41428dd22021\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 17 13:07:37 crc kubenswrapper[4955]: I0217 13:07:37.589621 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 17 13:07:37 crc kubenswrapper[4955]: I0217 13:07:37.992774 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 17 13:07:38 crc kubenswrapper[4955]: W0217 13:07:38.002692 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod0a1ce3e3_775f_4b8f_8a70_41428dd22021.slice/crio-a6e97cafd37f20137e96ff17ce7eed163a11f0bb5af5e80f418eb521df182f70 WatchSource:0}: Error finding container a6e97cafd37f20137e96ff17ce7eed163a11f0bb5af5e80f418eb521df182f70: Status 404 returned error can't find the container with id a6e97cafd37f20137e96ff17ce7eed163a11f0bb5af5e80f418eb521df182f70 Feb 17 13:07:38 crc kubenswrapper[4955]: I0217 13:07:38.193259 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"0a1ce3e3-775f-4b8f-8a70-41428dd22021","Type":"ContainerStarted","Data":"a6e97cafd37f20137e96ff17ce7eed163a11f0bb5af5e80f418eb521df182f70"} Feb 17 13:07:39 crc kubenswrapper[4955]: I0217 13:07:39.201750 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"0a1ce3e3-775f-4b8f-8a70-41428dd22021","Type":"ContainerStarted","Data":"80fe36ba761dc228b5cbb9f8ed6c535954127eb594cbe8d5ade11a8f78d21667"} Feb 17 13:07:40 crc kubenswrapper[4955]: I0217 13:07:40.222934 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=3.222890924 podStartE2EDuration="3.222890924s" podCreationTimestamp="2026-02-17 13:07:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:07:40.22017072 +0000 UTC m=+198.742900263" watchObservedRunningTime="2026-02-17 13:07:40.222890924 +0000 UTC m=+198.745620467" Feb 17 13:07:42 crc kubenswrapper[4955]: I0217 13:07:42.613854 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pf44z" Feb 17 13:07:42 crc kubenswrapper[4955]: I0217 13:07:42.676331 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pf44z" Feb 17 13:07:42 crc kubenswrapper[4955]: I0217 13:07:42.921731 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-f8t7f" Feb 17 13:07:42 crc kubenswrapper[4955]: I0217 13:07:42.962121 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-f8t7f" Feb 17 13:07:43 crc kubenswrapper[4955]: I0217 13:07:43.042240 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5kwdn" Feb 17 13:07:43 crc kubenswrapper[4955]: I0217 13:07:43.091213 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5kwdn" Feb 17 13:07:43 crc kubenswrapper[4955]: I0217 13:07:43.673011 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f8t7f"] Feb 17 13:07:44 crc kubenswrapper[4955]: I0217 13:07:44.228162 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-f8t7f" podUID="d1ff2d84-9dcc-43f3-821f-989fbeec8e93" containerName="registry-server" containerID="cri-o://e2fef23908a1dcdb6e8491f45cee5be3d9b90e13172cb832b30d5007fe216764" gracePeriod=2 Feb 17 13:07:45 crc kubenswrapper[4955]: I0217 13:07:45.233697 4955 generic.go:334] "Generic (PLEG): container finished" podID="d1ff2d84-9dcc-43f3-821f-989fbeec8e93" containerID="e2fef23908a1dcdb6e8491f45cee5be3d9b90e13172cb832b30d5007fe216764" exitCode=0 Feb 17 13:07:45 crc kubenswrapper[4955]: I0217 13:07:45.233747 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f8t7f" event={"ID":"d1ff2d84-9dcc-43f3-821f-989fbeec8e93","Type":"ContainerDied","Data":"e2fef23908a1dcdb6e8491f45cee5be3d9b90e13172cb832b30d5007fe216764"} Feb 17 13:07:45 crc kubenswrapper[4955]: I0217 13:07:45.460825 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5kwdn"] Feb 17 13:07:45 crc kubenswrapper[4955]: I0217 13:07:45.461328 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5kwdn" podUID="7a9eeae4-44ec-4f00-a49f-2af0b006e5cb" containerName="registry-server" containerID="cri-o://9e5aaeadddbfe4f4e422de498fb4eaec110b68f02e8eacbd1daf7fc78d54d6b4" gracePeriod=2 Feb 17 13:07:45 crc kubenswrapper[4955]: I0217 13:07:45.729248 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7cjtv" Feb 17 13:07:45 crc kubenswrapper[4955]: I0217 13:07:45.783428 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7cjtv" Feb 17 13:07:46 crc kubenswrapper[4955]: I0217 13:07:46.054615 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8c8cn" Feb 17 13:07:46 crc kubenswrapper[4955]: I0217 13:07:46.094242 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8c8cn" Feb 17 13:07:46 crc kubenswrapper[4955]: I0217 13:07:46.247905 4955 generic.go:334] "Generic (PLEG): container finished" podID="7a9eeae4-44ec-4f00-a49f-2af0b006e5cb" containerID="9e5aaeadddbfe4f4e422de498fb4eaec110b68f02e8eacbd1daf7fc78d54d6b4" exitCode=0 Feb 17 13:07:46 crc kubenswrapper[4955]: I0217 13:07:46.247992 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5kwdn" event={"ID":"7a9eeae4-44ec-4f00-a49f-2af0b006e5cb","Type":"ContainerDied","Data":"9e5aaeadddbfe4f4e422de498fb4eaec110b68f02e8eacbd1daf7fc78d54d6b4"} Feb 17 13:07:47 crc kubenswrapper[4955]: I0217 13:07:47.460078 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f8t7f" Feb 17 13:07:47 crc kubenswrapper[4955]: I0217 13:07:47.580676 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1ff2d84-9dcc-43f3-821f-989fbeec8e93-utilities\") pod \"d1ff2d84-9dcc-43f3-821f-989fbeec8e93\" (UID: \"d1ff2d84-9dcc-43f3-821f-989fbeec8e93\") " Feb 17 13:07:47 crc kubenswrapper[4955]: I0217 13:07:47.581136 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1ff2d84-9dcc-43f3-821f-989fbeec8e93-catalog-content\") pod \"d1ff2d84-9dcc-43f3-821f-989fbeec8e93\" (UID: \"d1ff2d84-9dcc-43f3-821f-989fbeec8e93\") " Feb 17 13:07:47 crc kubenswrapper[4955]: I0217 13:07:47.581196 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pbt48\" (UniqueName: \"kubernetes.io/projected/d1ff2d84-9dcc-43f3-821f-989fbeec8e93-kube-api-access-pbt48\") pod \"d1ff2d84-9dcc-43f3-821f-989fbeec8e93\" (UID: \"d1ff2d84-9dcc-43f3-821f-989fbeec8e93\") " Feb 17 13:07:47 crc kubenswrapper[4955]: I0217 13:07:47.582552 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1ff2d84-9dcc-43f3-821f-989fbeec8e93-utilities" (OuterVolumeSpecName: "utilities") pod "d1ff2d84-9dcc-43f3-821f-989fbeec8e93" (UID: "d1ff2d84-9dcc-43f3-821f-989fbeec8e93"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:07:47 crc kubenswrapper[4955]: I0217 13:07:47.590521 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1ff2d84-9dcc-43f3-821f-989fbeec8e93-kube-api-access-pbt48" (OuterVolumeSpecName: "kube-api-access-pbt48") pod "d1ff2d84-9dcc-43f3-821f-989fbeec8e93" (UID: "d1ff2d84-9dcc-43f3-821f-989fbeec8e93"). InnerVolumeSpecName "kube-api-access-pbt48". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:07:47 crc kubenswrapper[4955]: I0217 13:07:47.647437 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1ff2d84-9dcc-43f3-821f-989fbeec8e93-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d1ff2d84-9dcc-43f3-821f-989fbeec8e93" (UID: "d1ff2d84-9dcc-43f3-821f-989fbeec8e93"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:07:47 crc kubenswrapper[4955]: I0217 13:07:47.656715 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5kwdn" Feb 17 13:07:47 crc kubenswrapper[4955]: I0217 13:07:47.687505 4955 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1ff2d84-9dcc-43f3-821f-989fbeec8e93-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 13:07:47 crc kubenswrapper[4955]: I0217 13:07:47.687536 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pbt48\" (UniqueName: \"kubernetes.io/projected/d1ff2d84-9dcc-43f3-821f-989fbeec8e93-kube-api-access-pbt48\") on node \"crc\" DevicePath \"\"" Feb 17 13:07:47 crc kubenswrapper[4955]: I0217 13:07:47.687551 4955 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1ff2d84-9dcc-43f3-821f-989fbeec8e93-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 13:07:47 crc kubenswrapper[4955]: I0217 13:07:47.788082 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a9eeae4-44ec-4f00-a49f-2af0b006e5cb-catalog-content\") pod \"7a9eeae4-44ec-4f00-a49f-2af0b006e5cb\" (UID: \"7a9eeae4-44ec-4f00-a49f-2af0b006e5cb\") " Feb 17 13:07:47 crc kubenswrapper[4955]: I0217 13:07:47.788179 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4r6tg\" (UniqueName: \"kubernetes.io/projected/7a9eeae4-44ec-4f00-a49f-2af0b006e5cb-kube-api-access-4r6tg\") pod \"7a9eeae4-44ec-4f00-a49f-2af0b006e5cb\" (UID: \"7a9eeae4-44ec-4f00-a49f-2af0b006e5cb\") " Feb 17 13:07:47 crc kubenswrapper[4955]: I0217 13:07:47.788274 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a9eeae4-44ec-4f00-a49f-2af0b006e5cb-utilities\") pod \"7a9eeae4-44ec-4f00-a49f-2af0b006e5cb\" (UID: \"7a9eeae4-44ec-4f00-a49f-2af0b006e5cb\") " Feb 17 13:07:47 crc kubenswrapper[4955]: I0217 13:07:47.789269 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a9eeae4-44ec-4f00-a49f-2af0b006e5cb-utilities" (OuterVolumeSpecName: "utilities") pod "7a9eeae4-44ec-4f00-a49f-2af0b006e5cb" (UID: "7a9eeae4-44ec-4f00-a49f-2af0b006e5cb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:07:47 crc kubenswrapper[4955]: I0217 13:07:47.792845 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a9eeae4-44ec-4f00-a49f-2af0b006e5cb-kube-api-access-4r6tg" (OuterVolumeSpecName: "kube-api-access-4r6tg") pod "7a9eeae4-44ec-4f00-a49f-2af0b006e5cb" (UID: "7a9eeae4-44ec-4f00-a49f-2af0b006e5cb"). InnerVolumeSpecName "kube-api-access-4r6tg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:07:47 crc kubenswrapper[4955]: I0217 13:07:47.852008 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a9eeae4-44ec-4f00-a49f-2af0b006e5cb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7a9eeae4-44ec-4f00-a49f-2af0b006e5cb" (UID: "7a9eeae4-44ec-4f00-a49f-2af0b006e5cb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:07:47 crc kubenswrapper[4955]: I0217 13:07:47.890102 4955 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a9eeae4-44ec-4f00-a49f-2af0b006e5cb-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 13:07:47 crc kubenswrapper[4955]: I0217 13:07:47.890139 4955 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a9eeae4-44ec-4f00-a49f-2af0b006e5cb-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 13:07:47 crc kubenswrapper[4955]: I0217 13:07:47.890151 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4r6tg\" (UniqueName: \"kubernetes.io/projected/7a9eeae4-44ec-4f00-a49f-2af0b006e5cb-kube-api-access-4r6tg\") on node \"crc\" DevicePath \"\"" Feb 17 13:07:48 crc kubenswrapper[4955]: I0217 13:07:48.260396 4955 generic.go:334] "Generic (PLEG): container finished" podID="6d73442c-9d91-4b5b-b0dd-efc137a55959" containerID="eecb2ead972ac212b57fbc4fa9ac053eca89ad687e169c499ea9df1790eb70d9" exitCode=0 Feb 17 13:07:48 crc kubenswrapper[4955]: I0217 13:07:48.260476 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xlxd" event={"ID":"6d73442c-9d91-4b5b-b0dd-efc137a55959","Type":"ContainerDied","Data":"eecb2ead972ac212b57fbc4fa9ac053eca89ad687e169c499ea9df1790eb70d9"} Feb 17 13:07:48 crc kubenswrapper[4955]: I0217 13:07:48.262949 4955 generic.go:334] "Generic (PLEG): container finished" podID="6b4cdbfa-dd6b-4586-9127-23b2605d94b4" containerID="11b97789e6925c406972974292f1fd4412d98ffd7aa18d082d9221cab8dfeeb4" exitCode=0 Feb 17 13:07:48 crc kubenswrapper[4955]: I0217 13:07:48.262999 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9mrhv" event={"ID":"6b4cdbfa-dd6b-4586-9127-23b2605d94b4","Type":"ContainerDied","Data":"11b97789e6925c406972974292f1fd4412d98ffd7aa18d082d9221cab8dfeeb4"} Feb 17 13:07:48 crc kubenswrapper[4955]: I0217 13:07:48.266914 4955 generic.go:334] "Generic (PLEG): container finished" podID="700722b8-137a-4fcb-83bb-3cd0806c676c" containerID="968555c3f7f2323d0d9605a31e076b3396d018e1fee294aa23ad245cd2eac7b6" exitCode=0 Feb 17 13:07:48 crc kubenswrapper[4955]: I0217 13:07:48.267053 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mxmt9" event={"ID":"700722b8-137a-4fcb-83bb-3cd0806c676c","Type":"ContainerDied","Data":"968555c3f7f2323d0d9605a31e076b3396d018e1fee294aa23ad245cd2eac7b6"} Feb 17 13:07:48 crc kubenswrapper[4955]: I0217 13:07:48.275018 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5kwdn" event={"ID":"7a9eeae4-44ec-4f00-a49f-2af0b006e5cb","Type":"ContainerDied","Data":"ba52cde8153159687724b3fcc5c1bfca99223b16c5c451ec2386361448aa9241"} Feb 17 13:07:48 crc kubenswrapper[4955]: I0217 13:07:48.275065 4955 scope.go:117] "RemoveContainer" containerID="9e5aaeadddbfe4f4e422de498fb4eaec110b68f02e8eacbd1daf7fc78d54d6b4" Feb 17 13:07:48 crc kubenswrapper[4955]: I0217 13:07:48.275096 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5kwdn" Feb 17 13:07:48 crc kubenswrapper[4955]: I0217 13:07:48.278410 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f8t7f" event={"ID":"d1ff2d84-9dcc-43f3-821f-989fbeec8e93","Type":"ContainerDied","Data":"4ec34fa41d133037332717c6813a39a4521a4cf4290d290277655f78abed90fc"} Feb 17 13:07:48 crc kubenswrapper[4955]: I0217 13:07:48.278519 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f8t7f" Feb 17 13:07:48 crc kubenswrapper[4955]: I0217 13:07:48.309127 4955 scope.go:117] "RemoveContainer" containerID="d9b3be727e3c8fe306a4a17f4efb2478e2b8969d98a3083a5371a88d252a7185" Feb 17 13:07:48 crc kubenswrapper[4955]: I0217 13:07:48.339523 4955 scope.go:117] "RemoveContainer" containerID="3a20d1087687a07f7f544aab5b1a406c0667b4adc019b4ac5459239b9a668f04" Feb 17 13:07:48 crc kubenswrapper[4955]: I0217 13:07:48.358760 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5kwdn"] Feb 17 13:07:48 crc kubenswrapper[4955]: I0217 13:07:48.362332 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5kwdn"] Feb 17 13:07:48 crc kubenswrapper[4955]: I0217 13:07:48.364699 4955 scope.go:117] "RemoveContainer" containerID="e2fef23908a1dcdb6e8491f45cee5be3d9b90e13172cb832b30d5007fe216764" Feb 17 13:07:48 crc kubenswrapper[4955]: I0217 13:07:48.382057 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f8t7f"] Feb 17 13:07:48 crc kubenswrapper[4955]: I0217 13:07:48.384796 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-f8t7f"] Feb 17 13:07:48 crc kubenswrapper[4955]: I0217 13:07:48.391007 4955 scope.go:117] "RemoveContainer" containerID="9fff0794b0a3f8ac54d4b89ee18e2397711499eadc31e7e20fd1d8ce41a8b500" Feb 17 13:07:48 crc kubenswrapper[4955]: I0217 13:07:48.410593 4955 scope.go:117] "RemoveContainer" containerID="9d29fd01df2f747e6753f6a518a53d94be5d5d392283929731819e995ae462dc" Feb 17 13:07:48 crc kubenswrapper[4955]: I0217 13:07:48.457882 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8c8cn"] Feb 17 13:07:48 crc kubenswrapper[4955]: I0217 13:07:48.458177 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8c8cn" podUID="0bd63c82-a771-47d1-8ba4-a4036d456e0a" containerName="registry-server" containerID="cri-o://bb7138407c15fec362fd25409711ec8aee6010f94ae67c0743c3555f05c6cdca" gracePeriod=2 Feb 17 13:07:48 crc kubenswrapper[4955]: I0217 13:07:48.815752 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8c8cn" Feb 17 13:07:48 crc kubenswrapper[4955]: I0217 13:07:48.909298 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bd63c82-a771-47d1-8ba4-a4036d456e0a-catalog-content\") pod \"0bd63c82-a771-47d1-8ba4-a4036d456e0a\" (UID: \"0bd63c82-a771-47d1-8ba4-a4036d456e0a\") " Feb 17 13:07:48 crc kubenswrapper[4955]: I0217 13:07:48.909464 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bd63c82-a771-47d1-8ba4-a4036d456e0a-utilities\") pod \"0bd63c82-a771-47d1-8ba4-a4036d456e0a\" (UID: \"0bd63c82-a771-47d1-8ba4-a4036d456e0a\") " Feb 17 13:07:48 crc kubenswrapper[4955]: I0217 13:07:48.909508 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5r6zt\" (UniqueName: \"kubernetes.io/projected/0bd63c82-a771-47d1-8ba4-a4036d456e0a-kube-api-access-5r6zt\") pod \"0bd63c82-a771-47d1-8ba4-a4036d456e0a\" (UID: \"0bd63c82-a771-47d1-8ba4-a4036d456e0a\") " Feb 17 13:07:48 crc kubenswrapper[4955]: I0217 13:07:48.910656 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0bd63c82-a771-47d1-8ba4-a4036d456e0a-utilities" (OuterVolumeSpecName: "utilities") pod "0bd63c82-a771-47d1-8ba4-a4036d456e0a" (UID: "0bd63c82-a771-47d1-8ba4-a4036d456e0a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:07:48 crc kubenswrapper[4955]: I0217 13:07:48.914340 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bd63c82-a771-47d1-8ba4-a4036d456e0a-kube-api-access-5r6zt" (OuterVolumeSpecName: "kube-api-access-5r6zt") pod "0bd63c82-a771-47d1-8ba4-a4036d456e0a" (UID: "0bd63c82-a771-47d1-8ba4-a4036d456e0a"). InnerVolumeSpecName "kube-api-access-5r6zt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:07:49 crc kubenswrapper[4955]: I0217 13:07:49.010847 4955 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bd63c82-a771-47d1-8ba4-a4036d456e0a-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 13:07:49 crc kubenswrapper[4955]: I0217 13:07:49.010883 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5r6zt\" (UniqueName: \"kubernetes.io/projected/0bd63c82-a771-47d1-8ba4-a4036d456e0a-kube-api-access-5r6zt\") on node \"crc\" DevicePath \"\"" Feb 17 13:07:49 crc kubenswrapper[4955]: I0217 13:07:49.036707 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0bd63c82-a771-47d1-8ba4-a4036d456e0a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0bd63c82-a771-47d1-8ba4-a4036d456e0a" (UID: "0bd63c82-a771-47d1-8ba4-a4036d456e0a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:07:49 crc kubenswrapper[4955]: I0217 13:07:49.111629 4955 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bd63c82-a771-47d1-8ba4-a4036d456e0a-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 13:07:49 crc kubenswrapper[4955]: I0217 13:07:49.287415 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9mrhv" event={"ID":"6b4cdbfa-dd6b-4586-9127-23b2605d94b4","Type":"ContainerStarted","Data":"1ac2bad7a2d6b4d967a644de3235bb2249d27fa8e0ef0f07fea9fee3ce9aa56c"} Feb 17 13:07:49 crc kubenswrapper[4955]: I0217 13:07:49.289439 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mxmt9" event={"ID":"700722b8-137a-4fcb-83bb-3cd0806c676c","Type":"ContainerStarted","Data":"250cfe09d80307d0ceef83aaf478c513dc79912f5886cb2ab9bee8e6452d58cd"} Feb 17 13:07:49 crc kubenswrapper[4955]: I0217 13:07:49.291436 4955 generic.go:334] "Generic (PLEG): container finished" podID="0bd63c82-a771-47d1-8ba4-a4036d456e0a" containerID="bb7138407c15fec362fd25409711ec8aee6010f94ae67c0743c3555f05c6cdca" exitCode=0 Feb 17 13:07:49 crc kubenswrapper[4955]: I0217 13:07:49.291527 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8c8cn" event={"ID":"0bd63c82-a771-47d1-8ba4-a4036d456e0a","Type":"ContainerDied","Data":"bb7138407c15fec362fd25409711ec8aee6010f94ae67c0743c3555f05c6cdca"} Feb 17 13:07:49 crc kubenswrapper[4955]: I0217 13:07:49.291565 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8c8cn" event={"ID":"0bd63c82-a771-47d1-8ba4-a4036d456e0a","Type":"ContainerDied","Data":"168a75ffcf3acea83ca2de33c238e20282a74cfefdcffa122a401b6a9fcf6794"} Feb 17 13:07:49 crc kubenswrapper[4955]: I0217 13:07:49.291587 4955 scope.go:117] "RemoveContainer" containerID="bb7138407c15fec362fd25409711ec8aee6010f94ae67c0743c3555f05c6cdca" Feb 17 13:07:49 crc kubenswrapper[4955]: I0217 13:07:49.292023 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8c8cn" Feb 17 13:07:49 crc kubenswrapper[4955]: I0217 13:07:49.294507 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xlxd" event={"ID":"6d73442c-9d91-4b5b-b0dd-efc137a55959","Type":"ContainerStarted","Data":"a7e789ce5032ffe565391646ae6f22e89fb488f2420d1cb3130d5fba459d3241"} Feb 17 13:07:49 crc kubenswrapper[4955]: I0217 13:07:49.821085 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9mrhv" podStartSLOduration=2.754087229 podStartE2EDuration="55.821068318s" podCreationTimestamp="2026-02-17 13:06:54 +0000 UTC" firstStartedPulling="2026-02-17 13:06:55.639547586 +0000 UTC m=+154.162277139" lastFinishedPulling="2026-02-17 13:07:48.706528675 +0000 UTC m=+207.229258228" observedRunningTime="2026-02-17 13:07:49.819426437 +0000 UTC m=+208.342155980" watchObservedRunningTime="2026-02-17 13:07:49.821068318 +0000 UTC m=+208.343797861" Feb 17 13:07:49 crc kubenswrapper[4955]: I0217 13:07:49.842752 4955 scope.go:117] "RemoveContainer" containerID="98ad7e49c398a2de9469eab67c57d3fd091fe96d3b384065b4e3679b4dcf6148" Feb 17 13:07:49 crc kubenswrapper[4955]: I0217 13:07:49.852339 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6xlxd" podStartSLOduration=3.69758474 podStartE2EDuration="57.852315069s" podCreationTimestamp="2026-02-17 13:06:52 +0000 UTC" firstStartedPulling="2026-02-17 13:06:54.582264342 +0000 UTC m=+153.104993895" lastFinishedPulling="2026-02-17 13:07:48.736994681 +0000 UTC m=+207.259724224" observedRunningTime="2026-02-17 13:07:49.85073996 +0000 UTC m=+208.373469513" watchObservedRunningTime="2026-02-17 13:07:49.852315069 +0000 UTC m=+208.375044612" Feb 17 13:07:49 crc kubenswrapper[4955]: I0217 13:07:49.867194 4955 scope.go:117] "RemoveContainer" containerID="ef9ddd759d5fd2c654467d9c2894226dc3b4b44d11432dccc92b98b8a3450ddc" Feb 17 13:07:49 crc kubenswrapper[4955]: I0217 13:07:49.882805 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mxmt9" podStartSLOduration=2.834787474 podStartE2EDuration="55.882764335s" podCreationTimestamp="2026-02-17 13:06:54 +0000 UTC" firstStartedPulling="2026-02-17 13:06:55.63299311 +0000 UTC m=+154.155722683" lastFinishedPulling="2026-02-17 13:07:48.680970001 +0000 UTC m=+207.203699544" observedRunningTime="2026-02-17 13:07:49.871489795 +0000 UTC m=+208.394219338" watchObservedRunningTime="2026-02-17 13:07:49.882764335 +0000 UTC m=+208.405493868" Feb 17 13:07:49 crc kubenswrapper[4955]: I0217 13:07:49.885069 4955 scope.go:117] "RemoveContainer" containerID="bb7138407c15fec362fd25409711ec8aee6010f94ae67c0743c3555f05c6cdca" Feb 17 13:07:49 crc kubenswrapper[4955]: E0217 13:07:49.886957 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb7138407c15fec362fd25409711ec8aee6010f94ae67c0743c3555f05c6cdca\": container with ID starting with bb7138407c15fec362fd25409711ec8aee6010f94ae67c0743c3555f05c6cdca not found: ID does not exist" containerID="bb7138407c15fec362fd25409711ec8aee6010f94ae67c0743c3555f05c6cdca" Feb 17 13:07:49 crc kubenswrapper[4955]: I0217 13:07:49.887001 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb7138407c15fec362fd25409711ec8aee6010f94ae67c0743c3555f05c6cdca"} err="failed to get container status \"bb7138407c15fec362fd25409711ec8aee6010f94ae67c0743c3555f05c6cdca\": rpc error: code = NotFound desc = could not find container \"bb7138407c15fec362fd25409711ec8aee6010f94ae67c0743c3555f05c6cdca\": container with ID starting with bb7138407c15fec362fd25409711ec8aee6010f94ae67c0743c3555f05c6cdca not found: ID does not exist" Feb 17 13:07:49 crc kubenswrapper[4955]: I0217 13:07:49.887064 4955 scope.go:117] "RemoveContainer" containerID="98ad7e49c398a2de9469eab67c57d3fd091fe96d3b384065b4e3679b4dcf6148" Feb 17 13:07:49 crc kubenswrapper[4955]: I0217 13:07:49.887674 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8c8cn"] Feb 17 13:07:49 crc kubenswrapper[4955]: E0217 13:07:49.887997 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98ad7e49c398a2de9469eab67c57d3fd091fe96d3b384065b4e3679b4dcf6148\": container with ID starting with 98ad7e49c398a2de9469eab67c57d3fd091fe96d3b384065b4e3679b4dcf6148 not found: ID does not exist" containerID="98ad7e49c398a2de9469eab67c57d3fd091fe96d3b384065b4e3679b4dcf6148" Feb 17 13:07:49 crc kubenswrapper[4955]: I0217 13:07:49.888038 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98ad7e49c398a2de9469eab67c57d3fd091fe96d3b384065b4e3679b4dcf6148"} err="failed to get container status \"98ad7e49c398a2de9469eab67c57d3fd091fe96d3b384065b4e3679b4dcf6148\": rpc error: code = NotFound desc = could not find container \"98ad7e49c398a2de9469eab67c57d3fd091fe96d3b384065b4e3679b4dcf6148\": container with ID starting with 98ad7e49c398a2de9469eab67c57d3fd091fe96d3b384065b4e3679b4dcf6148 not found: ID does not exist" Feb 17 13:07:49 crc kubenswrapper[4955]: I0217 13:07:49.888067 4955 scope.go:117] "RemoveContainer" containerID="ef9ddd759d5fd2c654467d9c2894226dc3b4b44d11432dccc92b98b8a3450ddc" Feb 17 13:07:49 crc kubenswrapper[4955]: E0217 13:07:49.889873 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef9ddd759d5fd2c654467d9c2894226dc3b4b44d11432dccc92b98b8a3450ddc\": container with ID starting with ef9ddd759d5fd2c654467d9c2894226dc3b4b44d11432dccc92b98b8a3450ddc not found: ID does not exist" containerID="ef9ddd759d5fd2c654467d9c2894226dc3b4b44d11432dccc92b98b8a3450ddc" Feb 17 13:07:49 crc kubenswrapper[4955]: I0217 13:07:49.889904 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef9ddd759d5fd2c654467d9c2894226dc3b4b44d11432dccc92b98b8a3450ddc"} err="failed to get container status \"ef9ddd759d5fd2c654467d9c2894226dc3b4b44d11432dccc92b98b8a3450ddc\": rpc error: code = NotFound desc = could not find container \"ef9ddd759d5fd2c654467d9c2894226dc3b4b44d11432dccc92b98b8a3450ddc\": container with ID starting with ef9ddd759d5fd2c654467d9c2894226dc3b4b44d11432dccc92b98b8a3450ddc not found: ID does not exist" Feb 17 13:07:49 crc kubenswrapper[4955]: I0217 13:07:49.894518 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8c8cn"] Feb 17 13:07:50 crc kubenswrapper[4955]: I0217 13:07:50.230300 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0bd63c82-a771-47d1-8ba4-a4036d456e0a" path="/var/lib/kubelet/pods/0bd63c82-a771-47d1-8ba4-a4036d456e0a/volumes" Feb 17 13:07:50 crc kubenswrapper[4955]: I0217 13:07:50.231427 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a9eeae4-44ec-4f00-a49f-2af0b006e5cb" path="/var/lib/kubelet/pods/7a9eeae4-44ec-4f00-a49f-2af0b006e5cb/volumes" Feb 17 13:07:50 crc kubenswrapper[4955]: I0217 13:07:50.232134 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1ff2d84-9dcc-43f3-821f-989fbeec8e93" path="/var/lib/kubelet/pods/d1ff2d84-9dcc-43f3-821f-989fbeec8e93/volumes" Feb 17 13:07:52 crc kubenswrapper[4955]: I0217 13:07:52.755941 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6xlxd" Feb 17 13:07:52 crc kubenswrapper[4955]: I0217 13:07:52.756327 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6xlxd" Feb 17 13:07:52 crc kubenswrapper[4955]: I0217 13:07:52.807909 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6xlxd" Feb 17 13:07:53 crc kubenswrapper[4955]: I0217 13:07:53.400412 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6xlxd" Feb 17 13:07:54 crc kubenswrapper[4955]: I0217 13:07:54.411717 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mxmt9" Feb 17 13:07:54 crc kubenswrapper[4955]: I0217 13:07:54.411807 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mxmt9" Feb 17 13:07:54 crc kubenswrapper[4955]: I0217 13:07:54.485088 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mxmt9" Feb 17 13:07:54 crc kubenswrapper[4955]: I0217 13:07:54.773990 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9mrhv" Feb 17 13:07:54 crc kubenswrapper[4955]: I0217 13:07:54.774049 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9mrhv" Feb 17 13:07:54 crc kubenswrapper[4955]: I0217 13:07:54.832536 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9mrhv" Feb 17 13:07:55 crc kubenswrapper[4955]: I0217 13:07:55.410132 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9mrhv" Feb 17 13:07:55 crc kubenswrapper[4955]: I0217 13:07:55.410484 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mxmt9" Feb 17 13:07:56 crc kubenswrapper[4955]: I0217 13:07:56.262941 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9mrhv"] Feb 17 13:07:57 crc kubenswrapper[4955]: I0217 13:07:57.371532 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9mrhv" podUID="6b4cdbfa-dd6b-4586-9127-23b2605d94b4" containerName="registry-server" containerID="cri-o://1ac2bad7a2d6b4d967a644de3235bb2249d27fa8e0ef0f07fea9fee3ce9aa56c" gracePeriod=2 Feb 17 13:07:57 crc kubenswrapper[4955]: I0217 13:07:57.818714 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9mrhv" Feb 17 13:07:57 crc kubenswrapper[4955]: I0217 13:07:57.834153 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b4cdbfa-dd6b-4586-9127-23b2605d94b4-catalog-content\") pod \"6b4cdbfa-dd6b-4586-9127-23b2605d94b4\" (UID: \"6b4cdbfa-dd6b-4586-9127-23b2605d94b4\") " Feb 17 13:07:57 crc kubenswrapper[4955]: I0217 13:07:57.834254 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b4cdbfa-dd6b-4586-9127-23b2605d94b4-utilities\") pod \"6b4cdbfa-dd6b-4586-9127-23b2605d94b4\" (UID: \"6b4cdbfa-dd6b-4586-9127-23b2605d94b4\") " Feb 17 13:07:57 crc kubenswrapper[4955]: I0217 13:07:57.834287 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n99kp\" (UniqueName: \"kubernetes.io/projected/6b4cdbfa-dd6b-4586-9127-23b2605d94b4-kube-api-access-n99kp\") pod \"6b4cdbfa-dd6b-4586-9127-23b2605d94b4\" (UID: \"6b4cdbfa-dd6b-4586-9127-23b2605d94b4\") " Feb 17 13:07:57 crc kubenswrapper[4955]: I0217 13:07:57.835232 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b4cdbfa-dd6b-4586-9127-23b2605d94b4-utilities" (OuterVolumeSpecName: "utilities") pod "6b4cdbfa-dd6b-4586-9127-23b2605d94b4" (UID: "6b4cdbfa-dd6b-4586-9127-23b2605d94b4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:07:57 crc kubenswrapper[4955]: I0217 13:07:57.840934 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b4cdbfa-dd6b-4586-9127-23b2605d94b4-kube-api-access-n99kp" (OuterVolumeSpecName: "kube-api-access-n99kp") pod "6b4cdbfa-dd6b-4586-9127-23b2605d94b4" (UID: "6b4cdbfa-dd6b-4586-9127-23b2605d94b4"). InnerVolumeSpecName "kube-api-access-n99kp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:07:57 crc kubenswrapper[4955]: I0217 13:07:57.865829 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b4cdbfa-dd6b-4586-9127-23b2605d94b4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6b4cdbfa-dd6b-4586-9127-23b2605d94b4" (UID: "6b4cdbfa-dd6b-4586-9127-23b2605d94b4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:07:57 crc kubenswrapper[4955]: I0217 13:07:57.935300 4955 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b4cdbfa-dd6b-4586-9127-23b2605d94b4-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 13:07:57 crc kubenswrapper[4955]: I0217 13:07:57.935646 4955 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b4cdbfa-dd6b-4586-9127-23b2605d94b4-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 13:07:57 crc kubenswrapper[4955]: I0217 13:07:57.935713 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n99kp\" (UniqueName: \"kubernetes.io/projected/6b4cdbfa-dd6b-4586-9127-23b2605d94b4-kube-api-access-n99kp\") on node \"crc\" DevicePath \"\"" Feb 17 13:07:58 crc kubenswrapper[4955]: I0217 13:07:58.381915 4955 generic.go:334] "Generic (PLEG): container finished" podID="6b4cdbfa-dd6b-4586-9127-23b2605d94b4" containerID="1ac2bad7a2d6b4d967a644de3235bb2249d27fa8e0ef0f07fea9fee3ce9aa56c" exitCode=0 Feb 17 13:07:58 crc kubenswrapper[4955]: I0217 13:07:58.381978 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9mrhv" event={"ID":"6b4cdbfa-dd6b-4586-9127-23b2605d94b4","Type":"ContainerDied","Data":"1ac2bad7a2d6b4d967a644de3235bb2249d27fa8e0ef0f07fea9fee3ce9aa56c"} Feb 17 13:07:58 crc kubenswrapper[4955]: I0217 13:07:58.382017 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9mrhv" event={"ID":"6b4cdbfa-dd6b-4586-9127-23b2605d94b4","Type":"ContainerDied","Data":"4af0a8de45aed09f11485d2c269d8ac3002558203f61670af50d44c43a191b73"} Feb 17 13:07:58 crc kubenswrapper[4955]: I0217 13:07:58.382038 4955 scope.go:117] "RemoveContainer" containerID="1ac2bad7a2d6b4d967a644de3235bb2249d27fa8e0ef0f07fea9fee3ce9aa56c" Feb 17 13:07:58 crc kubenswrapper[4955]: I0217 13:07:58.384053 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9mrhv" Feb 17 13:07:58 crc kubenswrapper[4955]: I0217 13:07:58.406444 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9mrhv"] Feb 17 13:07:58 crc kubenswrapper[4955]: I0217 13:07:58.413809 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9mrhv"] Feb 17 13:07:58 crc kubenswrapper[4955]: I0217 13:07:58.417999 4955 scope.go:117] "RemoveContainer" containerID="11b97789e6925c406972974292f1fd4412d98ffd7aa18d082d9221cab8dfeeb4" Feb 17 13:07:58 crc kubenswrapper[4955]: I0217 13:07:58.451816 4955 scope.go:117] "RemoveContainer" containerID="263a9bfa1557ea4609f633f59559743f457fba14eb0d627cdf702bc80ae8ab93" Feb 17 13:07:58 crc kubenswrapper[4955]: I0217 13:07:58.477666 4955 scope.go:117] "RemoveContainer" containerID="1ac2bad7a2d6b4d967a644de3235bb2249d27fa8e0ef0f07fea9fee3ce9aa56c" Feb 17 13:07:58 crc kubenswrapper[4955]: E0217 13:07:58.478241 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ac2bad7a2d6b4d967a644de3235bb2249d27fa8e0ef0f07fea9fee3ce9aa56c\": container with ID starting with 1ac2bad7a2d6b4d967a644de3235bb2249d27fa8e0ef0f07fea9fee3ce9aa56c not found: ID does not exist" containerID="1ac2bad7a2d6b4d967a644de3235bb2249d27fa8e0ef0f07fea9fee3ce9aa56c" Feb 17 13:07:58 crc kubenswrapper[4955]: I0217 13:07:58.478308 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ac2bad7a2d6b4d967a644de3235bb2249d27fa8e0ef0f07fea9fee3ce9aa56c"} err="failed to get container status \"1ac2bad7a2d6b4d967a644de3235bb2249d27fa8e0ef0f07fea9fee3ce9aa56c\": rpc error: code = NotFound desc = could not find container \"1ac2bad7a2d6b4d967a644de3235bb2249d27fa8e0ef0f07fea9fee3ce9aa56c\": container with ID starting with 1ac2bad7a2d6b4d967a644de3235bb2249d27fa8e0ef0f07fea9fee3ce9aa56c not found: ID does not exist" Feb 17 13:07:58 crc kubenswrapper[4955]: I0217 13:07:58.478344 4955 scope.go:117] "RemoveContainer" containerID="11b97789e6925c406972974292f1fd4412d98ffd7aa18d082d9221cab8dfeeb4" Feb 17 13:07:58 crc kubenswrapper[4955]: E0217 13:07:58.478925 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11b97789e6925c406972974292f1fd4412d98ffd7aa18d082d9221cab8dfeeb4\": container with ID starting with 11b97789e6925c406972974292f1fd4412d98ffd7aa18d082d9221cab8dfeeb4 not found: ID does not exist" containerID="11b97789e6925c406972974292f1fd4412d98ffd7aa18d082d9221cab8dfeeb4" Feb 17 13:07:58 crc kubenswrapper[4955]: I0217 13:07:58.478961 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11b97789e6925c406972974292f1fd4412d98ffd7aa18d082d9221cab8dfeeb4"} err="failed to get container status \"11b97789e6925c406972974292f1fd4412d98ffd7aa18d082d9221cab8dfeeb4\": rpc error: code = NotFound desc = could not find container \"11b97789e6925c406972974292f1fd4412d98ffd7aa18d082d9221cab8dfeeb4\": container with ID starting with 11b97789e6925c406972974292f1fd4412d98ffd7aa18d082d9221cab8dfeeb4 not found: ID does not exist" Feb 17 13:07:58 crc kubenswrapper[4955]: I0217 13:07:58.478989 4955 scope.go:117] "RemoveContainer" containerID="263a9bfa1557ea4609f633f59559743f457fba14eb0d627cdf702bc80ae8ab93" Feb 17 13:07:58 crc kubenswrapper[4955]: E0217 13:07:58.479263 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"263a9bfa1557ea4609f633f59559743f457fba14eb0d627cdf702bc80ae8ab93\": container with ID starting with 263a9bfa1557ea4609f633f59559743f457fba14eb0d627cdf702bc80ae8ab93 not found: ID does not exist" containerID="263a9bfa1557ea4609f633f59559743f457fba14eb0d627cdf702bc80ae8ab93" Feb 17 13:07:58 crc kubenswrapper[4955]: I0217 13:07:58.479292 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"263a9bfa1557ea4609f633f59559743f457fba14eb0d627cdf702bc80ae8ab93"} err="failed to get container status \"263a9bfa1557ea4609f633f59559743f457fba14eb0d627cdf702bc80ae8ab93\": rpc error: code = NotFound desc = could not find container \"263a9bfa1557ea4609f633f59559743f457fba14eb0d627cdf702bc80ae8ab93\": container with ID starting with 263a9bfa1557ea4609f633f59559743f457fba14eb0d627cdf702bc80ae8ab93 not found: ID does not exist" Feb 17 13:08:00 crc kubenswrapper[4955]: I0217 13:08:00.229943 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b4cdbfa-dd6b-4586-9127-23b2605d94b4" path="/var/lib/kubelet/pods/6b4cdbfa-dd6b-4586-9127-23b2605d94b4/volumes" Feb 17 13:08:04 crc kubenswrapper[4955]: I0217 13:08:04.156180 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-vfbvk"] Feb 17 13:08:04 crc kubenswrapper[4955]: I0217 13:08:04.574521 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 13:08:04 crc kubenswrapper[4955]: I0217 13:08:04.574578 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 13:08:04 crc kubenswrapper[4955]: I0217 13:08:04.574621 4955 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" Feb 17 13:08:04 crc kubenswrapper[4955]: I0217 13:08:04.575172 4955 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9bec78f9e9f6416d6b600f6231f0d24652473b48b95bf0e0675cc85c9208d124"} pod="openshift-machine-config-operator/machine-config-daemon-29qxq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 17 13:08:04 crc kubenswrapper[4955]: I0217 13:08:04.575227 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" containerID="cri-o://9bec78f9e9f6416d6b600f6231f0d24652473b48b95bf0e0675cc85c9208d124" gracePeriod=600 Feb 17 13:08:05 crc kubenswrapper[4955]: I0217 13:08:05.419671 4955 generic.go:334] "Generic (PLEG): container finished" podID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerID="9bec78f9e9f6416d6b600f6231f0d24652473b48b95bf0e0675cc85c9208d124" exitCode=0 Feb 17 13:08:05 crc kubenswrapper[4955]: I0217 13:08:05.419740 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" event={"ID":"5bf36887-2a2b-4193-abdb-a15beaf46300","Type":"ContainerDied","Data":"9bec78f9e9f6416d6b600f6231f0d24652473b48b95bf0e0675cc85c9208d124"} Feb 17 13:08:05 crc kubenswrapper[4955]: I0217 13:08:05.420226 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" event={"ID":"5bf36887-2a2b-4193-abdb-a15beaf46300","Type":"ContainerStarted","Data":"3d5433cfcc5f7c8d75620ee096f13026d74f5b8c91209da3a81b1363052f9063"} Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.457005 4955 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 17 13:08:16 crc kubenswrapper[4955]: E0217 13:08:16.457838 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a9eeae4-44ec-4f00-a49f-2af0b006e5cb" containerName="registry-server" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.457856 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a9eeae4-44ec-4f00-a49f-2af0b006e5cb" containerName="registry-server" Feb 17 13:08:16 crc kubenswrapper[4955]: E0217 13:08:16.457865 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b4cdbfa-dd6b-4586-9127-23b2605d94b4" containerName="extract-content" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.457873 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b4cdbfa-dd6b-4586-9127-23b2605d94b4" containerName="extract-content" Feb 17 13:08:16 crc kubenswrapper[4955]: E0217 13:08:16.457881 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1ff2d84-9dcc-43f3-821f-989fbeec8e93" containerName="extract-utilities" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.457888 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1ff2d84-9dcc-43f3-821f-989fbeec8e93" containerName="extract-utilities" Feb 17 13:08:16 crc kubenswrapper[4955]: E0217 13:08:16.457902 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1ff2d84-9dcc-43f3-821f-989fbeec8e93" containerName="extract-content" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.457910 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1ff2d84-9dcc-43f3-821f-989fbeec8e93" containerName="extract-content" Feb 17 13:08:16 crc kubenswrapper[4955]: E0217 13:08:16.457928 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b4cdbfa-dd6b-4586-9127-23b2605d94b4" containerName="registry-server" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.457935 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b4cdbfa-dd6b-4586-9127-23b2605d94b4" containerName="registry-server" Feb 17 13:08:16 crc kubenswrapper[4955]: E0217 13:08:16.457950 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bd63c82-a771-47d1-8ba4-a4036d456e0a" containerName="extract-utilities" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.457957 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bd63c82-a771-47d1-8ba4-a4036d456e0a" containerName="extract-utilities" Feb 17 13:08:16 crc kubenswrapper[4955]: E0217 13:08:16.457964 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bd63c82-a771-47d1-8ba4-a4036d456e0a" containerName="registry-server" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.457971 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bd63c82-a771-47d1-8ba4-a4036d456e0a" containerName="registry-server" Feb 17 13:08:16 crc kubenswrapper[4955]: E0217 13:08:16.457982 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bd63c82-a771-47d1-8ba4-a4036d456e0a" containerName="extract-content" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.457990 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bd63c82-a771-47d1-8ba4-a4036d456e0a" containerName="extract-content" Feb 17 13:08:16 crc kubenswrapper[4955]: E0217 13:08:16.458005 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a9eeae4-44ec-4f00-a49f-2af0b006e5cb" containerName="extract-utilities" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.458012 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a9eeae4-44ec-4f00-a49f-2af0b006e5cb" containerName="extract-utilities" Feb 17 13:08:16 crc kubenswrapper[4955]: E0217 13:08:16.458022 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a9eeae4-44ec-4f00-a49f-2af0b006e5cb" containerName="extract-content" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.458028 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a9eeae4-44ec-4f00-a49f-2af0b006e5cb" containerName="extract-content" Feb 17 13:08:16 crc kubenswrapper[4955]: E0217 13:08:16.458039 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b4cdbfa-dd6b-4586-9127-23b2605d94b4" containerName="extract-utilities" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.458046 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b4cdbfa-dd6b-4586-9127-23b2605d94b4" containerName="extract-utilities" Feb 17 13:08:16 crc kubenswrapper[4955]: E0217 13:08:16.458055 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1ff2d84-9dcc-43f3-821f-989fbeec8e93" containerName="registry-server" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.458062 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1ff2d84-9dcc-43f3-821f-989fbeec8e93" containerName="registry-server" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.458173 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b4cdbfa-dd6b-4586-9127-23b2605d94b4" containerName="registry-server" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.458206 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bd63c82-a771-47d1-8ba4-a4036d456e0a" containerName="registry-server" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.458217 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a9eeae4-44ec-4f00-a49f-2af0b006e5cb" containerName="registry-server" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.458227 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1ff2d84-9dcc-43f3-821f-989fbeec8e93" containerName="registry-server" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.458635 4955 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.458946 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9" gracePeriod=15 Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.459028 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.459080 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611" gracePeriod=15 Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.459039 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3" gracePeriod=15 Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.459140 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5" gracePeriod=15 Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.459260 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac" gracePeriod=15 Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.460506 4955 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 17 13:08:16 crc kubenswrapper[4955]: E0217 13:08:16.460851 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.460880 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 17 13:08:16 crc kubenswrapper[4955]: E0217 13:08:16.460906 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.460923 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 17 13:08:16 crc kubenswrapper[4955]: E0217 13:08:16.460940 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.460952 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 17 13:08:16 crc kubenswrapper[4955]: E0217 13:08:16.460986 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.460999 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 17 13:08:16 crc kubenswrapper[4955]: E0217 13:08:16.461016 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.461027 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 17 13:08:16 crc kubenswrapper[4955]: E0217 13:08:16.461046 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.461058 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 17 13:08:16 crc kubenswrapper[4955]: E0217 13:08:16.461075 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.461087 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.463300 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.463360 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.463393 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.463413 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.463445 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.463473 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.595644 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.596013 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.596038 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.596080 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.596105 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.596146 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.596168 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.596193 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.697345 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.697410 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.697449 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.697458 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.697499 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.697526 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.697544 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.697562 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.697577 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.697606 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.697620 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.697642 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.697673 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.697708 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.697724 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 13:08:16 crc kubenswrapper[4955]: I0217 13:08:16.697743 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 13:08:17 crc kubenswrapper[4955]: I0217 13:08:17.497726 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 17 13:08:17 crc kubenswrapper[4955]: I0217 13:08:17.499967 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 17 13:08:17 crc kubenswrapper[4955]: I0217 13:08:17.500809 4955 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac" exitCode=0 Feb 17 13:08:17 crc kubenswrapper[4955]: I0217 13:08:17.500836 4955 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3" exitCode=0 Feb 17 13:08:17 crc kubenswrapper[4955]: I0217 13:08:17.500868 4955 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611" exitCode=0 Feb 17 13:08:17 crc kubenswrapper[4955]: I0217 13:08:17.500877 4955 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5" exitCode=2 Feb 17 13:08:17 crc kubenswrapper[4955]: I0217 13:08:17.500924 4955 scope.go:117] "RemoveContainer" containerID="d8f9a5c85b4c6ed840f9f5642314722f4e852e430092ab9c6ea994a5242ce472" Feb 17 13:08:17 crc kubenswrapper[4955]: I0217 13:08:17.504032 4955 generic.go:334] "Generic (PLEG): container finished" podID="0a1ce3e3-775f-4b8f-8a70-41428dd22021" containerID="80fe36ba761dc228b5cbb9f8ed6c535954127eb594cbe8d5ade11a8f78d21667" exitCode=0 Feb 17 13:08:17 crc kubenswrapper[4955]: I0217 13:08:17.504121 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"0a1ce3e3-775f-4b8f-8a70-41428dd22021","Type":"ContainerDied","Data":"80fe36ba761dc228b5cbb9f8ed6c535954127eb594cbe8d5ade11a8f78d21667"} Feb 17 13:08:17 crc kubenswrapper[4955]: I0217 13:08:17.506270 4955 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" Feb 17 13:08:17 crc kubenswrapper[4955]: I0217 13:08:17.508212 4955 status_manager.go:851] "Failed to get status for pod" podUID="0a1ce3e3-775f-4b8f-8a70-41428dd22021" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" Feb 17 13:08:18 crc kubenswrapper[4955]: I0217 13:08:18.516208 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 17 13:08:18 crc kubenswrapper[4955]: I0217 13:08:18.878175 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 17 13:08:18 crc kubenswrapper[4955]: I0217 13:08:18.879566 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 13:08:18 crc kubenswrapper[4955]: I0217 13:08:18.880161 4955 status_manager.go:851] "Failed to get status for pod" podUID="0a1ce3e3-775f-4b8f-8a70-41428dd22021" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" Feb 17 13:08:18 crc kubenswrapper[4955]: I0217 13:08:18.880470 4955 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" Feb 17 13:08:18 crc kubenswrapper[4955]: I0217 13:08:18.881367 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 17 13:08:18 crc kubenswrapper[4955]: I0217 13:08:18.881731 4955 status_manager.go:851] "Failed to get status for pod" podUID="0a1ce3e3-775f-4b8f-8a70-41428dd22021" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" Feb 17 13:08:18 crc kubenswrapper[4955]: I0217 13:08:18.882053 4955 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.028200 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0a1ce3e3-775f-4b8f-8a70-41428dd22021-kubelet-dir\") pod \"0a1ce3e3-775f-4b8f-8a70-41428dd22021\" (UID: \"0a1ce3e3-775f-4b8f-8a70-41428dd22021\") " Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.028297 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0a1ce3e3-775f-4b8f-8a70-41428dd22021-kube-api-access\") pod \"0a1ce3e3-775f-4b8f-8a70-41428dd22021\" (UID: \"0a1ce3e3-775f-4b8f-8a70-41428dd22021\") " Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.028305 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0a1ce3e3-775f-4b8f-8a70-41428dd22021-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "0a1ce3e3-775f-4b8f-8a70-41428dd22021" (UID: "0a1ce3e3-775f-4b8f-8a70-41428dd22021"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.028330 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.028408 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.028465 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.028516 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/0a1ce3e3-775f-4b8f-8a70-41428dd22021-var-lock\") pod \"0a1ce3e3-775f-4b8f-8a70-41428dd22021\" (UID: \"0a1ce3e3-775f-4b8f-8a70-41428dd22021\") " Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.028520 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.028594 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0a1ce3e3-775f-4b8f-8a70-41428dd22021-var-lock" (OuterVolumeSpecName: "var-lock") pod "0a1ce3e3-775f-4b8f-8a70-41428dd22021" (UID: "0a1ce3e3-775f-4b8f-8a70-41428dd22021"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.028606 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.028649 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.029031 4955 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0a1ce3e3-775f-4b8f-8a70-41428dd22021-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.029046 4955 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.029058 4955 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.029068 4955 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/0a1ce3e3-775f-4b8f-8a70-41428dd22021-var-lock\") on node \"crc\" DevicePath \"\"" Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.029076 4955 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.036034 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a1ce3e3-775f-4b8f-8a70-41428dd22021-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0a1ce3e3-775f-4b8f-8a70-41428dd22021" (UID: "0a1ce3e3-775f-4b8f-8a70-41428dd22021"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.129931 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0a1ce3e3-775f-4b8f-8a70-41428dd22021-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.531455 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.534891 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.534923 4955 scope.go:117] "RemoveContainer" containerID="2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac" Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.534869 4955 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9" exitCode=0 Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.545053 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"0a1ce3e3-775f-4b8f-8a70-41428dd22021","Type":"ContainerDied","Data":"a6e97cafd37f20137e96ff17ce7eed163a11f0bb5af5e80f418eb521df182f70"} Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.546106 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.545104 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6e97cafd37f20137e96ff17ce7eed163a11f0bb5af5e80f418eb521df182f70" Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.558088 4955 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.558645 4955 status_manager.go:851] "Failed to get status for pod" podUID="0a1ce3e3-775f-4b8f-8a70-41428dd22021" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.566441 4955 status_manager.go:851] "Failed to get status for pod" podUID="0a1ce3e3-775f-4b8f-8a70-41428dd22021" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.566949 4955 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.567520 4955 scope.go:117] "RemoveContainer" containerID="c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3" Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.584735 4955 scope.go:117] "RemoveContainer" containerID="e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611" Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.612568 4955 scope.go:117] "RemoveContainer" containerID="016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5" Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.626493 4955 scope.go:117] "RemoveContainer" containerID="23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9" Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.643319 4955 scope.go:117] "RemoveContainer" containerID="706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e" Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.663199 4955 scope.go:117] "RemoveContainer" containerID="2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac" Feb 17 13:08:19 crc kubenswrapper[4955]: E0217 13:08:19.664403 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac\": container with ID starting with 2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac not found: ID does not exist" containerID="2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac" Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.664458 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac"} err="failed to get container status \"2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac\": rpc error: code = NotFound desc = could not find container \"2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac\": container with ID starting with 2a39b10279078871cee535e99ed7219bb13c407cc5b660c13df822d69a75b4ac not found: ID does not exist" Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.664489 4955 scope.go:117] "RemoveContainer" containerID="c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3" Feb 17 13:08:19 crc kubenswrapper[4955]: E0217 13:08:19.664844 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\": container with ID starting with c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3 not found: ID does not exist" containerID="c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3" Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.664871 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3"} err="failed to get container status \"c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\": rpc error: code = NotFound desc = could not find container \"c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3\": container with ID starting with c202dda2eba1d758595437d0a85e0da1868964239576fcb9d0efa8875330ffa3 not found: ID does not exist" Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.664888 4955 scope.go:117] "RemoveContainer" containerID="e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611" Feb 17 13:08:19 crc kubenswrapper[4955]: E0217 13:08:19.665100 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\": container with ID starting with e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611 not found: ID does not exist" containerID="e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611" Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.665129 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611"} err="failed to get container status \"e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\": rpc error: code = NotFound desc = could not find container \"e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611\": container with ID starting with e8912f2955adc9255b1643056fce00748b5cb8332706bf3728a84573e9922611 not found: ID does not exist" Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.665145 4955 scope.go:117] "RemoveContainer" containerID="016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5" Feb 17 13:08:19 crc kubenswrapper[4955]: E0217 13:08:19.665344 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\": container with ID starting with 016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5 not found: ID does not exist" containerID="016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5" Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.665369 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5"} err="failed to get container status \"016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\": rpc error: code = NotFound desc = could not find container \"016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5\": container with ID starting with 016875a92fe91622f08a45132f55021d760c0b4bbad9953bd7dc13bf36f27dd5 not found: ID does not exist" Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.665382 4955 scope.go:117] "RemoveContainer" containerID="23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9" Feb 17 13:08:19 crc kubenswrapper[4955]: E0217 13:08:19.665594 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\": container with ID starting with 23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9 not found: ID does not exist" containerID="23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9" Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.665618 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9"} err="failed to get container status \"23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\": rpc error: code = NotFound desc = could not find container \"23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9\": container with ID starting with 23c73aa1e50a4d0a42096351ac77504316700f0411c81b595bd379b2b5b9b8a9 not found: ID does not exist" Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.666422 4955 scope.go:117] "RemoveContainer" containerID="706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e" Feb 17 13:08:19 crc kubenswrapper[4955]: E0217 13:08:19.666880 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\": container with ID starting with 706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e not found: ID does not exist" containerID="706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e" Feb 17 13:08:19 crc kubenswrapper[4955]: I0217 13:08:19.666907 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e"} err="failed to get container status \"706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\": rpc error: code = NotFound desc = could not find container \"706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e\": container with ID starting with 706a3edd68d31abd4da09cd411315d2ee307588044d70b94f3b3040b4dcd988e not found: ID does not exist" Feb 17 13:08:20 crc kubenswrapper[4955]: I0217 13:08:20.228747 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Feb 17 13:08:21 crc kubenswrapper[4955]: E0217 13:08:21.509530 4955 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.46:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 13:08:21 crc kubenswrapper[4955]: I0217 13:08:21.510318 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 13:08:21 crc kubenswrapper[4955]: W0217 13:08:21.546911 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-d12a3a6692fca787d353755e6db5f9871fbc2c4aa6d2e2d33dd052df541d0f1c WatchSource:0}: Error finding container d12a3a6692fca787d353755e6db5f9871fbc2c4aa6d2e2d33dd052df541d0f1c: Status 404 returned error can't find the container with id d12a3a6692fca787d353755e6db5f9871fbc2c4aa6d2e2d33dd052df541d0f1c Feb 17 13:08:21 crc kubenswrapper[4955]: E0217 13:08:21.550881 4955 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.46:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.18950a9ddb940917 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-17 13:08:21.550213399 +0000 UTC m=+240.072942942,LastTimestamp:2026-02-17 13:08:21.550213399 +0000 UTC m=+240.072942942,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 17 13:08:21 crc kubenswrapper[4955]: I0217 13:08:21.561072 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"d12a3a6692fca787d353755e6db5f9871fbc2c4aa6d2e2d33dd052df541d0f1c"} Feb 17 13:08:22 crc kubenswrapper[4955]: I0217 13:08:22.225431 4955 status_manager.go:851] "Failed to get status for pod" podUID="0a1ce3e3-775f-4b8f-8a70-41428dd22021" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" Feb 17 13:08:22 crc kubenswrapper[4955]: I0217 13:08:22.571137 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"480234690c7b53b1ee80941ab8487a4fd60f8e5f57154994e813eb498e132330"} Feb 17 13:08:22 crc kubenswrapper[4955]: I0217 13:08:22.571733 4955 status_manager.go:851] "Failed to get status for pod" podUID="0a1ce3e3-775f-4b8f-8a70-41428dd22021" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" Feb 17 13:08:22 crc kubenswrapper[4955]: E0217 13:08:22.571755 4955 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.46:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 13:08:23 crc kubenswrapper[4955]: E0217 13:08:23.578865 4955 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.46:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 13:08:23 crc kubenswrapper[4955]: E0217 13:08:23.866041 4955 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.46:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.18950a9ddb940917 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-17 13:08:21.550213399 +0000 UTC m=+240.072942942,LastTimestamp:2026-02-17 13:08:21.550213399 +0000 UTC m=+240.072942942,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 17 13:08:23 crc kubenswrapper[4955]: E0217 13:08:23.873371 4955 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.46:6443: connect: connection refused" Feb 17 13:08:23 crc kubenswrapper[4955]: E0217 13:08:23.873871 4955 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.46:6443: connect: connection refused" Feb 17 13:08:23 crc kubenswrapper[4955]: E0217 13:08:23.874324 4955 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.46:6443: connect: connection refused" Feb 17 13:08:23 crc kubenswrapper[4955]: E0217 13:08:23.874824 4955 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.46:6443: connect: connection refused" Feb 17 13:08:23 crc kubenswrapper[4955]: E0217 13:08:23.875342 4955 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.46:6443: connect: connection refused" Feb 17 13:08:23 crc kubenswrapper[4955]: I0217 13:08:23.875401 4955 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Feb 17 13:08:23 crc kubenswrapper[4955]: E0217 13:08:23.875853 4955 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.46:6443: connect: connection refused" interval="200ms" Feb 17 13:08:24 crc kubenswrapper[4955]: E0217 13:08:24.076977 4955 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.46:6443: connect: connection refused" interval="400ms" Feb 17 13:08:24 crc kubenswrapper[4955]: E0217 13:08:24.477515 4955 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.46:6443: connect: connection refused" interval="800ms" Feb 17 13:08:24 crc kubenswrapper[4955]: E0217 13:08:24.670470 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:08:24Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:08:24Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:08:24Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-17T13:08:24Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.46:6443: connect: connection refused" Feb 17 13:08:24 crc kubenswrapper[4955]: E0217 13:08:24.672392 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.46:6443: connect: connection refused" Feb 17 13:08:24 crc kubenswrapper[4955]: E0217 13:08:24.673165 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.46:6443: connect: connection refused" Feb 17 13:08:24 crc kubenswrapper[4955]: E0217 13:08:24.673814 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.46:6443: connect: connection refused" Feb 17 13:08:24 crc kubenswrapper[4955]: E0217 13:08:24.674134 4955 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.46:6443: connect: connection refused" Feb 17 13:08:24 crc kubenswrapper[4955]: E0217 13:08:24.674306 4955 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 17 13:08:25 crc kubenswrapper[4955]: E0217 13:08:25.279882 4955 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.46:6443: connect: connection refused" interval="1.6s" Feb 17 13:08:26 crc kubenswrapper[4955]: E0217 13:08:26.880773 4955 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.46:6443: connect: connection refused" interval="3.2s" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.189281 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" podUID="f7b03b61-99a7-4db3-be68-e2ad2149f44b" containerName="oauth-openshift" containerID="cri-o://cc506ca7959b82fc864ec0e4adc9ad768b07457682de68c5eae57837298f4c1b" gracePeriod=15 Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.222744 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.223662 4955 status_manager.go:851] "Failed to get status for pod" podUID="0a1ce3e3-775f-4b8f-8a70-41428dd22021" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.330647 4955 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d05aa39d-94e6-49d4-885e-2c1f986310f1" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.330970 4955 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d05aa39d-94e6-49d4-885e-2c1f986310f1" Feb 17 13:08:29 crc kubenswrapper[4955]: E0217 13:08:29.331509 4955 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.332130 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.555066 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.555551 4955 status_manager.go:851] "Failed to get status for pod" podUID="0a1ce3e3-775f-4b8f-8a70-41428dd22021" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.555911 4955 status_manager.go:851] "Failed to get status for pod" podUID="f7b03b61-99a7-4db3-be68-e2ad2149f44b" pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-vfbvk\": dial tcp 38.102.83.46:6443: connect: connection refused" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.610456 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f577bba65b696bc79a90ae7527501165382a38fe7bf024a5867678b4e887c3b6"} Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.610742 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"4b8fc4898e8cc5fcb7578528dac9896690a0950746bfa59193c227177b4f2836"} Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.611076 4955 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d05aa39d-94e6-49d4-885e-2c1f986310f1" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.611105 4955 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d05aa39d-94e6-49d4-885e-2c1f986310f1" Feb 17 13:08:29 crc kubenswrapper[4955]: E0217 13:08:29.611573 4955 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.611577 4955 status_manager.go:851] "Failed to get status for pod" podUID="0a1ce3e3-775f-4b8f-8a70-41428dd22021" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.611977 4955 status_manager.go:851] "Failed to get status for pod" podUID="f7b03b61-99a7-4db3-be68-e2ad2149f44b" pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-vfbvk\": dial tcp 38.102.83.46:6443: connect: connection refused" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.613851 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.613896 4955 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5" exitCode=1 Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.613959 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5"} Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.614409 4955 scope.go:117] "RemoveContainer" containerID="d353761b01014cf3a4259b141b58dd1bff8c52092a8cf76c4682914a9e0552b5" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.614411 4955 status_manager.go:851] "Failed to get status for pod" podUID="f7b03b61-99a7-4db3-be68-e2ad2149f44b" pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-vfbvk\": dial tcp 38.102.83.46:6443: connect: connection refused" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.614767 4955 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.614996 4955 status_manager.go:851] "Failed to get status for pod" podUID="0a1ce3e3-775f-4b8f-8a70-41428dd22021" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.616009 4955 generic.go:334] "Generic (PLEG): container finished" podID="f7b03b61-99a7-4db3-be68-e2ad2149f44b" containerID="cc506ca7959b82fc864ec0e4adc9ad768b07457682de68c5eae57837298f4c1b" exitCode=0 Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.616040 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.616045 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" event={"ID":"f7b03b61-99a7-4db3-be68-e2ad2149f44b","Type":"ContainerDied","Data":"cc506ca7959b82fc864ec0e4adc9ad768b07457682de68c5eae57837298f4c1b"} Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.616076 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" event={"ID":"f7b03b61-99a7-4db3-be68-e2ad2149f44b","Type":"ContainerDied","Data":"200b6ace981980dbd7db743ebc8b8bf7b7ddc033d8a0d9365e8fe8c1b91fb251"} Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.616097 4955 scope.go:117] "RemoveContainer" containerID="cc506ca7959b82fc864ec0e4adc9ad768b07457682de68c5eae57837298f4c1b" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.616464 4955 status_manager.go:851] "Failed to get status for pod" podUID="0a1ce3e3-775f-4b8f-8a70-41428dd22021" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.616721 4955 status_manager.go:851] "Failed to get status for pod" podUID="f7b03b61-99a7-4db3-be68-e2ad2149f44b" pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-vfbvk\": dial tcp 38.102.83.46:6443: connect: connection refused" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.617040 4955 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.663305 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-user-template-login\") pod \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.663398 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-system-cliconfig\") pod \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.663499 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-system-serving-cert\") pod \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.663527 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-system-ocp-branding-template\") pod \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.663593 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f7b03b61-99a7-4db3-be68-e2ad2149f44b-audit-policies\") pod \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.663631 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-system-service-ca\") pod \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.663654 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-system-session\") pod \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.663678 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-user-idp-0-file-data\") pod \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.664264 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7b03b61-99a7-4db3-be68-e2ad2149f44b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "f7b03b61-99a7-4db3-be68-e2ad2149f44b" (UID: "f7b03b61-99a7-4db3-be68-e2ad2149f44b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.664278 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "f7b03b61-99a7-4db3-be68-e2ad2149f44b" (UID: "f7b03b61-99a7-4db3-be68-e2ad2149f44b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.664422 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-user-template-provider-selection\") pod \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.664657 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-user-template-error\") pod \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.664696 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-system-router-certs\") pod \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.664715 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-system-trusted-ca-bundle\") pod \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.664733 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f7b03b61-99a7-4db3-be68-e2ad2149f44b-audit-dir\") pod \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.664717 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "f7b03b61-99a7-4db3-be68-e2ad2149f44b" (UID: "f7b03b61-99a7-4db3-be68-e2ad2149f44b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.664760 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5ftl\" (UniqueName: \"kubernetes.io/projected/f7b03b61-99a7-4db3-be68-e2ad2149f44b-kube-api-access-g5ftl\") pod \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\" (UID: \"f7b03b61-99a7-4db3-be68-e2ad2149f44b\") " Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.665016 4955 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.665030 4955 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f7b03b61-99a7-4db3-be68-e2ad2149f44b-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.665041 4955 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.665128 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7b03b61-99a7-4db3-be68-e2ad2149f44b-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f7b03b61-99a7-4db3-be68-e2ad2149f44b" (UID: "f7b03b61-99a7-4db3-be68-e2ad2149f44b"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.665368 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "f7b03b61-99a7-4db3-be68-e2ad2149f44b" (UID: "f7b03b61-99a7-4db3-be68-e2ad2149f44b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.668977 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "f7b03b61-99a7-4db3-be68-e2ad2149f44b" (UID: "f7b03b61-99a7-4db3-be68-e2ad2149f44b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.670108 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "f7b03b61-99a7-4db3-be68-e2ad2149f44b" (UID: "f7b03b61-99a7-4db3-be68-e2ad2149f44b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.670215 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "f7b03b61-99a7-4db3-be68-e2ad2149f44b" (UID: "f7b03b61-99a7-4db3-be68-e2ad2149f44b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.670243 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7b03b61-99a7-4db3-be68-e2ad2149f44b-kube-api-access-g5ftl" (OuterVolumeSpecName: "kube-api-access-g5ftl") pod "f7b03b61-99a7-4db3-be68-e2ad2149f44b" (UID: "f7b03b61-99a7-4db3-be68-e2ad2149f44b"). InnerVolumeSpecName "kube-api-access-g5ftl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.670316 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "f7b03b61-99a7-4db3-be68-e2ad2149f44b" (UID: "f7b03b61-99a7-4db3-be68-e2ad2149f44b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.670603 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "f7b03b61-99a7-4db3-be68-e2ad2149f44b" (UID: "f7b03b61-99a7-4db3-be68-e2ad2149f44b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.670727 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "f7b03b61-99a7-4db3-be68-e2ad2149f44b" (UID: "f7b03b61-99a7-4db3-be68-e2ad2149f44b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.670879 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "f7b03b61-99a7-4db3-be68-e2ad2149f44b" (UID: "f7b03b61-99a7-4db3-be68-e2ad2149f44b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.671116 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "f7b03b61-99a7-4db3-be68-e2ad2149f44b" (UID: "f7b03b61-99a7-4db3-be68-e2ad2149f44b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.686748 4955 scope.go:117] "RemoveContainer" containerID="cc506ca7959b82fc864ec0e4adc9ad768b07457682de68c5eae57837298f4c1b" Feb 17 13:08:29 crc kubenswrapper[4955]: E0217 13:08:29.691277 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc506ca7959b82fc864ec0e4adc9ad768b07457682de68c5eae57837298f4c1b\": container with ID starting with cc506ca7959b82fc864ec0e4adc9ad768b07457682de68c5eae57837298f4c1b not found: ID does not exist" containerID="cc506ca7959b82fc864ec0e4adc9ad768b07457682de68c5eae57837298f4c1b" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.691361 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc506ca7959b82fc864ec0e4adc9ad768b07457682de68c5eae57837298f4c1b"} err="failed to get container status \"cc506ca7959b82fc864ec0e4adc9ad768b07457682de68c5eae57837298f4c1b\": rpc error: code = NotFound desc = could not find container \"cc506ca7959b82fc864ec0e4adc9ad768b07457682de68c5eae57837298f4c1b\": container with ID starting with cc506ca7959b82fc864ec0e4adc9ad768b07457682de68c5eae57837298f4c1b not found: ID does not exist" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.765764 4955 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.765816 4955 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.765829 4955 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.765839 4955 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.765849 4955 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.765859 4955 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f7b03b61-99a7-4db3-be68-e2ad2149f44b-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.765868 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5ftl\" (UniqueName: \"kubernetes.io/projected/f7b03b61-99a7-4db3-be68-e2ad2149f44b-kube-api-access-g5ftl\") on node \"crc\" DevicePath \"\"" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.765877 4955 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.765886 4955 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.765895 4955 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.765903 4955 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f7b03b61-99a7-4db3-be68-e2ad2149f44b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.932759 4955 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.933240 4955 status_manager.go:851] "Failed to get status for pod" podUID="0a1ce3e3-775f-4b8f-8a70-41428dd22021" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" Feb 17 13:08:29 crc kubenswrapper[4955]: I0217 13:08:29.933606 4955 status_manager.go:851] "Failed to get status for pod" podUID="f7b03b61-99a7-4db3-be68-e2ad2149f44b" pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-vfbvk\": dial tcp 38.102.83.46:6443: connect: connection refused" Feb 17 13:08:30 crc kubenswrapper[4955]: E0217 13:08:30.081869 4955 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.46:6443: connect: connection refused" interval="6.4s" Feb 17 13:08:30 crc kubenswrapper[4955]: E0217 13:08:30.271057 4955 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.46:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" volumeName="registry-storage" Feb 17 13:08:30 crc kubenswrapper[4955]: I0217 13:08:30.625900 4955 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="f577bba65b696bc79a90ae7527501165382a38fe7bf024a5867678b4e887c3b6" exitCode=0 Feb 17 13:08:30 crc kubenswrapper[4955]: I0217 13:08:30.625999 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"f577bba65b696bc79a90ae7527501165382a38fe7bf024a5867678b4e887c3b6"} Feb 17 13:08:30 crc kubenswrapper[4955]: I0217 13:08:30.626364 4955 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d05aa39d-94e6-49d4-885e-2c1f986310f1" Feb 17 13:08:30 crc kubenswrapper[4955]: I0217 13:08:30.626390 4955 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d05aa39d-94e6-49d4-885e-2c1f986310f1" Feb 17 13:08:30 crc kubenswrapper[4955]: E0217 13:08:30.626884 4955 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 13:08:30 crc kubenswrapper[4955]: I0217 13:08:30.627322 4955 status_manager.go:851] "Failed to get status for pod" podUID="f7b03b61-99a7-4db3-be68-e2ad2149f44b" pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-vfbvk\": dial tcp 38.102.83.46:6443: connect: connection refused" Feb 17 13:08:30 crc kubenswrapper[4955]: I0217 13:08:30.627825 4955 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" Feb 17 13:08:30 crc kubenswrapper[4955]: I0217 13:08:30.629168 4955 status_manager.go:851] "Failed to get status for pod" podUID="0a1ce3e3-775f-4b8f-8a70-41428dd22021" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" Feb 17 13:08:30 crc kubenswrapper[4955]: I0217 13:08:30.630565 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 17 13:08:30 crc kubenswrapper[4955]: I0217 13:08:30.630672 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ce0253d3fc9045cfcfb872654561d32826f0ebd88da0b9d9903976942fdfd6d3"} Feb 17 13:08:30 crc kubenswrapper[4955]: I0217 13:08:30.631587 4955 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" Feb 17 13:08:30 crc kubenswrapper[4955]: I0217 13:08:30.632097 4955 status_manager.go:851] "Failed to get status for pod" podUID="0a1ce3e3-775f-4b8f-8a70-41428dd22021" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.46:6443: connect: connection refused" Feb 17 13:08:30 crc kubenswrapper[4955]: I0217 13:08:30.632606 4955 status_manager.go:851] "Failed to get status for pod" podUID="f7b03b61-99a7-4db3-be68-e2ad2149f44b" pod="openshift-authentication/oauth-openshift-558db77b4-vfbvk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-vfbvk\": dial tcp 38.102.83.46:6443: connect: connection refused" Feb 17 13:08:31 crc kubenswrapper[4955]: I0217 13:08:31.644106 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"30e3176f8440eefd266a8fb5c4aeedfc7e11a11dc3157f999926af0e3934e3f4"} Feb 17 13:08:31 crc kubenswrapper[4955]: I0217 13:08:31.644413 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"599c25b1d093e3e1f4c2c4ec431cb11f1b9ffabbe93b4ca610b1a9487b2777c5"} Feb 17 13:08:31 crc kubenswrapper[4955]: I0217 13:08:31.644440 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b17f10ad2eabaaaf3f47838b966de54f60d0081b11c8ec7c4da8f17b869f7cd8"} Feb 17 13:08:32 crc kubenswrapper[4955]: I0217 13:08:32.651355 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"9f334658ec557b2dea2db894018376c4b19afbbdaf8173de3390e3b43e0409a9"} Feb 17 13:08:32 crc kubenswrapper[4955]: I0217 13:08:32.651405 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b38bd3006fcae8de9c84c6874f1fd000f02112cb3f725e9516001867dcb65ac7"} Feb 17 13:08:32 crc kubenswrapper[4955]: I0217 13:08:32.651573 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 13:08:32 crc kubenswrapper[4955]: I0217 13:08:32.651733 4955 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d05aa39d-94e6-49d4-885e-2c1f986310f1" Feb 17 13:08:32 crc kubenswrapper[4955]: I0217 13:08:32.651761 4955 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d05aa39d-94e6-49d4-885e-2c1f986310f1" Feb 17 13:08:34 crc kubenswrapper[4955]: I0217 13:08:34.333316 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 13:08:34 crc kubenswrapper[4955]: I0217 13:08:34.333731 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 13:08:34 crc kubenswrapper[4955]: I0217 13:08:34.339177 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 13:08:36 crc kubenswrapper[4955]: I0217 13:08:36.574542 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 17 13:08:37 crc kubenswrapper[4955]: I0217 13:08:37.659075 4955 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 13:08:37 crc kubenswrapper[4955]: I0217 13:08:37.674366 4955 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d05aa39d-94e6-49d4-885e-2c1f986310f1" Feb 17 13:08:37 crc kubenswrapper[4955]: I0217 13:08:37.674403 4955 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d05aa39d-94e6-49d4-885e-2c1f986310f1" Feb 17 13:08:37 crc kubenswrapper[4955]: I0217 13:08:37.677413 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 13:08:37 crc kubenswrapper[4955]: I0217 13:08:37.682164 4955 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="9cc545ef-f02f-437f-b888-00f1e04b2bc3" Feb 17 13:08:38 crc kubenswrapper[4955]: I0217 13:08:38.579369 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 17 13:08:38 crc kubenswrapper[4955]: I0217 13:08:38.582730 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 17 13:08:38 crc kubenswrapper[4955]: I0217 13:08:38.679179 4955 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d05aa39d-94e6-49d4-885e-2c1f986310f1" Feb 17 13:08:38 crc kubenswrapper[4955]: I0217 13:08:38.679210 4955 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d05aa39d-94e6-49d4-885e-2c1f986310f1" Feb 17 13:08:42 crc kubenswrapper[4955]: I0217 13:08:42.236863 4955 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="9cc545ef-f02f-437f-b888-00f1e04b2bc3" Feb 17 13:08:46 crc kubenswrapper[4955]: I0217 13:08:46.582227 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 17 13:08:47 crc kubenswrapper[4955]: I0217 13:08:47.865811 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 17 13:08:48 crc kubenswrapper[4955]: I0217 13:08:48.031541 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 17 13:08:48 crc kubenswrapper[4955]: I0217 13:08:48.061024 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 17 13:08:48 crc kubenswrapper[4955]: I0217 13:08:48.132769 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 17 13:08:48 crc kubenswrapper[4955]: I0217 13:08:48.356308 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 17 13:08:48 crc kubenswrapper[4955]: I0217 13:08:48.924126 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 17 13:08:48 crc kubenswrapper[4955]: I0217 13:08:48.942759 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 17 13:08:49 crc kubenswrapper[4955]: I0217 13:08:49.053579 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 17 13:08:49 crc kubenswrapper[4955]: I0217 13:08:49.147097 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 17 13:08:49 crc kubenswrapper[4955]: I0217 13:08:49.162121 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 17 13:08:49 crc kubenswrapper[4955]: I0217 13:08:49.189123 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 17 13:08:49 crc kubenswrapper[4955]: I0217 13:08:49.239706 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 17 13:08:49 crc kubenswrapper[4955]: I0217 13:08:49.530625 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 17 13:08:49 crc kubenswrapper[4955]: I0217 13:08:49.644117 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 17 13:08:49 crc kubenswrapper[4955]: I0217 13:08:49.998169 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 17 13:08:50 crc kubenswrapper[4955]: I0217 13:08:50.134005 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 17 13:08:50 crc kubenswrapper[4955]: I0217 13:08:50.189192 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 17 13:08:50 crc kubenswrapper[4955]: I0217 13:08:50.234466 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 17 13:08:50 crc kubenswrapper[4955]: I0217 13:08:50.375578 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 17 13:08:50 crc kubenswrapper[4955]: I0217 13:08:50.462719 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 17 13:08:50 crc kubenswrapper[4955]: I0217 13:08:50.539411 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 17 13:08:50 crc kubenswrapper[4955]: I0217 13:08:50.664610 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 17 13:08:50 crc kubenswrapper[4955]: I0217 13:08:50.829402 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 17 13:08:50 crc kubenswrapper[4955]: I0217 13:08:50.926916 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 17 13:08:51 crc kubenswrapper[4955]: I0217 13:08:51.000953 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 17 13:08:51 crc kubenswrapper[4955]: I0217 13:08:51.019968 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 17 13:08:51 crc kubenswrapper[4955]: I0217 13:08:51.160720 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 17 13:08:51 crc kubenswrapper[4955]: I0217 13:08:51.177276 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 17 13:08:51 crc kubenswrapper[4955]: I0217 13:08:51.248291 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 17 13:08:51 crc kubenswrapper[4955]: I0217 13:08:51.384563 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 17 13:08:51 crc kubenswrapper[4955]: I0217 13:08:51.430020 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 17 13:08:51 crc kubenswrapper[4955]: I0217 13:08:51.471003 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 17 13:08:51 crc kubenswrapper[4955]: I0217 13:08:51.626424 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 17 13:08:51 crc kubenswrapper[4955]: I0217 13:08:51.689747 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 17 13:08:51 crc kubenswrapper[4955]: I0217 13:08:51.702239 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 17 13:08:51 crc kubenswrapper[4955]: I0217 13:08:51.741546 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 17 13:08:51 crc kubenswrapper[4955]: I0217 13:08:51.746843 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 17 13:08:51 crc kubenswrapper[4955]: I0217 13:08:51.788294 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 17 13:08:51 crc kubenswrapper[4955]: I0217 13:08:51.821051 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 17 13:08:51 crc kubenswrapper[4955]: I0217 13:08:51.827052 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 17 13:08:51 crc kubenswrapper[4955]: I0217 13:08:51.898316 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 17 13:08:51 crc kubenswrapper[4955]: I0217 13:08:51.907412 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 17 13:08:51 crc kubenswrapper[4955]: I0217 13:08:51.932803 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 17 13:08:51 crc kubenswrapper[4955]: I0217 13:08:51.992482 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 17 13:08:52 crc kubenswrapper[4955]: I0217 13:08:52.026016 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 17 13:08:52 crc kubenswrapper[4955]: I0217 13:08:52.051122 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 17 13:08:52 crc kubenswrapper[4955]: I0217 13:08:52.056314 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 17 13:08:52 crc kubenswrapper[4955]: I0217 13:08:52.127432 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 17 13:08:52 crc kubenswrapper[4955]: I0217 13:08:52.211367 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 17 13:08:52 crc kubenswrapper[4955]: I0217 13:08:52.240482 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 17 13:08:52 crc kubenswrapper[4955]: I0217 13:08:52.242523 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 17 13:08:52 crc kubenswrapper[4955]: I0217 13:08:52.243124 4955 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 17 13:08:52 crc kubenswrapper[4955]: I0217 13:08:52.390328 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 17 13:08:52 crc kubenswrapper[4955]: I0217 13:08:52.438849 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 17 13:08:52 crc kubenswrapper[4955]: I0217 13:08:52.498882 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 17 13:08:52 crc kubenswrapper[4955]: I0217 13:08:52.498954 4955 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 17 13:08:52 crc kubenswrapper[4955]: I0217 13:08:52.568991 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 17 13:08:52 crc kubenswrapper[4955]: I0217 13:08:52.660676 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 17 13:08:52 crc kubenswrapper[4955]: I0217 13:08:52.714657 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 17 13:08:52 crc kubenswrapper[4955]: I0217 13:08:52.788180 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 17 13:08:52 crc kubenswrapper[4955]: I0217 13:08:52.828326 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 17 13:08:52 crc kubenswrapper[4955]: I0217 13:08:52.867201 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 17 13:08:52 crc kubenswrapper[4955]: I0217 13:08:52.909332 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 17 13:08:52 crc kubenswrapper[4955]: I0217 13:08:52.970044 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 17 13:08:53 crc kubenswrapper[4955]: I0217 13:08:53.004414 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 17 13:08:53 crc kubenswrapper[4955]: I0217 13:08:53.085770 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 17 13:08:53 crc kubenswrapper[4955]: I0217 13:08:53.172693 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 17 13:08:53 crc kubenswrapper[4955]: I0217 13:08:53.509078 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 17 13:08:53 crc kubenswrapper[4955]: I0217 13:08:53.535166 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 17 13:08:53 crc kubenswrapper[4955]: I0217 13:08:53.568869 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 17 13:08:53 crc kubenswrapper[4955]: I0217 13:08:53.623585 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 17 13:08:53 crc kubenswrapper[4955]: I0217 13:08:53.641285 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 17 13:08:53 crc kubenswrapper[4955]: I0217 13:08:53.684812 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 17 13:08:53 crc kubenswrapper[4955]: I0217 13:08:53.688124 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 17 13:08:53 crc kubenswrapper[4955]: I0217 13:08:53.766126 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 17 13:08:53 crc kubenswrapper[4955]: I0217 13:08:53.781162 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 17 13:08:53 crc kubenswrapper[4955]: I0217 13:08:53.823986 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 17 13:08:53 crc kubenswrapper[4955]: I0217 13:08:53.840508 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 17 13:08:54 crc kubenswrapper[4955]: I0217 13:08:54.033534 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 17 13:08:54 crc kubenswrapper[4955]: I0217 13:08:54.067283 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 17 13:08:54 crc kubenswrapper[4955]: I0217 13:08:54.152194 4955 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 17 13:08:54 crc kubenswrapper[4955]: I0217 13:08:54.224155 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 17 13:08:54 crc kubenswrapper[4955]: I0217 13:08:54.317520 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 17 13:08:54 crc kubenswrapper[4955]: I0217 13:08:54.523031 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 17 13:08:54 crc kubenswrapper[4955]: I0217 13:08:54.713714 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 17 13:08:54 crc kubenswrapper[4955]: I0217 13:08:54.769753 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 17 13:08:54 crc kubenswrapper[4955]: I0217 13:08:54.801826 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 17 13:08:54 crc kubenswrapper[4955]: I0217 13:08:54.801982 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 17 13:08:54 crc kubenswrapper[4955]: I0217 13:08:54.930042 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 17 13:08:54 crc kubenswrapper[4955]: I0217 13:08:54.958260 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 17 13:08:54 crc kubenswrapper[4955]: I0217 13:08:54.959410 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 17 13:08:54 crc kubenswrapper[4955]: I0217 13:08:54.980661 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 17 13:08:55 crc kubenswrapper[4955]: I0217 13:08:55.027091 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 17 13:08:55 crc kubenswrapper[4955]: I0217 13:08:55.055886 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 17 13:08:55 crc kubenswrapper[4955]: I0217 13:08:55.160985 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 17 13:08:55 crc kubenswrapper[4955]: I0217 13:08:55.192985 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 17 13:08:55 crc kubenswrapper[4955]: I0217 13:08:55.285301 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 17 13:08:55 crc kubenswrapper[4955]: I0217 13:08:55.290200 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 17 13:08:55 crc kubenswrapper[4955]: I0217 13:08:55.377849 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 17 13:08:55 crc kubenswrapper[4955]: I0217 13:08:55.494079 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 17 13:08:55 crc kubenswrapper[4955]: I0217 13:08:55.557172 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 17 13:08:55 crc kubenswrapper[4955]: I0217 13:08:55.559511 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 17 13:08:55 crc kubenswrapper[4955]: I0217 13:08:55.590604 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 17 13:08:55 crc kubenswrapper[4955]: I0217 13:08:55.703343 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 17 13:08:55 crc kubenswrapper[4955]: I0217 13:08:55.714738 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 17 13:08:55 crc kubenswrapper[4955]: I0217 13:08:55.775045 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 17 13:08:55 crc kubenswrapper[4955]: I0217 13:08:55.909483 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 17 13:08:55 crc kubenswrapper[4955]: I0217 13:08:55.936147 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 17 13:08:55 crc kubenswrapper[4955]: I0217 13:08:55.943153 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 17 13:08:56 crc kubenswrapper[4955]: I0217 13:08:56.069923 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 17 13:08:56 crc kubenswrapper[4955]: I0217 13:08:56.091318 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 17 13:08:56 crc kubenswrapper[4955]: I0217 13:08:56.112346 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 17 13:08:56 crc kubenswrapper[4955]: I0217 13:08:56.166192 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 17 13:08:56 crc kubenswrapper[4955]: I0217 13:08:56.171340 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 17 13:08:56 crc kubenswrapper[4955]: I0217 13:08:56.199317 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 17 13:08:56 crc kubenswrapper[4955]: I0217 13:08:56.224246 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 17 13:08:56 crc kubenswrapper[4955]: I0217 13:08:56.283298 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 17 13:08:56 crc kubenswrapper[4955]: I0217 13:08:56.417381 4955 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 17 13:08:56 crc kubenswrapper[4955]: I0217 13:08:56.434966 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 17 13:08:56 crc kubenswrapper[4955]: I0217 13:08:56.446864 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 17 13:08:56 crc kubenswrapper[4955]: I0217 13:08:56.485350 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 17 13:08:56 crc kubenswrapper[4955]: I0217 13:08:56.529534 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 17 13:08:56 crc kubenswrapper[4955]: I0217 13:08:56.534096 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 17 13:08:56 crc kubenswrapper[4955]: I0217 13:08:56.546446 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 17 13:08:56 crc kubenswrapper[4955]: I0217 13:08:56.559874 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 17 13:08:56 crc kubenswrapper[4955]: I0217 13:08:56.607884 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 17 13:08:56 crc kubenswrapper[4955]: I0217 13:08:56.608513 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 17 13:08:56 crc kubenswrapper[4955]: I0217 13:08:56.636426 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 17 13:08:56 crc kubenswrapper[4955]: I0217 13:08:56.680112 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 17 13:08:56 crc kubenswrapper[4955]: I0217 13:08:56.695429 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 17 13:08:56 crc kubenswrapper[4955]: I0217 13:08:56.764065 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 17 13:08:56 crc kubenswrapper[4955]: I0217 13:08:56.809423 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 17 13:08:56 crc kubenswrapper[4955]: I0217 13:08:56.827190 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 17 13:08:56 crc kubenswrapper[4955]: I0217 13:08:56.854960 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 17 13:08:57 crc kubenswrapper[4955]: I0217 13:08:57.037692 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 17 13:08:57 crc kubenswrapper[4955]: I0217 13:08:57.075225 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 17 13:08:57 crc kubenswrapper[4955]: I0217 13:08:57.255016 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 17 13:08:57 crc kubenswrapper[4955]: I0217 13:08:57.269256 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 17 13:08:57 crc kubenswrapper[4955]: I0217 13:08:57.365046 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 17 13:08:57 crc kubenswrapper[4955]: I0217 13:08:57.432946 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 17 13:08:57 crc kubenswrapper[4955]: I0217 13:08:57.578340 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 17 13:08:57 crc kubenswrapper[4955]: I0217 13:08:57.590322 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 17 13:08:57 crc kubenswrapper[4955]: I0217 13:08:57.623161 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 17 13:08:57 crc kubenswrapper[4955]: I0217 13:08:57.658994 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 17 13:08:57 crc kubenswrapper[4955]: I0217 13:08:57.674171 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 17 13:08:57 crc kubenswrapper[4955]: I0217 13:08:57.699592 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 17 13:08:57 crc kubenswrapper[4955]: I0217 13:08:57.734387 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 17 13:08:57 crc kubenswrapper[4955]: I0217 13:08:57.738748 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 17 13:08:57 crc kubenswrapper[4955]: I0217 13:08:57.765328 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 17 13:08:57 crc kubenswrapper[4955]: I0217 13:08:57.844953 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 17 13:08:57 crc kubenswrapper[4955]: I0217 13:08:57.856972 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 17 13:08:57 crc kubenswrapper[4955]: I0217 13:08:57.994148 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.032616 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.083204 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.109133 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.129695 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.187440 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.280354 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.327307 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.342193 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.389498 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.398943 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.402869 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.427816 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.539045 4955 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.542967 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-vfbvk","openshift-kube-apiserver/kube-apiserver-crc"] Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.543024 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-54f7c55fd8-897r9"] Feb 17 13:08:58 crc kubenswrapper[4955]: E0217 13:08:58.543189 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7b03b61-99a7-4db3-be68-e2ad2149f44b" containerName="oauth-openshift" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.543206 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7b03b61-99a7-4db3-be68-e2ad2149f44b" containerName="oauth-openshift" Feb 17 13:08:58 crc kubenswrapper[4955]: E0217 13:08:58.543220 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a1ce3e3-775f-4b8f-8a70-41428dd22021" containerName="installer" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.543226 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a1ce3e3-775f-4b8f-8a70-41428dd22021" containerName="installer" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.543481 4955 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d05aa39d-94e6-49d4-885e-2c1f986310f1" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.543516 4955 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d05aa39d-94e6-49d4-885e-2c1f986310f1" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.543582 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a1ce3e3-775f-4b8f-8a70-41428dd22021" containerName="installer" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.543623 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7b03b61-99a7-4db3-be68-e2ad2149f44b" containerName="oauth-openshift" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.544269 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.548048 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.549439 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.549881 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.550280 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.550380 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.551437 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.551616 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.553257 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.553293 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.553615 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.554378 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.554956 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.555116 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.558063 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.560749 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.568443 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.569103 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=21.569085349 podStartE2EDuration="21.569085349s" podCreationTimestamp="2026-02-17 13:08:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:08:58.56812139 +0000 UTC m=+277.090850943" watchObservedRunningTime="2026-02-17 13:08:58.569085349 +0000 UTC m=+277.091814892" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.579411 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.604754 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.648647 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qc6ll\" (UniqueName: \"kubernetes.io/projected/5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3-kube-api-access-qc6ll\") pod \"oauth-openshift-54f7c55fd8-897r9\" (UID: \"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.648692 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3-v4-0-config-system-router-certs\") pod \"oauth-openshift-54f7c55fd8-897r9\" (UID: \"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.648712 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3-v4-0-config-system-serving-cert\") pod \"oauth-openshift-54f7c55fd8-897r9\" (UID: \"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.648740 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-54f7c55fd8-897r9\" (UID: \"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.650686 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3-v4-0-config-system-session\") pod \"oauth-openshift-54f7c55fd8-897r9\" (UID: \"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.653522 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3-audit-dir\") pod \"oauth-openshift-54f7c55fd8-897r9\" (UID: \"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.653579 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-54f7c55fd8-897r9\" (UID: \"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.653630 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-54f7c55fd8-897r9\" (UID: \"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.653654 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3-v4-0-config-system-cliconfig\") pod \"oauth-openshift-54f7c55fd8-897r9\" (UID: \"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.653702 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3-v4-0-config-system-service-ca\") pod \"oauth-openshift-54f7c55fd8-897r9\" (UID: \"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.653729 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3-v4-0-config-user-template-login\") pod \"oauth-openshift-54f7c55fd8-897r9\" (UID: \"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.653749 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3-v4-0-config-user-template-error\") pod \"oauth-openshift-54f7c55fd8-897r9\" (UID: \"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.653794 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-54f7c55fd8-897r9\" (UID: \"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.653830 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3-audit-policies\") pod \"oauth-openshift-54f7c55fd8-897r9\" (UID: \"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.755151 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qc6ll\" (UniqueName: \"kubernetes.io/projected/5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3-kube-api-access-qc6ll\") pod \"oauth-openshift-54f7c55fd8-897r9\" (UID: \"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.755207 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3-v4-0-config-system-router-certs\") pod \"oauth-openshift-54f7c55fd8-897r9\" (UID: \"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.755261 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3-v4-0-config-system-serving-cert\") pod \"oauth-openshift-54f7c55fd8-897r9\" (UID: \"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.755287 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-54f7c55fd8-897r9\" (UID: \"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.755325 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3-v4-0-config-system-session\") pod \"oauth-openshift-54f7c55fd8-897r9\" (UID: \"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.755359 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3-audit-dir\") pod \"oauth-openshift-54f7c55fd8-897r9\" (UID: \"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.755409 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-54f7c55fd8-897r9\" (UID: \"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.755452 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-54f7c55fd8-897r9\" (UID: \"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.755488 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3-v4-0-config-system-cliconfig\") pod \"oauth-openshift-54f7c55fd8-897r9\" (UID: \"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.755533 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3-v4-0-config-system-service-ca\") pod \"oauth-openshift-54f7c55fd8-897r9\" (UID: \"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.755561 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3-v4-0-config-user-template-error\") pod \"oauth-openshift-54f7c55fd8-897r9\" (UID: \"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.755599 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3-v4-0-config-user-template-login\") pod \"oauth-openshift-54f7c55fd8-897r9\" (UID: \"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.755633 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-54f7c55fd8-897r9\" (UID: \"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.755705 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3-audit-policies\") pod \"oauth-openshift-54f7c55fd8-897r9\" (UID: \"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.756017 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3-audit-dir\") pod \"oauth-openshift-54f7c55fd8-897r9\" (UID: \"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.756712 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3-audit-policies\") pod \"oauth-openshift-54f7c55fd8-897r9\" (UID: \"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.757055 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3-v4-0-config-system-cliconfig\") pod \"oauth-openshift-54f7c55fd8-897r9\" (UID: \"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.757186 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3-v4-0-config-system-service-ca\") pod \"oauth-openshift-54f7c55fd8-897r9\" (UID: \"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.758159 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-54f7c55fd8-897r9\" (UID: \"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.761717 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-54f7c55fd8-897r9\" (UID: \"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.761900 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3-v4-0-config-user-template-login\") pod \"oauth-openshift-54f7c55fd8-897r9\" (UID: \"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.761904 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-54f7c55fd8-897r9\" (UID: \"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.762215 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3-v4-0-config-user-template-error\") pod \"oauth-openshift-54f7c55fd8-897r9\" (UID: \"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.762391 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.763990 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3-v4-0-config-system-serving-cert\") pod \"oauth-openshift-54f7c55fd8-897r9\" (UID: \"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.766515 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3-v4-0-config-system-session\") pod \"oauth-openshift-54f7c55fd8-897r9\" (UID: \"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.775418 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3-v4-0-config-system-router-certs\") pod \"oauth-openshift-54f7c55fd8-897r9\" (UID: \"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.778359 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qc6ll\" (UniqueName: \"kubernetes.io/projected/5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3-kube-api-access-qc6ll\") pod \"oauth-openshift-54f7c55fd8-897r9\" (UID: \"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.780770 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-54f7c55fd8-897r9\" (UID: \"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.836579 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.871289 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.937197 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 17 13:08:58 crc kubenswrapper[4955]: I0217 13:08:58.939526 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 17 13:08:59 crc kubenswrapper[4955]: I0217 13:08:59.010397 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 17 13:08:59 crc kubenswrapper[4955]: I0217 13:08:59.050494 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 17 13:08:59 crc kubenswrapper[4955]: I0217 13:08:59.343133 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-54f7c55fd8-897r9"] Feb 17 13:08:59 crc kubenswrapper[4955]: I0217 13:08:59.362106 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 17 13:08:59 crc kubenswrapper[4955]: I0217 13:08:59.400127 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 17 13:08:59 crc kubenswrapper[4955]: I0217 13:08:59.443198 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 17 13:08:59 crc kubenswrapper[4955]: I0217 13:08:59.475776 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 17 13:08:59 crc kubenswrapper[4955]: I0217 13:08:59.476732 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 17 13:08:59 crc kubenswrapper[4955]: I0217 13:08:59.552726 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 17 13:08:59 crc kubenswrapper[4955]: I0217 13:08:59.606135 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 17 13:08:59 crc kubenswrapper[4955]: I0217 13:08:59.672727 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 17 13:08:59 crc kubenswrapper[4955]: I0217 13:08:59.724750 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 17 13:08:59 crc kubenswrapper[4955]: I0217 13:08:59.795942 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 17 13:08:59 crc kubenswrapper[4955]: I0217 13:08:59.801267 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 17 13:08:59 crc kubenswrapper[4955]: I0217 13:08:59.826391 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 17 13:08:59 crc kubenswrapper[4955]: I0217 13:08:59.841619 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" event={"ID":"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3","Type":"ContainerStarted","Data":"cc3b6ebe544e06583e36c2e65fe792e8ef8418f1b5c0be2e1686e47a7302c28c"} Feb 17 13:08:59 crc kubenswrapper[4955]: I0217 13:08:59.841844 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" event={"ID":"5c4c0e3d-7d3d-469c-8bf6-b49ac2b337d3","Type":"ContainerStarted","Data":"d4483571c5d1f44ddce697465a93977592f7e906d264ad18a91c815b8fefac91"} Feb 17 13:08:59 crc kubenswrapper[4955]: I0217 13:08:59.841954 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:08:59 crc kubenswrapper[4955]: I0217 13:08:59.864543 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" podStartSLOduration=55.86452058 podStartE2EDuration="55.86452058s" podCreationTimestamp="2026-02-17 13:08:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:08:59.863278283 +0000 UTC m=+278.386007886" watchObservedRunningTime="2026-02-17 13:08:59.86452058 +0000 UTC m=+278.387250123" Feb 17 13:08:59 crc kubenswrapper[4955]: I0217 13:08:59.985242 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 17 13:09:00 crc kubenswrapper[4955]: I0217 13:09:00.038703 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-54f7c55fd8-897r9" Feb 17 13:09:00 crc kubenswrapper[4955]: I0217 13:09:00.041615 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 17 13:09:00 crc kubenswrapper[4955]: I0217 13:09:00.084995 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 17 13:09:00 crc kubenswrapper[4955]: I0217 13:09:00.153413 4955 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 17 13:09:00 crc kubenswrapper[4955]: I0217 13:09:00.153838 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://480234690c7b53b1ee80941ab8487a4fd60f8e5f57154994e813eb498e132330" gracePeriod=5 Feb 17 13:09:00 crc kubenswrapper[4955]: I0217 13:09:00.214697 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 17 13:09:00 crc kubenswrapper[4955]: I0217 13:09:00.230370 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7b03b61-99a7-4db3-be68-e2ad2149f44b" path="/var/lib/kubelet/pods/f7b03b61-99a7-4db3-be68-e2ad2149f44b/volumes" Feb 17 13:09:00 crc kubenswrapper[4955]: I0217 13:09:00.300267 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 17 13:09:00 crc kubenswrapper[4955]: I0217 13:09:00.304513 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 17 13:09:00 crc kubenswrapper[4955]: I0217 13:09:00.305975 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 17 13:09:00 crc kubenswrapper[4955]: I0217 13:09:00.335622 4955 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 17 13:09:00 crc kubenswrapper[4955]: I0217 13:09:00.486946 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 17 13:09:00 crc kubenswrapper[4955]: I0217 13:09:00.503028 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 17 13:09:00 crc kubenswrapper[4955]: I0217 13:09:00.531163 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 17 13:09:00 crc kubenswrapper[4955]: I0217 13:09:00.590083 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 17 13:09:00 crc kubenswrapper[4955]: I0217 13:09:00.598575 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 17 13:09:00 crc kubenswrapper[4955]: I0217 13:09:00.630319 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 17 13:09:00 crc kubenswrapper[4955]: I0217 13:09:00.633997 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 17 13:09:00 crc kubenswrapper[4955]: I0217 13:09:00.730241 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 17 13:09:00 crc kubenswrapper[4955]: I0217 13:09:00.760486 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 17 13:09:00 crc kubenswrapper[4955]: I0217 13:09:00.801483 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 17 13:09:00 crc kubenswrapper[4955]: I0217 13:09:00.886740 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 17 13:09:00 crc kubenswrapper[4955]: I0217 13:09:00.919553 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 17 13:09:01 crc kubenswrapper[4955]: I0217 13:09:01.019190 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 17 13:09:01 crc kubenswrapper[4955]: I0217 13:09:01.071736 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 17 13:09:01 crc kubenswrapper[4955]: I0217 13:09:01.113656 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 17 13:09:01 crc kubenswrapper[4955]: I0217 13:09:01.337932 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 17 13:09:01 crc kubenswrapper[4955]: I0217 13:09:01.552962 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 17 13:09:01 crc kubenswrapper[4955]: I0217 13:09:01.680850 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 17 13:09:01 crc kubenswrapper[4955]: I0217 13:09:01.694558 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 17 13:09:01 crc kubenswrapper[4955]: I0217 13:09:01.759656 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 17 13:09:01 crc kubenswrapper[4955]: I0217 13:09:01.767739 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 17 13:09:01 crc kubenswrapper[4955]: I0217 13:09:01.781206 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 17 13:09:01 crc kubenswrapper[4955]: I0217 13:09:01.856240 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 17 13:09:01 crc kubenswrapper[4955]: I0217 13:09:01.977987 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 17 13:09:02 crc kubenswrapper[4955]: I0217 13:09:02.061892 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 17 13:09:02 crc kubenswrapper[4955]: I0217 13:09:02.086777 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 17 13:09:02 crc kubenswrapper[4955]: I0217 13:09:02.208163 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 17 13:09:02 crc kubenswrapper[4955]: I0217 13:09:02.507554 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 17 13:09:02 crc kubenswrapper[4955]: I0217 13:09:02.591905 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 17 13:09:02 crc kubenswrapper[4955]: I0217 13:09:02.683623 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 17 13:09:02 crc kubenswrapper[4955]: I0217 13:09:02.765855 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 17 13:09:02 crc kubenswrapper[4955]: I0217 13:09:02.845404 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 17 13:09:02 crc kubenswrapper[4955]: I0217 13:09:02.865540 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 17 13:09:03 crc kubenswrapper[4955]: I0217 13:09:03.127688 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 17 13:09:03 crc kubenswrapper[4955]: I0217 13:09:03.198471 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 17 13:09:03 crc kubenswrapper[4955]: I0217 13:09:03.364248 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 17 13:09:03 crc kubenswrapper[4955]: I0217 13:09:03.409894 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 17 13:09:03 crc kubenswrapper[4955]: I0217 13:09:03.887638 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 17 13:09:03 crc kubenswrapper[4955]: I0217 13:09:03.964822 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 17 13:09:03 crc kubenswrapper[4955]: I0217 13:09:03.975012 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 17 13:09:04 crc kubenswrapper[4955]: I0217 13:09:04.676923 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 17 13:09:04 crc kubenswrapper[4955]: I0217 13:09:04.847967 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 17 13:09:05 crc kubenswrapper[4955]: I0217 13:09:05.394426 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 17 13:09:05 crc kubenswrapper[4955]: I0217 13:09:05.723444 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 17 13:09:05 crc kubenswrapper[4955]: I0217 13:09:05.723508 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 13:09:05 crc kubenswrapper[4955]: I0217 13:09:05.797072 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 17 13:09:05 crc kubenswrapper[4955]: I0217 13:09:05.797199 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 13:09:05 crc kubenswrapper[4955]: I0217 13:09:05.797485 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 17 13:09:05 crc kubenswrapper[4955]: I0217 13:09:05.797607 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 17 13:09:05 crc kubenswrapper[4955]: I0217 13:09:05.797666 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 17 13:09:05 crc kubenswrapper[4955]: I0217 13:09:05.797696 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 13:09:05 crc kubenswrapper[4955]: I0217 13:09:05.797721 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 17 13:09:05 crc kubenswrapper[4955]: I0217 13:09:05.797970 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 13:09:05 crc kubenswrapper[4955]: I0217 13:09:05.798334 4955 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Feb 17 13:09:05 crc kubenswrapper[4955]: I0217 13:09:05.798371 4955 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Feb 17 13:09:05 crc kubenswrapper[4955]: I0217 13:09:05.798396 4955 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 17 13:09:05 crc kubenswrapper[4955]: I0217 13:09:05.798622 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 13:09:05 crc kubenswrapper[4955]: I0217 13:09:05.807542 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 13:09:05 crc kubenswrapper[4955]: I0217 13:09:05.893939 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 17 13:09:05 crc kubenswrapper[4955]: I0217 13:09:05.894001 4955 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="480234690c7b53b1ee80941ab8487a4fd60f8e5f57154994e813eb498e132330" exitCode=137 Feb 17 13:09:05 crc kubenswrapper[4955]: I0217 13:09:05.894051 4955 scope.go:117] "RemoveContainer" containerID="480234690c7b53b1ee80941ab8487a4fd60f8e5f57154994e813eb498e132330" Feb 17 13:09:05 crc kubenswrapper[4955]: I0217 13:09:05.894105 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 17 13:09:05 crc kubenswrapper[4955]: I0217 13:09:05.899699 4955 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Feb 17 13:09:05 crc kubenswrapper[4955]: I0217 13:09:05.899760 4955 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 17 13:09:05 crc kubenswrapper[4955]: I0217 13:09:05.927984 4955 scope.go:117] "RemoveContainer" containerID="480234690c7b53b1ee80941ab8487a4fd60f8e5f57154994e813eb498e132330" Feb 17 13:09:05 crc kubenswrapper[4955]: E0217 13:09:05.928444 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"480234690c7b53b1ee80941ab8487a4fd60f8e5f57154994e813eb498e132330\": container with ID starting with 480234690c7b53b1ee80941ab8487a4fd60f8e5f57154994e813eb498e132330 not found: ID does not exist" containerID="480234690c7b53b1ee80941ab8487a4fd60f8e5f57154994e813eb498e132330" Feb 17 13:09:05 crc kubenswrapper[4955]: I0217 13:09:05.928508 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"480234690c7b53b1ee80941ab8487a4fd60f8e5f57154994e813eb498e132330"} err="failed to get container status \"480234690c7b53b1ee80941ab8487a4fd60f8e5f57154994e813eb498e132330\": rpc error: code = NotFound desc = could not find container \"480234690c7b53b1ee80941ab8487a4fd60f8e5f57154994e813eb498e132330\": container with ID starting with 480234690c7b53b1ee80941ab8487a4fd60f8e5f57154994e813eb498e132330 not found: ID does not exist" Feb 17 13:09:06 crc kubenswrapper[4955]: I0217 13:09:06.234958 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Feb 17 13:09:06 crc kubenswrapper[4955]: I0217 13:09:06.839365 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 17 13:09:21 crc kubenswrapper[4955]: I0217 13:09:21.980708 4955 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Feb 17 13:09:23 crc kubenswrapper[4955]: I0217 13:09:23.009269 4955 generic.go:334] "Generic (PLEG): container finished" podID="84d8965a-d181-4f61-921f-1802347f6606" containerID="11d49f6fc9e716b23a59cefaaaff05e16ef8d468c92c03e402a3224d7476fe31" exitCode=0 Feb 17 13:09:23 crc kubenswrapper[4955]: I0217 13:09:23.009357 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9c9zp" event={"ID":"84d8965a-d181-4f61-921f-1802347f6606","Type":"ContainerDied","Data":"11d49f6fc9e716b23a59cefaaaff05e16ef8d468c92c03e402a3224d7476fe31"} Feb 17 13:09:23 crc kubenswrapper[4955]: I0217 13:09:23.009912 4955 scope.go:117] "RemoveContainer" containerID="11d49f6fc9e716b23a59cefaaaff05e16ef8d468c92c03e402a3224d7476fe31" Feb 17 13:09:24 crc kubenswrapper[4955]: I0217 13:09:24.020042 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9c9zp" event={"ID":"84d8965a-d181-4f61-921f-1802347f6606","Type":"ContainerStarted","Data":"1973286b55d00e9ed0c19af324b11bd225aa27aa5e3eebd49c9295c83656042d"} Feb 17 13:09:24 crc kubenswrapper[4955]: I0217 13:09:24.020651 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-9c9zp" Feb 17 13:09:24 crc kubenswrapper[4955]: I0217 13:09:24.023048 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-9c9zp" Feb 17 13:09:28 crc kubenswrapper[4955]: I0217 13:09:28.217159 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-prtk8"] Feb 17 13:09:28 crc kubenswrapper[4955]: I0217 13:09:28.217388 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-prtk8" podUID="7db0eeb4-4ffe-4375-a854-0f47efce6a26" containerName="controller-manager" containerID="cri-o://fe421b07e4e9c29655c274a8cefc6a5ff5d1505159119a1ac18382323bc08690" gracePeriod=30 Feb 17 13:09:28 crc kubenswrapper[4955]: I0217 13:09:28.314076 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-trrhn"] Feb 17 13:09:28 crc kubenswrapper[4955]: I0217 13:09:28.314305 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-trrhn" podUID="32ea5b76-e8e5-4906-bc55-99d5ec780ff2" containerName="route-controller-manager" containerID="cri-o://cc3917f4c922095ef859cc70871d72ee960aae7344cb604f284832b5d076b35a" gracePeriod=30 Feb 17 13:09:28 crc kubenswrapper[4955]: I0217 13:09:28.835907 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-trrhn" Feb 17 13:09:28 crc kubenswrapper[4955]: I0217 13:09:28.945008 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/32ea5b76-e8e5-4906-bc55-99d5ec780ff2-client-ca\") pod \"32ea5b76-e8e5-4906-bc55-99d5ec780ff2\" (UID: \"32ea5b76-e8e5-4906-bc55-99d5ec780ff2\") " Feb 17 13:09:28 crc kubenswrapper[4955]: I0217 13:09:28.945200 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32ea5b76-e8e5-4906-bc55-99d5ec780ff2-config\") pod \"32ea5b76-e8e5-4906-bc55-99d5ec780ff2\" (UID: \"32ea5b76-e8e5-4906-bc55-99d5ec780ff2\") " Feb 17 13:09:28 crc kubenswrapper[4955]: I0217 13:09:28.945224 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32ea5b76-e8e5-4906-bc55-99d5ec780ff2-serving-cert\") pod \"32ea5b76-e8e5-4906-bc55-99d5ec780ff2\" (UID: \"32ea5b76-e8e5-4906-bc55-99d5ec780ff2\") " Feb 17 13:09:28 crc kubenswrapper[4955]: I0217 13:09:28.945253 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qn6qd\" (UniqueName: \"kubernetes.io/projected/32ea5b76-e8e5-4906-bc55-99d5ec780ff2-kube-api-access-qn6qd\") pod \"32ea5b76-e8e5-4906-bc55-99d5ec780ff2\" (UID: \"32ea5b76-e8e5-4906-bc55-99d5ec780ff2\") " Feb 17 13:09:28 crc kubenswrapper[4955]: I0217 13:09:28.946193 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32ea5b76-e8e5-4906-bc55-99d5ec780ff2-client-ca" (OuterVolumeSpecName: "client-ca") pod "32ea5b76-e8e5-4906-bc55-99d5ec780ff2" (UID: "32ea5b76-e8e5-4906-bc55-99d5ec780ff2"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:09:28 crc kubenswrapper[4955]: I0217 13:09:28.946235 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32ea5b76-e8e5-4906-bc55-99d5ec780ff2-config" (OuterVolumeSpecName: "config") pod "32ea5b76-e8e5-4906-bc55-99d5ec780ff2" (UID: "32ea5b76-e8e5-4906-bc55-99d5ec780ff2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:09:28 crc kubenswrapper[4955]: I0217 13:09:28.951713 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32ea5b76-e8e5-4906-bc55-99d5ec780ff2-kube-api-access-qn6qd" (OuterVolumeSpecName: "kube-api-access-qn6qd") pod "32ea5b76-e8e5-4906-bc55-99d5ec780ff2" (UID: "32ea5b76-e8e5-4906-bc55-99d5ec780ff2"). InnerVolumeSpecName "kube-api-access-qn6qd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:09:28 crc kubenswrapper[4955]: I0217 13:09:28.951835 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32ea5b76-e8e5-4906-bc55-99d5ec780ff2-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "32ea5b76-e8e5-4906-bc55-99d5ec780ff2" (UID: "32ea5b76-e8e5-4906-bc55-99d5ec780ff2"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.046809 4955 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/32ea5b76-e8e5-4906-bc55-99d5ec780ff2-client-ca\") on node \"crc\" DevicePath \"\"" Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.046885 4955 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32ea5b76-e8e5-4906-bc55-99d5ec780ff2-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.046898 4955 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32ea5b76-e8e5-4906-bc55-99d5ec780ff2-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.046910 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qn6qd\" (UniqueName: \"kubernetes.io/projected/32ea5b76-e8e5-4906-bc55-99d5ec780ff2-kube-api-access-qn6qd\") on node \"crc\" DevicePath \"\"" Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.047938 4955 generic.go:334] "Generic (PLEG): container finished" podID="7db0eeb4-4ffe-4375-a854-0f47efce6a26" containerID="fe421b07e4e9c29655c274a8cefc6a5ff5d1505159119a1ac18382323bc08690" exitCode=0 Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.048008 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-prtk8" event={"ID":"7db0eeb4-4ffe-4375-a854-0f47efce6a26","Type":"ContainerDied","Data":"fe421b07e4e9c29655c274a8cefc6a5ff5d1505159119a1ac18382323bc08690"} Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.049421 4955 generic.go:334] "Generic (PLEG): container finished" podID="32ea5b76-e8e5-4906-bc55-99d5ec780ff2" containerID="cc3917f4c922095ef859cc70871d72ee960aae7344cb604f284832b5d076b35a" exitCode=0 Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.049445 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-trrhn" event={"ID":"32ea5b76-e8e5-4906-bc55-99d5ec780ff2","Type":"ContainerDied","Data":"cc3917f4c922095ef859cc70871d72ee960aae7344cb604f284832b5d076b35a"} Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.049459 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-trrhn" event={"ID":"32ea5b76-e8e5-4906-bc55-99d5ec780ff2","Type":"ContainerDied","Data":"08906d556aa78f139c86352c20de5ae166084c72fcef01bc0329fd92a9e4948a"} Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.049478 4955 scope.go:117] "RemoveContainer" containerID="cc3917f4c922095ef859cc70871d72ee960aae7344cb604f284832b5d076b35a" Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.049550 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-trrhn" Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.071902 4955 scope.go:117] "RemoveContainer" containerID="cc3917f4c922095ef859cc70871d72ee960aae7344cb604f284832b5d076b35a" Feb 17 13:09:29 crc kubenswrapper[4955]: E0217 13:09:29.072289 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc3917f4c922095ef859cc70871d72ee960aae7344cb604f284832b5d076b35a\": container with ID starting with cc3917f4c922095ef859cc70871d72ee960aae7344cb604f284832b5d076b35a not found: ID does not exist" containerID="cc3917f4c922095ef859cc70871d72ee960aae7344cb604f284832b5d076b35a" Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.072328 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc3917f4c922095ef859cc70871d72ee960aae7344cb604f284832b5d076b35a"} err="failed to get container status \"cc3917f4c922095ef859cc70871d72ee960aae7344cb604f284832b5d076b35a\": rpc error: code = NotFound desc = could not find container \"cc3917f4c922095ef859cc70871d72ee960aae7344cb604f284832b5d076b35a\": container with ID starting with cc3917f4c922095ef859cc70871d72ee960aae7344cb604f284832b5d076b35a not found: ID does not exist" Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.080010 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-trrhn"] Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.084701 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-trrhn"] Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.088711 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-prtk8" Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.248104 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7db0eeb4-4ffe-4375-a854-0f47efce6a26-serving-cert\") pod \"7db0eeb4-4ffe-4375-a854-0f47efce6a26\" (UID: \"7db0eeb4-4ffe-4375-a854-0f47efce6a26\") " Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.248156 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7db0eeb4-4ffe-4375-a854-0f47efce6a26-proxy-ca-bundles\") pod \"7db0eeb4-4ffe-4375-a854-0f47efce6a26\" (UID: \"7db0eeb4-4ffe-4375-a854-0f47efce6a26\") " Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.248206 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7db0eeb4-4ffe-4375-a854-0f47efce6a26-client-ca\") pod \"7db0eeb4-4ffe-4375-a854-0f47efce6a26\" (UID: \"7db0eeb4-4ffe-4375-a854-0f47efce6a26\") " Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.248499 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmtdf\" (UniqueName: \"kubernetes.io/projected/7db0eeb4-4ffe-4375-a854-0f47efce6a26-kube-api-access-cmtdf\") pod \"7db0eeb4-4ffe-4375-a854-0f47efce6a26\" (UID: \"7db0eeb4-4ffe-4375-a854-0f47efce6a26\") " Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.248565 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7db0eeb4-4ffe-4375-a854-0f47efce6a26-config\") pod \"7db0eeb4-4ffe-4375-a854-0f47efce6a26\" (UID: \"7db0eeb4-4ffe-4375-a854-0f47efce6a26\") " Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.249763 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7db0eeb4-4ffe-4375-a854-0f47efce6a26-client-ca" (OuterVolumeSpecName: "client-ca") pod "7db0eeb4-4ffe-4375-a854-0f47efce6a26" (UID: "7db0eeb4-4ffe-4375-a854-0f47efce6a26"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.249842 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7db0eeb4-4ffe-4375-a854-0f47efce6a26-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7db0eeb4-4ffe-4375-a854-0f47efce6a26" (UID: "7db0eeb4-4ffe-4375-a854-0f47efce6a26"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.249881 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7db0eeb4-4ffe-4375-a854-0f47efce6a26-config" (OuterVolumeSpecName: "config") pod "7db0eeb4-4ffe-4375-a854-0f47efce6a26" (UID: "7db0eeb4-4ffe-4375-a854-0f47efce6a26"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.252650 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7db0eeb4-4ffe-4375-a854-0f47efce6a26-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7db0eeb4-4ffe-4375-a854-0f47efce6a26" (UID: "7db0eeb4-4ffe-4375-a854-0f47efce6a26"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.253861 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7db0eeb4-4ffe-4375-a854-0f47efce6a26-kube-api-access-cmtdf" (OuterVolumeSpecName: "kube-api-access-cmtdf") pod "7db0eeb4-4ffe-4375-a854-0f47efce6a26" (UID: "7db0eeb4-4ffe-4375-a854-0f47efce6a26"). InnerVolumeSpecName "kube-api-access-cmtdf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.350580 4955 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7db0eeb4-4ffe-4375-a854-0f47efce6a26-client-ca\") on node \"crc\" DevicePath \"\"" Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.350645 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmtdf\" (UniqueName: \"kubernetes.io/projected/7db0eeb4-4ffe-4375-a854-0f47efce6a26-kube-api-access-cmtdf\") on node \"crc\" DevicePath \"\"" Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.350674 4955 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7db0eeb4-4ffe-4375-a854-0f47efce6a26-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.350698 4955 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7db0eeb4-4ffe-4375-a854-0f47efce6a26-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.350724 4955 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7db0eeb4-4ffe-4375-a854-0f47efce6a26-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.945611 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7db5bf594-sbbgc"] Feb 17 13:09:29 crc kubenswrapper[4955]: E0217 13:09:29.946189 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32ea5b76-e8e5-4906-bc55-99d5ec780ff2" containerName="route-controller-manager" Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.946205 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="32ea5b76-e8e5-4906-bc55-99d5ec780ff2" containerName="route-controller-manager" Feb 17 13:09:29 crc kubenswrapper[4955]: E0217 13:09:29.946223 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.946233 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 17 13:09:29 crc kubenswrapper[4955]: E0217 13:09:29.946251 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7db0eeb4-4ffe-4375-a854-0f47efce6a26" containerName="controller-manager" Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.946260 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="7db0eeb4-4ffe-4375-a854-0f47efce6a26" containerName="controller-manager" Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.946379 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="7db0eeb4-4ffe-4375-a854-0f47efce6a26" containerName="controller-manager" Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.946400 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.946408 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="32ea5b76-e8e5-4906-bc55-99d5ec780ff2" containerName="route-controller-manager" Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.946849 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7db5bf594-sbbgc" Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.949162 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7584f948bb-dlvtn"] Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.949708 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7584f948bb-dlvtn" Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.951811 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.951961 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.954620 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.954993 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.956070 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.957884 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.968328 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7db5bf594-sbbgc"] Feb 17 13:09:29 crc kubenswrapper[4955]: I0217 13:09:29.972338 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7584f948bb-dlvtn"] Feb 17 13:09:30 crc kubenswrapper[4955]: I0217 13:09:30.058274 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-prtk8" event={"ID":"7db0eeb4-4ffe-4375-a854-0f47efce6a26","Type":"ContainerDied","Data":"b56a5c634f4e287c54fee13d4c6d793ec5b8e65999090f699d2f6746bc6a79bb"} Feb 17 13:09:30 crc kubenswrapper[4955]: I0217 13:09:30.058320 4955 scope.go:117] "RemoveContainer" containerID="fe421b07e4e9c29655c274a8cefc6a5ff5d1505159119a1ac18382323bc08690" Feb 17 13:09:30 crc kubenswrapper[4955]: I0217 13:09:30.058376 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-prtk8" Feb 17 13:09:30 crc kubenswrapper[4955]: I0217 13:09:30.058920 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3cf775f4-cb5f-4fb9-96e3-3fd76976f522-client-ca\") pod \"controller-manager-7db5bf594-sbbgc\" (UID: \"3cf775f4-cb5f-4fb9-96e3-3fd76976f522\") " pod="openshift-controller-manager/controller-manager-7db5bf594-sbbgc" Feb 17 13:09:30 crc kubenswrapper[4955]: I0217 13:09:30.058950 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3cf775f4-cb5f-4fb9-96e3-3fd76976f522-serving-cert\") pod \"controller-manager-7db5bf594-sbbgc\" (UID: \"3cf775f4-cb5f-4fb9-96e3-3fd76976f522\") " pod="openshift-controller-manager/controller-manager-7db5bf594-sbbgc" Feb 17 13:09:30 crc kubenswrapper[4955]: I0217 13:09:30.058978 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zl46p\" (UniqueName: \"kubernetes.io/projected/8c22eb11-d490-4e5d-9276-a732918e1b37-kube-api-access-zl46p\") pod \"route-controller-manager-7584f948bb-dlvtn\" (UID: \"8c22eb11-d490-4e5d-9276-a732918e1b37\") " pod="openshift-route-controller-manager/route-controller-manager-7584f948bb-dlvtn" Feb 17 13:09:30 crc kubenswrapper[4955]: I0217 13:09:30.058999 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c22eb11-d490-4e5d-9276-a732918e1b37-config\") pod \"route-controller-manager-7584f948bb-dlvtn\" (UID: \"8c22eb11-d490-4e5d-9276-a732918e1b37\") " pod="openshift-route-controller-manager/route-controller-manager-7584f948bb-dlvtn" Feb 17 13:09:30 crc kubenswrapper[4955]: I0217 13:09:30.059048 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wt85w\" (UniqueName: \"kubernetes.io/projected/3cf775f4-cb5f-4fb9-96e3-3fd76976f522-kube-api-access-wt85w\") pod \"controller-manager-7db5bf594-sbbgc\" (UID: \"3cf775f4-cb5f-4fb9-96e3-3fd76976f522\") " pod="openshift-controller-manager/controller-manager-7db5bf594-sbbgc" Feb 17 13:09:30 crc kubenswrapper[4955]: I0217 13:09:30.059068 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3cf775f4-cb5f-4fb9-96e3-3fd76976f522-config\") pod \"controller-manager-7db5bf594-sbbgc\" (UID: \"3cf775f4-cb5f-4fb9-96e3-3fd76976f522\") " pod="openshift-controller-manager/controller-manager-7db5bf594-sbbgc" Feb 17 13:09:30 crc kubenswrapper[4955]: I0217 13:09:30.059090 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8c22eb11-d490-4e5d-9276-a732918e1b37-client-ca\") pod \"route-controller-manager-7584f948bb-dlvtn\" (UID: \"8c22eb11-d490-4e5d-9276-a732918e1b37\") " pod="openshift-route-controller-manager/route-controller-manager-7584f948bb-dlvtn" Feb 17 13:09:30 crc kubenswrapper[4955]: I0217 13:09:30.059121 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8c22eb11-d490-4e5d-9276-a732918e1b37-serving-cert\") pod \"route-controller-manager-7584f948bb-dlvtn\" (UID: \"8c22eb11-d490-4e5d-9276-a732918e1b37\") " pod="openshift-route-controller-manager/route-controller-manager-7584f948bb-dlvtn" Feb 17 13:09:30 crc kubenswrapper[4955]: I0217 13:09:30.059144 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3cf775f4-cb5f-4fb9-96e3-3fd76976f522-proxy-ca-bundles\") pod \"controller-manager-7db5bf594-sbbgc\" (UID: \"3cf775f4-cb5f-4fb9-96e3-3fd76976f522\") " pod="openshift-controller-manager/controller-manager-7db5bf594-sbbgc" Feb 17 13:09:30 crc kubenswrapper[4955]: I0217 13:09:30.084824 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-prtk8"] Feb 17 13:09:30 crc kubenswrapper[4955]: I0217 13:09:30.091194 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-prtk8"] Feb 17 13:09:30 crc kubenswrapper[4955]: I0217 13:09:30.159919 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zl46p\" (UniqueName: \"kubernetes.io/projected/8c22eb11-d490-4e5d-9276-a732918e1b37-kube-api-access-zl46p\") pod \"route-controller-manager-7584f948bb-dlvtn\" (UID: \"8c22eb11-d490-4e5d-9276-a732918e1b37\") " pod="openshift-route-controller-manager/route-controller-manager-7584f948bb-dlvtn" Feb 17 13:09:30 crc kubenswrapper[4955]: I0217 13:09:30.159970 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c22eb11-d490-4e5d-9276-a732918e1b37-config\") pod \"route-controller-manager-7584f948bb-dlvtn\" (UID: \"8c22eb11-d490-4e5d-9276-a732918e1b37\") " pod="openshift-route-controller-manager/route-controller-manager-7584f948bb-dlvtn" Feb 17 13:09:30 crc kubenswrapper[4955]: I0217 13:09:30.160049 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wt85w\" (UniqueName: \"kubernetes.io/projected/3cf775f4-cb5f-4fb9-96e3-3fd76976f522-kube-api-access-wt85w\") pod \"controller-manager-7db5bf594-sbbgc\" (UID: \"3cf775f4-cb5f-4fb9-96e3-3fd76976f522\") " pod="openshift-controller-manager/controller-manager-7db5bf594-sbbgc" Feb 17 13:09:30 crc kubenswrapper[4955]: I0217 13:09:30.160077 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3cf775f4-cb5f-4fb9-96e3-3fd76976f522-config\") pod \"controller-manager-7db5bf594-sbbgc\" (UID: \"3cf775f4-cb5f-4fb9-96e3-3fd76976f522\") " pod="openshift-controller-manager/controller-manager-7db5bf594-sbbgc" Feb 17 13:09:30 crc kubenswrapper[4955]: I0217 13:09:30.160105 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8c22eb11-d490-4e5d-9276-a732918e1b37-client-ca\") pod \"route-controller-manager-7584f948bb-dlvtn\" (UID: \"8c22eb11-d490-4e5d-9276-a732918e1b37\") " pod="openshift-route-controller-manager/route-controller-manager-7584f948bb-dlvtn" Feb 17 13:09:30 crc kubenswrapper[4955]: I0217 13:09:30.160134 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8c22eb11-d490-4e5d-9276-a732918e1b37-serving-cert\") pod \"route-controller-manager-7584f948bb-dlvtn\" (UID: \"8c22eb11-d490-4e5d-9276-a732918e1b37\") " pod="openshift-route-controller-manager/route-controller-manager-7584f948bb-dlvtn" Feb 17 13:09:30 crc kubenswrapper[4955]: I0217 13:09:30.160158 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3cf775f4-cb5f-4fb9-96e3-3fd76976f522-proxy-ca-bundles\") pod \"controller-manager-7db5bf594-sbbgc\" (UID: \"3cf775f4-cb5f-4fb9-96e3-3fd76976f522\") " pod="openshift-controller-manager/controller-manager-7db5bf594-sbbgc" Feb 17 13:09:30 crc kubenswrapper[4955]: I0217 13:09:30.160200 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3cf775f4-cb5f-4fb9-96e3-3fd76976f522-client-ca\") pod \"controller-manager-7db5bf594-sbbgc\" (UID: \"3cf775f4-cb5f-4fb9-96e3-3fd76976f522\") " pod="openshift-controller-manager/controller-manager-7db5bf594-sbbgc" Feb 17 13:09:30 crc kubenswrapper[4955]: I0217 13:09:30.160226 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3cf775f4-cb5f-4fb9-96e3-3fd76976f522-serving-cert\") pod \"controller-manager-7db5bf594-sbbgc\" (UID: \"3cf775f4-cb5f-4fb9-96e3-3fd76976f522\") " pod="openshift-controller-manager/controller-manager-7db5bf594-sbbgc" Feb 17 13:09:30 crc kubenswrapper[4955]: I0217 13:09:30.161546 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8c22eb11-d490-4e5d-9276-a732918e1b37-client-ca\") pod \"route-controller-manager-7584f948bb-dlvtn\" (UID: \"8c22eb11-d490-4e5d-9276-a732918e1b37\") " pod="openshift-route-controller-manager/route-controller-manager-7584f948bb-dlvtn" Feb 17 13:09:30 crc kubenswrapper[4955]: I0217 13:09:30.161723 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3cf775f4-cb5f-4fb9-96e3-3fd76976f522-proxy-ca-bundles\") pod \"controller-manager-7db5bf594-sbbgc\" (UID: \"3cf775f4-cb5f-4fb9-96e3-3fd76976f522\") " pod="openshift-controller-manager/controller-manager-7db5bf594-sbbgc" Feb 17 13:09:30 crc kubenswrapper[4955]: I0217 13:09:30.163032 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3cf775f4-cb5f-4fb9-96e3-3fd76976f522-config\") pod \"controller-manager-7db5bf594-sbbgc\" (UID: \"3cf775f4-cb5f-4fb9-96e3-3fd76976f522\") " pod="openshift-controller-manager/controller-manager-7db5bf594-sbbgc" Feb 17 13:09:30 crc kubenswrapper[4955]: I0217 13:09:30.163626 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3cf775f4-cb5f-4fb9-96e3-3fd76976f522-client-ca\") pod \"controller-manager-7db5bf594-sbbgc\" (UID: \"3cf775f4-cb5f-4fb9-96e3-3fd76976f522\") " pod="openshift-controller-manager/controller-manager-7db5bf594-sbbgc" Feb 17 13:09:30 crc kubenswrapper[4955]: I0217 13:09:30.164315 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3cf775f4-cb5f-4fb9-96e3-3fd76976f522-serving-cert\") pod \"controller-manager-7db5bf594-sbbgc\" (UID: \"3cf775f4-cb5f-4fb9-96e3-3fd76976f522\") " pod="openshift-controller-manager/controller-manager-7db5bf594-sbbgc" Feb 17 13:09:30 crc kubenswrapper[4955]: I0217 13:09:30.164645 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8c22eb11-d490-4e5d-9276-a732918e1b37-serving-cert\") pod \"route-controller-manager-7584f948bb-dlvtn\" (UID: \"8c22eb11-d490-4e5d-9276-a732918e1b37\") " pod="openshift-route-controller-manager/route-controller-manager-7584f948bb-dlvtn" Feb 17 13:09:30 crc kubenswrapper[4955]: I0217 13:09:30.173566 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c22eb11-d490-4e5d-9276-a732918e1b37-config\") pod \"route-controller-manager-7584f948bb-dlvtn\" (UID: \"8c22eb11-d490-4e5d-9276-a732918e1b37\") " pod="openshift-route-controller-manager/route-controller-manager-7584f948bb-dlvtn" Feb 17 13:09:30 crc kubenswrapper[4955]: I0217 13:09:30.181747 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zl46p\" (UniqueName: \"kubernetes.io/projected/8c22eb11-d490-4e5d-9276-a732918e1b37-kube-api-access-zl46p\") pod \"route-controller-manager-7584f948bb-dlvtn\" (UID: \"8c22eb11-d490-4e5d-9276-a732918e1b37\") " pod="openshift-route-controller-manager/route-controller-manager-7584f948bb-dlvtn" Feb 17 13:09:30 crc kubenswrapper[4955]: I0217 13:09:30.185585 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wt85w\" (UniqueName: \"kubernetes.io/projected/3cf775f4-cb5f-4fb9-96e3-3fd76976f522-kube-api-access-wt85w\") pod \"controller-manager-7db5bf594-sbbgc\" (UID: \"3cf775f4-cb5f-4fb9-96e3-3fd76976f522\") " pod="openshift-controller-manager/controller-manager-7db5bf594-sbbgc" Feb 17 13:09:30 crc kubenswrapper[4955]: I0217 13:09:30.231810 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32ea5b76-e8e5-4906-bc55-99d5ec780ff2" path="/var/lib/kubelet/pods/32ea5b76-e8e5-4906-bc55-99d5ec780ff2/volumes" Feb 17 13:09:30 crc kubenswrapper[4955]: I0217 13:09:30.233046 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7db0eeb4-4ffe-4375-a854-0f47efce6a26" path="/var/lib/kubelet/pods/7db0eeb4-4ffe-4375-a854-0f47efce6a26/volumes" Feb 17 13:09:30 crc kubenswrapper[4955]: I0217 13:09:30.266558 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7db5bf594-sbbgc" Feb 17 13:09:30 crc kubenswrapper[4955]: I0217 13:09:30.277797 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7584f948bb-dlvtn" Feb 17 13:09:30 crc kubenswrapper[4955]: I0217 13:09:30.519595 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7db5bf594-sbbgc"] Feb 17 13:09:30 crc kubenswrapper[4955]: I0217 13:09:30.550233 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7584f948bb-dlvtn"] Feb 17 13:09:30 crc kubenswrapper[4955]: W0217 13:09:30.555948 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c22eb11_d490_4e5d_9276_a732918e1b37.slice/crio-2888d3b074d39f3ac3c42091e6f6ed3e0cf9f9919f77a043135586de7dfbdaf3 WatchSource:0}: Error finding container 2888d3b074d39f3ac3c42091e6f6ed3e0cf9f9919f77a043135586de7dfbdaf3: Status 404 returned error can't find the container with id 2888d3b074d39f3ac3c42091e6f6ed3e0cf9f9919f77a043135586de7dfbdaf3 Feb 17 13:09:31 crc kubenswrapper[4955]: I0217 13:09:31.067106 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7db5bf594-sbbgc" event={"ID":"3cf775f4-cb5f-4fb9-96e3-3fd76976f522","Type":"ContainerStarted","Data":"8da4f349481f751e78b7e244a0717bff4607d677ed70d3d208018e7c4d729ea5"} Feb 17 13:09:31 crc kubenswrapper[4955]: I0217 13:09:31.067459 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7db5bf594-sbbgc" Feb 17 13:09:31 crc kubenswrapper[4955]: I0217 13:09:31.067474 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7db5bf594-sbbgc" event={"ID":"3cf775f4-cb5f-4fb9-96e3-3fd76976f522","Type":"ContainerStarted","Data":"98ae23ddb9e1541455bf0716429d701ce20854a8e119debe19560822e99f55bd"} Feb 17 13:09:31 crc kubenswrapper[4955]: I0217 13:09:31.069424 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7584f948bb-dlvtn" event={"ID":"8c22eb11-d490-4e5d-9276-a732918e1b37","Type":"ContainerStarted","Data":"6331427edb08b8c00ee2ebb4a0d2df7b7d9af185a30cbc21cd8e2371d5785a7c"} Feb 17 13:09:31 crc kubenswrapper[4955]: I0217 13:09:31.069479 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7584f948bb-dlvtn" event={"ID":"8c22eb11-d490-4e5d-9276-a732918e1b37","Type":"ContainerStarted","Data":"2888d3b074d39f3ac3c42091e6f6ed3e0cf9f9919f77a043135586de7dfbdaf3"} Feb 17 13:09:31 crc kubenswrapper[4955]: I0217 13:09:31.069811 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7584f948bb-dlvtn" Feb 17 13:09:31 crc kubenswrapper[4955]: I0217 13:09:31.077008 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7db5bf594-sbbgc" Feb 17 13:09:31 crc kubenswrapper[4955]: I0217 13:09:31.087164 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7db5bf594-sbbgc" podStartSLOduration=3.087124707 podStartE2EDuration="3.087124707s" podCreationTimestamp="2026-02-17 13:09:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:09:31.084976623 +0000 UTC m=+309.607706206" watchObservedRunningTime="2026-02-17 13:09:31.087124707 +0000 UTC m=+309.609854260" Feb 17 13:09:31 crc kubenswrapper[4955]: I0217 13:09:31.159818 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7584f948bb-dlvtn" podStartSLOduration=3.159799373 podStartE2EDuration="3.159799373s" podCreationTimestamp="2026-02-17 13:09:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:09:31.157626799 +0000 UTC m=+309.680356372" watchObservedRunningTime="2026-02-17 13:09:31.159799373 +0000 UTC m=+309.682528926" Feb 17 13:09:31 crc kubenswrapper[4955]: I0217 13:09:31.217518 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7584f948bb-dlvtn" Feb 17 13:09:39 crc kubenswrapper[4955]: I0217 13:09:39.365280 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7db5bf594-sbbgc"] Feb 17 13:09:39 crc kubenswrapper[4955]: I0217 13:09:39.365966 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7db5bf594-sbbgc" podUID="3cf775f4-cb5f-4fb9-96e3-3fd76976f522" containerName="controller-manager" containerID="cri-o://8da4f349481f751e78b7e244a0717bff4607d677ed70d3d208018e7c4d729ea5" gracePeriod=30 Feb 17 13:09:39 crc kubenswrapper[4955]: I0217 13:09:39.386271 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7584f948bb-dlvtn"] Feb 17 13:09:39 crc kubenswrapper[4955]: I0217 13:09:39.386463 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-7584f948bb-dlvtn" podUID="8c22eb11-d490-4e5d-9276-a732918e1b37" containerName="route-controller-manager" containerID="cri-o://6331427edb08b8c00ee2ebb4a0d2df7b7d9af185a30cbc21cd8e2371d5785a7c" gracePeriod=30 Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.130563 4955 generic.go:334] "Generic (PLEG): container finished" podID="8c22eb11-d490-4e5d-9276-a732918e1b37" containerID="6331427edb08b8c00ee2ebb4a0d2df7b7d9af185a30cbc21cd8e2371d5785a7c" exitCode=0 Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.130643 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7584f948bb-dlvtn" event={"ID":"8c22eb11-d490-4e5d-9276-a732918e1b37","Type":"ContainerDied","Data":"6331427edb08b8c00ee2ebb4a0d2df7b7d9af185a30cbc21cd8e2371d5785a7c"} Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.132296 4955 generic.go:334] "Generic (PLEG): container finished" podID="3cf775f4-cb5f-4fb9-96e3-3fd76976f522" containerID="8da4f349481f751e78b7e244a0717bff4607d677ed70d3d208018e7c4d729ea5" exitCode=0 Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.132327 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7db5bf594-sbbgc" event={"ID":"3cf775f4-cb5f-4fb9-96e3-3fd76976f522","Type":"ContainerDied","Data":"8da4f349481f751e78b7e244a0717bff4607d677ed70d3d208018e7c4d729ea5"} Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.267574 4955 patch_prober.go:28] interesting pod/controller-manager-7db5bf594-sbbgc container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.57:8443/healthz\": dial tcp 10.217.0.57:8443: connect: connection refused" start-of-body= Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.267642 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-7db5bf594-sbbgc" podUID="3cf775f4-cb5f-4fb9-96e3-3fd76976f522" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.57:8443/healthz\": dial tcp 10.217.0.57:8443: connect: connection refused" Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.279681 4955 patch_prober.go:28] interesting pod/route-controller-manager-7584f948bb-dlvtn container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.58:8443/healthz\": dial tcp 10.217.0.58:8443: connect: connection refused" start-of-body= Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.279742 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-7584f948bb-dlvtn" podUID="8c22eb11-d490-4e5d-9276-a732918e1b37" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.58:8443/healthz\": dial tcp 10.217.0.58:8443: connect: connection refused" Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.681034 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7db5bf594-sbbgc" Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.695657 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7584f948bb-dlvtn" Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.720435 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-74577df4c5-gw5v6"] Feb 17 13:09:40 crc kubenswrapper[4955]: E0217 13:09:40.720696 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c22eb11-d490-4e5d-9276-a732918e1b37" containerName="route-controller-manager" Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.720713 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c22eb11-d490-4e5d-9276-a732918e1b37" containerName="route-controller-manager" Feb 17 13:09:40 crc kubenswrapper[4955]: E0217 13:09:40.720740 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cf775f4-cb5f-4fb9-96e3-3fd76976f522" containerName="controller-manager" Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.720748 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cf775f4-cb5f-4fb9-96e3-3fd76976f522" containerName="controller-manager" Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.720869 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cf775f4-cb5f-4fb9-96e3-3fd76976f522" containerName="controller-manager" Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.720882 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c22eb11-d490-4e5d-9276-a732918e1b37" containerName="route-controller-manager" Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.721244 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74577df4c5-gw5v6" Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.744680 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-74577df4c5-gw5v6"] Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.754175 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8c22eb11-d490-4e5d-9276-a732918e1b37-client-ca\") pod \"8c22eb11-d490-4e5d-9276-a732918e1b37\" (UID: \"8c22eb11-d490-4e5d-9276-a732918e1b37\") " Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.754231 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8c22eb11-d490-4e5d-9276-a732918e1b37-serving-cert\") pod \"8c22eb11-d490-4e5d-9276-a732918e1b37\" (UID: \"8c22eb11-d490-4e5d-9276-a732918e1b37\") " Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.754261 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3cf775f4-cb5f-4fb9-96e3-3fd76976f522-client-ca\") pod \"3cf775f4-cb5f-4fb9-96e3-3fd76976f522\" (UID: \"3cf775f4-cb5f-4fb9-96e3-3fd76976f522\") " Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.754319 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3cf775f4-cb5f-4fb9-96e3-3fd76976f522-serving-cert\") pod \"3cf775f4-cb5f-4fb9-96e3-3fd76976f522\" (UID: \"3cf775f4-cb5f-4fb9-96e3-3fd76976f522\") " Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.754373 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3cf775f4-cb5f-4fb9-96e3-3fd76976f522-config\") pod \"3cf775f4-cb5f-4fb9-96e3-3fd76976f522\" (UID: \"3cf775f4-cb5f-4fb9-96e3-3fd76976f522\") " Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.754431 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3cf775f4-cb5f-4fb9-96e3-3fd76976f522-proxy-ca-bundles\") pod \"3cf775f4-cb5f-4fb9-96e3-3fd76976f522\" (UID: \"3cf775f4-cb5f-4fb9-96e3-3fd76976f522\") " Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.754473 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c22eb11-d490-4e5d-9276-a732918e1b37-config\") pod \"8c22eb11-d490-4e5d-9276-a732918e1b37\" (UID: \"8c22eb11-d490-4e5d-9276-a732918e1b37\") " Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.754505 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zl46p\" (UniqueName: \"kubernetes.io/projected/8c22eb11-d490-4e5d-9276-a732918e1b37-kube-api-access-zl46p\") pod \"8c22eb11-d490-4e5d-9276-a732918e1b37\" (UID: \"8c22eb11-d490-4e5d-9276-a732918e1b37\") " Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.754545 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wt85w\" (UniqueName: \"kubernetes.io/projected/3cf775f4-cb5f-4fb9-96e3-3fd76976f522-kube-api-access-wt85w\") pod \"3cf775f4-cb5f-4fb9-96e3-3fd76976f522\" (UID: \"3cf775f4-cb5f-4fb9-96e3-3fd76976f522\") " Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.754733 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ca3af582-adf9-49a2-a122-7a6d7eacb396-client-ca\") pod \"controller-manager-74577df4c5-gw5v6\" (UID: \"ca3af582-adf9-49a2-a122-7a6d7eacb396\") " pod="openshift-controller-manager/controller-manager-74577df4c5-gw5v6" Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.754802 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ca3af582-adf9-49a2-a122-7a6d7eacb396-serving-cert\") pod \"controller-manager-74577df4c5-gw5v6\" (UID: \"ca3af582-adf9-49a2-a122-7a6d7eacb396\") " pod="openshift-controller-manager/controller-manager-74577df4c5-gw5v6" Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.754830 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hzz4\" (UniqueName: \"kubernetes.io/projected/ca3af582-adf9-49a2-a122-7a6d7eacb396-kube-api-access-5hzz4\") pod \"controller-manager-74577df4c5-gw5v6\" (UID: \"ca3af582-adf9-49a2-a122-7a6d7eacb396\") " pod="openshift-controller-manager/controller-manager-74577df4c5-gw5v6" Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.754856 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ca3af582-adf9-49a2-a122-7a6d7eacb396-proxy-ca-bundles\") pod \"controller-manager-74577df4c5-gw5v6\" (UID: \"ca3af582-adf9-49a2-a122-7a6d7eacb396\") " pod="openshift-controller-manager/controller-manager-74577df4c5-gw5v6" Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.754882 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca3af582-adf9-49a2-a122-7a6d7eacb396-config\") pod \"controller-manager-74577df4c5-gw5v6\" (UID: \"ca3af582-adf9-49a2-a122-7a6d7eacb396\") " pod="openshift-controller-manager/controller-manager-74577df4c5-gw5v6" Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.754996 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c22eb11-d490-4e5d-9276-a732918e1b37-client-ca" (OuterVolumeSpecName: "client-ca") pod "8c22eb11-d490-4e5d-9276-a732918e1b37" (UID: "8c22eb11-d490-4e5d-9276-a732918e1b37"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.757372 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cf775f4-cb5f-4fb9-96e3-3fd76976f522-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "3cf775f4-cb5f-4fb9-96e3-3fd76976f522" (UID: "3cf775f4-cb5f-4fb9-96e3-3fd76976f522"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.757693 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c22eb11-d490-4e5d-9276-a732918e1b37-config" (OuterVolumeSpecName: "config") pod "8c22eb11-d490-4e5d-9276-a732918e1b37" (UID: "8c22eb11-d490-4e5d-9276-a732918e1b37"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.757887 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cf775f4-cb5f-4fb9-96e3-3fd76976f522-client-ca" (OuterVolumeSpecName: "client-ca") pod "3cf775f4-cb5f-4fb9-96e3-3fd76976f522" (UID: "3cf775f4-cb5f-4fb9-96e3-3fd76976f522"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.760581 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cf775f4-cb5f-4fb9-96e3-3fd76976f522-config" (OuterVolumeSpecName: "config") pod "3cf775f4-cb5f-4fb9-96e3-3fd76976f522" (UID: "3cf775f4-cb5f-4fb9-96e3-3fd76976f522"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.760969 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cf775f4-cb5f-4fb9-96e3-3fd76976f522-kube-api-access-wt85w" (OuterVolumeSpecName: "kube-api-access-wt85w") pod "3cf775f4-cb5f-4fb9-96e3-3fd76976f522" (UID: "3cf775f4-cb5f-4fb9-96e3-3fd76976f522"). InnerVolumeSpecName "kube-api-access-wt85w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.760990 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c22eb11-d490-4e5d-9276-a732918e1b37-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8c22eb11-d490-4e5d-9276-a732918e1b37" (UID: "8c22eb11-d490-4e5d-9276-a732918e1b37"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.772836 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cf775f4-cb5f-4fb9-96e3-3fd76976f522-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "3cf775f4-cb5f-4fb9-96e3-3fd76976f522" (UID: "3cf775f4-cb5f-4fb9-96e3-3fd76976f522"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.772894 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c22eb11-d490-4e5d-9276-a732918e1b37-kube-api-access-zl46p" (OuterVolumeSpecName: "kube-api-access-zl46p") pod "8c22eb11-d490-4e5d-9276-a732918e1b37" (UID: "8c22eb11-d490-4e5d-9276-a732918e1b37"). InnerVolumeSpecName "kube-api-access-zl46p". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.855602 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ca3af582-adf9-49a2-a122-7a6d7eacb396-client-ca\") pod \"controller-manager-74577df4c5-gw5v6\" (UID: \"ca3af582-adf9-49a2-a122-7a6d7eacb396\") " pod="openshift-controller-manager/controller-manager-74577df4c5-gw5v6" Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.855685 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ca3af582-adf9-49a2-a122-7a6d7eacb396-serving-cert\") pod \"controller-manager-74577df4c5-gw5v6\" (UID: \"ca3af582-adf9-49a2-a122-7a6d7eacb396\") " pod="openshift-controller-manager/controller-manager-74577df4c5-gw5v6" Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.855713 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hzz4\" (UniqueName: \"kubernetes.io/projected/ca3af582-adf9-49a2-a122-7a6d7eacb396-kube-api-access-5hzz4\") pod \"controller-manager-74577df4c5-gw5v6\" (UID: \"ca3af582-adf9-49a2-a122-7a6d7eacb396\") " pod="openshift-controller-manager/controller-manager-74577df4c5-gw5v6" Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.855740 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ca3af582-adf9-49a2-a122-7a6d7eacb396-proxy-ca-bundles\") pod \"controller-manager-74577df4c5-gw5v6\" (UID: \"ca3af582-adf9-49a2-a122-7a6d7eacb396\") " pod="openshift-controller-manager/controller-manager-74577df4c5-gw5v6" Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.855771 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca3af582-adf9-49a2-a122-7a6d7eacb396-config\") pod \"controller-manager-74577df4c5-gw5v6\" (UID: \"ca3af582-adf9-49a2-a122-7a6d7eacb396\") " pod="openshift-controller-manager/controller-manager-74577df4c5-gw5v6" Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.855858 4955 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3cf775f4-cb5f-4fb9-96e3-3fd76976f522-client-ca\") on node \"crc\" DevicePath \"\"" Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.855872 4955 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3cf775f4-cb5f-4fb9-96e3-3fd76976f522-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.855883 4955 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3cf775f4-cb5f-4fb9-96e3-3fd76976f522-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.855894 4955 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3cf775f4-cb5f-4fb9-96e3-3fd76976f522-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.855905 4955 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c22eb11-d490-4e5d-9276-a732918e1b37-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.855915 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zl46p\" (UniqueName: \"kubernetes.io/projected/8c22eb11-d490-4e5d-9276-a732918e1b37-kube-api-access-zl46p\") on node \"crc\" DevicePath \"\"" Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.855925 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wt85w\" (UniqueName: \"kubernetes.io/projected/3cf775f4-cb5f-4fb9-96e3-3fd76976f522-kube-api-access-wt85w\") on node \"crc\" DevicePath \"\"" Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.855935 4955 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8c22eb11-d490-4e5d-9276-a732918e1b37-client-ca\") on node \"crc\" DevicePath \"\"" Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.855945 4955 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8c22eb11-d490-4e5d-9276-a732918e1b37-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.856625 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ca3af582-adf9-49a2-a122-7a6d7eacb396-client-ca\") pod \"controller-manager-74577df4c5-gw5v6\" (UID: \"ca3af582-adf9-49a2-a122-7a6d7eacb396\") " pod="openshift-controller-manager/controller-manager-74577df4c5-gw5v6" Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.857374 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ca3af582-adf9-49a2-a122-7a6d7eacb396-proxy-ca-bundles\") pod \"controller-manager-74577df4c5-gw5v6\" (UID: \"ca3af582-adf9-49a2-a122-7a6d7eacb396\") " pod="openshift-controller-manager/controller-manager-74577df4c5-gw5v6" Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.857546 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca3af582-adf9-49a2-a122-7a6d7eacb396-config\") pod \"controller-manager-74577df4c5-gw5v6\" (UID: \"ca3af582-adf9-49a2-a122-7a6d7eacb396\") " pod="openshift-controller-manager/controller-manager-74577df4c5-gw5v6" Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.859406 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ca3af582-adf9-49a2-a122-7a6d7eacb396-serving-cert\") pod \"controller-manager-74577df4c5-gw5v6\" (UID: \"ca3af582-adf9-49a2-a122-7a6d7eacb396\") " pod="openshift-controller-manager/controller-manager-74577df4c5-gw5v6" Feb 17 13:09:40 crc kubenswrapper[4955]: I0217 13:09:40.882120 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hzz4\" (UniqueName: \"kubernetes.io/projected/ca3af582-adf9-49a2-a122-7a6d7eacb396-kube-api-access-5hzz4\") pod \"controller-manager-74577df4c5-gw5v6\" (UID: \"ca3af582-adf9-49a2-a122-7a6d7eacb396\") " pod="openshift-controller-manager/controller-manager-74577df4c5-gw5v6" Feb 17 13:09:41 crc kubenswrapper[4955]: I0217 13:09:41.038117 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74577df4c5-gw5v6" Feb 17 13:09:41 crc kubenswrapper[4955]: I0217 13:09:41.146108 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7db5bf594-sbbgc" event={"ID":"3cf775f4-cb5f-4fb9-96e3-3fd76976f522","Type":"ContainerDied","Data":"98ae23ddb9e1541455bf0716429d701ce20854a8e119debe19560822e99f55bd"} Feb 17 13:09:41 crc kubenswrapper[4955]: I0217 13:09:41.146181 4955 scope.go:117] "RemoveContainer" containerID="8da4f349481f751e78b7e244a0717bff4607d677ed70d3d208018e7c4d729ea5" Feb 17 13:09:41 crc kubenswrapper[4955]: I0217 13:09:41.146188 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7db5bf594-sbbgc" Feb 17 13:09:41 crc kubenswrapper[4955]: I0217 13:09:41.153968 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7584f948bb-dlvtn" event={"ID":"8c22eb11-d490-4e5d-9276-a732918e1b37","Type":"ContainerDied","Data":"2888d3b074d39f3ac3c42091e6f6ed3e0cf9f9919f77a043135586de7dfbdaf3"} Feb 17 13:09:41 crc kubenswrapper[4955]: I0217 13:09:41.154044 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7584f948bb-dlvtn" Feb 17 13:09:41 crc kubenswrapper[4955]: I0217 13:09:41.184339 4955 scope.go:117] "RemoveContainer" containerID="6331427edb08b8c00ee2ebb4a0d2df7b7d9af185a30cbc21cd8e2371d5785a7c" Feb 17 13:09:41 crc kubenswrapper[4955]: I0217 13:09:41.244572 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7584f948bb-dlvtn"] Feb 17 13:09:41 crc kubenswrapper[4955]: I0217 13:09:41.246944 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7584f948bb-dlvtn"] Feb 17 13:09:41 crc kubenswrapper[4955]: I0217 13:09:41.258519 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7db5bf594-sbbgc"] Feb 17 13:09:41 crc kubenswrapper[4955]: I0217 13:09:41.261602 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7db5bf594-sbbgc"] Feb 17 13:09:41 crc kubenswrapper[4955]: I0217 13:09:41.596046 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-74577df4c5-gw5v6"] Feb 17 13:09:42 crc kubenswrapper[4955]: I0217 13:09:42.160507 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-74577df4c5-gw5v6" event={"ID":"ca3af582-adf9-49a2-a122-7a6d7eacb396","Type":"ContainerStarted","Data":"c27ed893a2f1c6d4b4025051d9c8f7f0b2a60171322bfe21e6a480f70b73b436"} Feb 17 13:09:42 crc kubenswrapper[4955]: I0217 13:09:42.160560 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-74577df4c5-gw5v6" event={"ID":"ca3af582-adf9-49a2-a122-7a6d7eacb396","Type":"ContainerStarted","Data":"27b18e30f820ddda02f47f33366b3efac74ecbd1a1a89260a058c988c2a6187b"} Feb 17 13:09:42 crc kubenswrapper[4955]: I0217 13:09:42.161766 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-74577df4c5-gw5v6" Feb 17 13:09:42 crc kubenswrapper[4955]: I0217 13:09:42.173842 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-74577df4c5-gw5v6" Feb 17 13:09:42 crc kubenswrapper[4955]: I0217 13:09:42.180610 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-74577df4c5-gw5v6" podStartSLOduration=3.180595561 podStartE2EDuration="3.180595561s" podCreationTimestamp="2026-02-17 13:09:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:09:42.179761496 +0000 UTC m=+320.702491059" watchObservedRunningTime="2026-02-17 13:09:42.180595561 +0000 UTC m=+320.703325124" Feb 17 13:09:42 crc kubenswrapper[4955]: I0217 13:09:42.238136 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cf775f4-cb5f-4fb9-96e3-3fd76976f522" path="/var/lib/kubelet/pods/3cf775f4-cb5f-4fb9-96e3-3fd76976f522/volumes" Feb 17 13:09:42 crc kubenswrapper[4955]: I0217 13:09:42.238846 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c22eb11-d490-4e5d-9276-a732918e1b37" path="/var/lib/kubelet/pods/8c22eb11-d490-4e5d-9276-a732918e1b37/volumes" Feb 17 13:09:42 crc kubenswrapper[4955]: I0217 13:09:42.961395 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c6ddf959-fmhdf"] Feb 17 13:09:42 crc kubenswrapper[4955]: I0217 13:09:42.963636 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5c6ddf959-fmhdf" Feb 17 13:09:42 crc kubenswrapper[4955]: I0217 13:09:42.966758 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 17 13:09:42 crc kubenswrapper[4955]: I0217 13:09:42.967318 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 17 13:09:42 crc kubenswrapper[4955]: I0217 13:09:42.967604 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 17 13:09:42 crc kubenswrapper[4955]: I0217 13:09:42.968028 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 17 13:09:42 crc kubenswrapper[4955]: I0217 13:09:42.968373 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 17 13:09:42 crc kubenswrapper[4955]: I0217 13:09:42.969860 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 17 13:09:42 crc kubenswrapper[4955]: I0217 13:09:42.977168 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c6ddf959-fmhdf"] Feb 17 13:09:42 crc kubenswrapper[4955]: I0217 13:09:42.984011 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61d9e898-4bc3-4e40-aade-42a310dad86d-config\") pod \"route-controller-manager-5c6ddf959-fmhdf\" (UID: \"61d9e898-4bc3-4e40-aade-42a310dad86d\") " pod="openshift-route-controller-manager/route-controller-manager-5c6ddf959-fmhdf" Feb 17 13:09:42 crc kubenswrapper[4955]: I0217 13:09:42.984070 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/61d9e898-4bc3-4e40-aade-42a310dad86d-serving-cert\") pod \"route-controller-manager-5c6ddf959-fmhdf\" (UID: \"61d9e898-4bc3-4e40-aade-42a310dad86d\") " pod="openshift-route-controller-manager/route-controller-manager-5c6ddf959-fmhdf" Feb 17 13:09:42 crc kubenswrapper[4955]: I0217 13:09:42.984167 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2n7pn\" (UniqueName: \"kubernetes.io/projected/61d9e898-4bc3-4e40-aade-42a310dad86d-kube-api-access-2n7pn\") pod \"route-controller-manager-5c6ddf959-fmhdf\" (UID: \"61d9e898-4bc3-4e40-aade-42a310dad86d\") " pod="openshift-route-controller-manager/route-controller-manager-5c6ddf959-fmhdf" Feb 17 13:09:42 crc kubenswrapper[4955]: I0217 13:09:42.984288 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/61d9e898-4bc3-4e40-aade-42a310dad86d-client-ca\") pod \"route-controller-manager-5c6ddf959-fmhdf\" (UID: \"61d9e898-4bc3-4e40-aade-42a310dad86d\") " pod="openshift-route-controller-manager/route-controller-manager-5c6ddf959-fmhdf" Feb 17 13:09:43 crc kubenswrapper[4955]: I0217 13:09:43.088302 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61d9e898-4bc3-4e40-aade-42a310dad86d-config\") pod \"route-controller-manager-5c6ddf959-fmhdf\" (UID: \"61d9e898-4bc3-4e40-aade-42a310dad86d\") " pod="openshift-route-controller-manager/route-controller-manager-5c6ddf959-fmhdf" Feb 17 13:09:43 crc kubenswrapper[4955]: I0217 13:09:43.088396 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/61d9e898-4bc3-4e40-aade-42a310dad86d-serving-cert\") pod \"route-controller-manager-5c6ddf959-fmhdf\" (UID: \"61d9e898-4bc3-4e40-aade-42a310dad86d\") " pod="openshift-route-controller-manager/route-controller-manager-5c6ddf959-fmhdf" Feb 17 13:09:43 crc kubenswrapper[4955]: I0217 13:09:43.088457 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2n7pn\" (UniqueName: \"kubernetes.io/projected/61d9e898-4bc3-4e40-aade-42a310dad86d-kube-api-access-2n7pn\") pod \"route-controller-manager-5c6ddf959-fmhdf\" (UID: \"61d9e898-4bc3-4e40-aade-42a310dad86d\") " pod="openshift-route-controller-manager/route-controller-manager-5c6ddf959-fmhdf" Feb 17 13:09:43 crc kubenswrapper[4955]: I0217 13:09:43.088520 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/61d9e898-4bc3-4e40-aade-42a310dad86d-client-ca\") pod \"route-controller-manager-5c6ddf959-fmhdf\" (UID: \"61d9e898-4bc3-4e40-aade-42a310dad86d\") " pod="openshift-route-controller-manager/route-controller-manager-5c6ddf959-fmhdf" Feb 17 13:09:43 crc kubenswrapper[4955]: I0217 13:09:43.091229 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61d9e898-4bc3-4e40-aade-42a310dad86d-config\") pod \"route-controller-manager-5c6ddf959-fmhdf\" (UID: \"61d9e898-4bc3-4e40-aade-42a310dad86d\") " pod="openshift-route-controller-manager/route-controller-manager-5c6ddf959-fmhdf" Feb 17 13:09:43 crc kubenswrapper[4955]: I0217 13:09:43.093308 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/61d9e898-4bc3-4e40-aade-42a310dad86d-client-ca\") pod \"route-controller-manager-5c6ddf959-fmhdf\" (UID: \"61d9e898-4bc3-4e40-aade-42a310dad86d\") " pod="openshift-route-controller-manager/route-controller-manager-5c6ddf959-fmhdf" Feb 17 13:09:43 crc kubenswrapper[4955]: I0217 13:09:43.103697 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/61d9e898-4bc3-4e40-aade-42a310dad86d-serving-cert\") pod \"route-controller-manager-5c6ddf959-fmhdf\" (UID: \"61d9e898-4bc3-4e40-aade-42a310dad86d\") " pod="openshift-route-controller-manager/route-controller-manager-5c6ddf959-fmhdf" Feb 17 13:09:43 crc kubenswrapper[4955]: I0217 13:09:43.119268 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2n7pn\" (UniqueName: \"kubernetes.io/projected/61d9e898-4bc3-4e40-aade-42a310dad86d-kube-api-access-2n7pn\") pod \"route-controller-manager-5c6ddf959-fmhdf\" (UID: \"61d9e898-4bc3-4e40-aade-42a310dad86d\") " pod="openshift-route-controller-manager/route-controller-manager-5c6ddf959-fmhdf" Feb 17 13:09:43 crc kubenswrapper[4955]: I0217 13:09:43.299728 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5c6ddf959-fmhdf" Feb 17 13:09:43 crc kubenswrapper[4955]: I0217 13:09:43.569472 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c6ddf959-fmhdf"] Feb 17 13:09:44 crc kubenswrapper[4955]: I0217 13:09:44.178738 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5c6ddf959-fmhdf" event={"ID":"61d9e898-4bc3-4e40-aade-42a310dad86d","Type":"ContainerStarted","Data":"69d55b5a36c2d8dcc1a8c048fd95451d5ea430a2cf9aef3bb94f1fb745cf5428"} Feb 17 13:09:44 crc kubenswrapper[4955]: I0217 13:09:44.178833 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5c6ddf959-fmhdf" event={"ID":"61d9e898-4bc3-4e40-aade-42a310dad86d","Type":"ContainerStarted","Data":"ea410f29a61ef118cd33927122c0eb33af45354941622cb278a10f8e613beddf"} Feb 17 13:09:44 crc kubenswrapper[4955]: I0217 13:09:44.179076 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5c6ddf959-fmhdf" Feb 17 13:09:44 crc kubenswrapper[4955]: I0217 13:09:44.197933 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5c6ddf959-fmhdf" podStartSLOduration=5.197915312 podStartE2EDuration="5.197915312s" podCreationTimestamp="2026-02-17 13:09:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:09:44.195734448 +0000 UTC m=+322.718463991" watchObservedRunningTime="2026-02-17 13:09:44.197915312 +0000 UTC m=+322.720644855" Feb 17 13:09:44 crc kubenswrapper[4955]: I0217 13:09:44.324669 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5c6ddf959-fmhdf" Feb 17 13:10:04 crc kubenswrapper[4955]: I0217 13:10:04.574415 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 13:10:04 crc kubenswrapper[4955]: I0217 13:10:04.575079 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 13:10:08 crc kubenswrapper[4955]: I0217 13:10:08.218498 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c6ddf959-fmhdf"] Feb 17 13:10:08 crc kubenswrapper[4955]: I0217 13:10:08.219054 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-5c6ddf959-fmhdf" podUID="61d9e898-4bc3-4e40-aade-42a310dad86d" containerName="route-controller-manager" containerID="cri-o://69d55b5a36c2d8dcc1a8c048fd95451d5ea430a2cf9aef3bb94f1fb745cf5428" gracePeriod=30 Feb 17 13:10:08 crc kubenswrapper[4955]: I0217 13:10:08.725257 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5c6ddf959-fmhdf" Feb 17 13:10:08 crc kubenswrapper[4955]: I0217 13:10:08.873182 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61d9e898-4bc3-4e40-aade-42a310dad86d-config\") pod \"61d9e898-4bc3-4e40-aade-42a310dad86d\" (UID: \"61d9e898-4bc3-4e40-aade-42a310dad86d\") " Feb 17 13:10:08 crc kubenswrapper[4955]: I0217 13:10:08.873282 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/61d9e898-4bc3-4e40-aade-42a310dad86d-serving-cert\") pod \"61d9e898-4bc3-4e40-aade-42a310dad86d\" (UID: \"61d9e898-4bc3-4e40-aade-42a310dad86d\") " Feb 17 13:10:08 crc kubenswrapper[4955]: I0217 13:10:08.873354 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/61d9e898-4bc3-4e40-aade-42a310dad86d-client-ca\") pod \"61d9e898-4bc3-4e40-aade-42a310dad86d\" (UID: \"61d9e898-4bc3-4e40-aade-42a310dad86d\") " Feb 17 13:10:08 crc kubenswrapper[4955]: I0217 13:10:08.873382 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2n7pn\" (UniqueName: \"kubernetes.io/projected/61d9e898-4bc3-4e40-aade-42a310dad86d-kube-api-access-2n7pn\") pod \"61d9e898-4bc3-4e40-aade-42a310dad86d\" (UID: \"61d9e898-4bc3-4e40-aade-42a310dad86d\") " Feb 17 13:10:08 crc kubenswrapper[4955]: I0217 13:10:08.875033 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61d9e898-4bc3-4e40-aade-42a310dad86d-client-ca" (OuterVolumeSpecName: "client-ca") pod "61d9e898-4bc3-4e40-aade-42a310dad86d" (UID: "61d9e898-4bc3-4e40-aade-42a310dad86d"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:10:08 crc kubenswrapper[4955]: I0217 13:10:08.875059 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61d9e898-4bc3-4e40-aade-42a310dad86d-config" (OuterVolumeSpecName: "config") pod "61d9e898-4bc3-4e40-aade-42a310dad86d" (UID: "61d9e898-4bc3-4e40-aade-42a310dad86d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:10:08 crc kubenswrapper[4955]: I0217 13:10:08.882540 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61d9e898-4bc3-4e40-aade-42a310dad86d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "61d9e898-4bc3-4e40-aade-42a310dad86d" (UID: "61d9e898-4bc3-4e40-aade-42a310dad86d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:10:08 crc kubenswrapper[4955]: I0217 13:10:08.882545 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61d9e898-4bc3-4e40-aade-42a310dad86d-kube-api-access-2n7pn" (OuterVolumeSpecName: "kube-api-access-2n7pn") pod "61d9e898-4bc3-4e40-aade-42a310dad86d" (UID: "61d9e898-4bc3-4e40-aade-42a310dad86d"). InnerVolumeSpecName "kube-api-access-2n7pn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:10:08 crc kubenswrapper[4955]: I0217 13:10:08.975045 4955 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/61d9e898-4bc3-4e40-aade-42a310dad86d-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 13:10:08 crc kubenswrapper[4955]: I0217 13:10:08.975075 4955 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/61d9e898-4bc3-4e40-aade-42a310dad86d-client-ca\") on node \"crc\" DevicePath \"\"" Feb 17 13:10:08 crc kubenswrapper[4955]: I0217 13:10:08.975086 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2n7pn\" (UniqueName: \"kubernetes.io/projected/61d9e898-4bc3-4e40-aade-42a310dad86d-kube-api-access-2n7pn\") on node \"crc\" DevicePath \"\"" Feb 17 13:10:08 crc kubenswrapper[4955]: I0217 13:10:08.975096 4955 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61d9e898-4bc3-4e40-aade-42a310dad86d-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:10:09 crc kubenswrapper[4955]: I0217 13:10:09.327934 4955 generic.go:334] "Generic (PLEG): container finished" podID="61d9e898-4bc3-4e40-aade-42a310dad86d" containerID="69d55b5a36c2d8dcc1a8c048fd95451d5ea430a2cf9aef3bb94f1fb745cf5428" exitCode=0 Feb 17 13:10:09 crc kubenswrapper[4955]: I0217 13:10:09.327978 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5c6ddf959-fmhdf" event={"ID":"61d9e898-4bc3-4e40-aade-42a310dad86d","Type":"ContainerDied","Data":"69d55b5a36c2d8dcc1a8c048fd95451d5ea430a2cf9aef3bb94f1fb745cf5428"} Feb 17 13:10:09 crc kubenswrapper[4955]: I0217 13:10:09.328007 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5c6ddf959-fmhdf" event={"ID":"61d9e898-4bc3-4e40-aade-42a310dad86d","Type":"ContainerDied","Data":"ea410f29a61ef118cd33927122c0eb33af45354941622cb278a10f8e613beddf"} Feb 17 13:10:09 crc kubenswrapper[4955]: I0217 13:10:09.328022 4955 scope.go:117] "RemoveContainer" containerID="69d55b5a36c2d8dcc1a8c048fd95451d5ea430a2cf9aef3bb94f1fb745cf5428" Feb 17 13:10:09 crc kubenswrapper[4955]: I0217 13:10:09.328131 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5c6ddf959-fmhdf" Feb 17 13:10:09 crc kubenswrapper[4955]: I0217 13:10:09.349674 4955 scope.go:117] "RemoveContainer" containerID="69d55b5a36c2d8dcc1a8c048fd95451d5ea430a2cf9aef3bb94f1fb745cf5428" Feb 17 13:10:09 crc kubenswrapper[4955]: E0217 13:10:09.358103 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69d55b5a36c2d8dcc1a8c048fd95451d5ea430a2cf9aef3bb94f1fb745cf5428\": container with ID starting with 69d55b5a36c2d8dcc1a8c048fd95451d5ea430a2cf9aef3bb94f1fb745cf5428 not found: ID does not exist" containerID="69d55b5a36c2d8dcc1a8c048fd95451d5ea430a2cf9aef3bb94f1fb745cf5428" Feb 17 13:10:09 crc kubenswrapper[4955]: I0217 13:10:09.358164 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69d55b5a36c2d8dcc1a8c048fd95451d5ea430a2cf9aef3bb94f1fb745cf5428"} err="failed to get container status \"69d55b5a36c2d8dcc1a8c048fd95451d5ea430a2cf9aef3bb94f1fb745cf5428\": rpc error: code = NotFound desc = could not find container \"69d55b5a36c2d8dcc1a8c048fd95451d5ea430a2cf9aef3bb94f1fb745cf5428\": container with ID starting with 69d55b5a36c2d8dcc1a8c048fd95451d5ea430a2cf9aef3bb94f1fb745cf5428 not found: ID does not exist" Feb 17 13:10:09 crc kubenswrapper[4955]: I0217 13:10:09.359634 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c6ddf959-fmhdf"] Feb 17 13:10:09 crc kubenswrapper[4955]: I0217 13:10:09.363892 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c6ddf959-fmhdf"] Feb 17 13:10:09 crc kubenswrapper[4955]: I0217 13:10:09.973330 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8688c6554-bzvmd"] Feb 17 13:10:09 crc kubenswrapper[4955]: E0217 13:10:09.973581 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61d9e898-4bc3-4e40-aade-42a310dad86d" containerName="route-controller-manager" Feb 17 13:10:09 crc kubenswrapper[4955]: I0217 13:10:09.973599 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="61d9e898-4bc3-4e40-aade-42a310dad86d" containerName="route-controller-manager" Feb 17 13:10:09 crc kubenswrapper[4955]: I0217 13:10:09.973746 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="61d9e898-4bc3-4e40-aade-42a310dad86d" containerName="route-controller-manager" Feb 17 13:10:09 crc kubenswrapper[4955]: I0217 13:10:09.974188 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8688c6554-bzvmd" Feb 17 13:10:09 crc kubenswrapper[4955]: I0217 13:10:09.979474 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 17 13:10:09 crc kubenswrapper[4955]: I0217 13:10:09.979695 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 17 13:10:09 crc kubenswrapper[4955]: I0217 13:10:09.979847 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 17 13:10:09 crc kubenswrapper[4955]: I0217 13:10:09.980013 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 17 13:10:09 crc kubenswrapper[4955]: I0217 13:10:09.980078 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 17 13:10:09 crc kubenswrapper[4955]: I0217 13:10:09.980100 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 17 13:10:09 crc kubenswrapper[4955]: I0217 13:10:09.987697 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67ff64a1-761d-44b7-bf01-44c42b714c7b-serving-cert\") pod \"route-controller-manager-8688c6554-bzvmd\" (UID: \"67ff64a1-761d-44b7-bf01-44c42b714c7b\") " pod="openshift-route-controller-manager/route-controller-manager-8688c6554-bzvmd" Feb 17 13:10:09 crc kubenswrapper[4955]: I0217 13:10:09.987745 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/67ff64a1-761d-44b7-bf01-44c42b714c7b-client-ca\") pod \"route-controller-manager-8688c6554-bzvmd\" (UID: \"67ff64a1-761d-44b7-bf01-44c42b714c7b\") " pod="openshift-route-controller-manager/route-controller-manager-8688c6554-bzvmd" Feb 17 13:10:09 crc kubenswrapper[4955]: I0217 13:10:09.987777 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67ff64a1-761d-44b7-bf01-44c42b714c7b-config\") pod \"route-controller-manager-8688c6554-bzvmd\" (UID: \"67ff64a1-761d-44b7-bf01-44c42b714c7b\") " pod="openshift-route-controller-manager/route-controller-manager-8688c6554-bzvmd" Feb 17 13:10:09 crc kubenswrapper[4955]: I0217 13:10:09.987966 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6stm\" (UniqueName: \"kubernetes.io/projected/67ff64a1-761d-44b7-bf01-44c42b714c7b-kube-api-access-v6stm\") pod \"route-controller-manager-8688c6554-bzvmd\" (UID: \"67ff64a1-761d-44b7-bf01-44c42b714c7b\") " pod="openshift-route-controller-manager/route-controller-manager-8688c6554-bzvmd" Feb 17 13:10:09 crc kubenswrapper[4955]: I0217 13:10:09.988747 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8688c6554-bzvmd"] Feb 17 13:10:10 crc kubenswrapper[4955]: I0217 13:10:10.088683 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/67ff64a1-761d-44b7-bf01-44c42b714c7b-client-ca\") pod \"route-controller-manager-8688c6554-bzvmd\" (UID: \"67ff64a1-761d-44b7-bf01-44c42b714c7b\") " pod="openshift-route-controller-manager/route-controller-manager-8688c6554-bzvmd" Feb 17 13:10:10 crc kubenswrapper[4955]: I0217 13:10:10.088738 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67ff64a1-761d-44b7-bf01-44c42b714c7b-serving-cert\") pod \"route-controller-manager-8688c6554-bzvmd\" (UID: \"67ff64a1-761d-44b7-bf01-44c42b714c7b\") " pod="openshift-route-controller-manager/route-controller-manager-8688c6554-bzvmd" Feb 17 13:10:10 crc kubenswrapper[4955]: I0217 13:10:10.088772 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67ff64a1-761d-44b7-bf01-44c42b714c7b-config\") pod \"route-controller-manager-8688c6554-bzvmd\" (UID: \"67ff64a1-761d-44b7-bf01-44c42b714c7b\") " pod="openshift-route-controller-manager/route-controller-manager-8688c6554-bzvmd" Feb 17 13:10:10 crc kubenswrapper[4955]: I0217 13:10:10.088849 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6stm\" (UniqueName: \"kubernetes.io/projected/67ff64a1-761d-44b7-bf01-44c42b714c7b-kube-api-access-v6stm\") pod \"route-controller-manager-8688c6554-bzvmd\" (UID: \"67ff64a1-761d-44b7-bf01-44c42b714c7b\") " pod="openshift-route-controller-manager/route-controller-manager-8688c6554-bzvmd" Feb 17 13:10:10 crc kubenswrapper[4955]: I0217 13:10:10.090439 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/67ff64a1-761d-44b7-bf01-44c42b714c7b-client-ca\") pod \"route-controller-manager-8688c6554-bzvmd\" (UID: \"67ff64a1-761d-44b7-bf01-44c42b714c7b\") " pod="openshift-route-controller-manager/route-controller-manager-8688c6554-bzvmd" Feb 17 13:10:10 crc kubenswrapper[4955]: I0217 13:10:10.090633 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67ff64a1-761d-44b7-bf01-44c42b714c7b-config\") pod \"route-controller-manager-8688c6554-bzvmd\" (UID: \"67ff64a1-761d-44b7-bf01-44c42b714c7b\") " pod="openshift-route-controller-manager/route-controller-manager-8688c6554-bzvmd" Feb 17 13:10:10 crc kubenswrapper[4955]: I0217 13:10:10.094912 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67ff64a1-761d-44b7-bf01-44c42b714c7b-serving-cert\") pod \"route-controller-manager-8688c6554-bzvmd\" (UID: \"67ff64a1-761d-44b7-bf01-44c42b714c7b\") " pod="openshift-route-controller-manager/route-controller-manager-8688c6554-bzvmd" Feb 17 13:10:10 crc kubenswrapper[4955]: I0217 13:10:10.106465 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6stm\" (UniqueName: \"kubernetes.io/projected/67ff64a1-761d-44b7-bf01-44c42b714c7b-kube-api-access-v6stm\") pod \"route-controller-manager-8688c6554-bzvmd\" (UID: \"67ff64a1-761d-44b7-bf01-44c42b714c7b\") " pod="openshift-route-controller-manager/route-controller-manager-8688c6554-bzvmd" Feb 17 13:10:10 crc kubenswrapper[4955]: I0217 13:10:10.228918 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61d9e898-4bc3-4e40-aade-42a310dad86d" path="/var/lib/kubelet/pods/61d9e898-4bc3-4e40-aade-42a310dad86d/volumes" Feb 17 13:10:10 crc kubenswrapper[4955]: I0217 13:10:10.297089 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8688c6554-bzvmd" Feb 17 13:10:10 crc kubenswrapper[4955]: I0217 13:10:10.737140 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8688c6554-bzvmd"] Feb 17 13:10:11 crc kubenswrapper[4955]: I0217 13:10:11.339675 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8688c6554-bzvmd" event={"ID":"67ff64a1-761d-44b7-bf01-44c42b714c7b","Type":"ContainerStarted","Data":"aa56d56e7e3b481bbc4581c0894af07a33e402fba01845b4d99c9eec7da1f40a"} Feb 17 13:10:11 crc kubenswrapper[4955]: I0217 13:10:11.340063 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-8688c6554-bzvmd" Feb 17 13:10:11 crc kubenswrapper[4955]: I0217 13:10:11.340083 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8688c6554-bzvmd" event={"ID":"67ff64a1-761d-44b7-bf01-44c42b714c7b","Type":"ContainerStarted","Data":"6ad8bcdb31000ea6dc13cba9f7faf519e20958dfa401741afe6fd8227e1647c9"} Feb 17 13:10:11 crc kubenswrapper[4955]: I0217 13:10:11.344083 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-8688c6554-bzvmd" Feb 17 13:10:11 crc kubenswrapper[4955]: I0217 13:10:11.356050 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-8688c6554-bzvmd" podStartSLOduration=3.356035237 podStartE2EDuration="3.356035237s" podCreationTimestamp="2026-02-17 13:10:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:10:11.352658928 +0000 UTC m=+349.875388481" watchObservedRunningTime="2026-02-17 13:10:11.356035237 +0000 UTC m=+349.878764780" Feb 17 13:10:24 crc kubenswrapper[4955]: I0217 13:10:24.297552 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-74pjh"] Feb 17 13:10:24 crc kubenswrapper[4955]: I0217 13:10:24.299757 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-74pjh" Feb 17 13:10:24 crc kubenswrapper[4955]: I0217 13:10:24.317749 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-74pjh"] Feb 17 13:10:24 crc kubenswrapper[4955]: I0217 13:10:24.473402 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-74pjh\" (UID: \"f5fa758b-b43b-49ea-bfe0-5105af318ea7\") " pod="openshift-image-registry/image-registry-66df7c8f76-74pjh" Feb 17 13:10:24 crc kubenswrapper[4955]: I0217 13:10:24.473981 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f5fa758b-b43b-49ea-bfe0-5105af318ea7-ca-trust-extracted\") pod \"image-registry-66df7c8f76-74pjh\" (UID: \"f5fa758b-b43b-49ea-bfe0-5105af318ea7\") " pod="openshift-image-registry/image-registry-66df7c8f76-74pjh" Feb 17 13:10:24 crc kubenswrapper[4955]: I0217 13:10:24.474255 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f5fa758b-b43b-49ea-bfe0-5105af318ea7-registry-tls\") pod \"image-registry-66df7c8f76-74pjh\" (UID: \"f5fa758b-b43b-49ea-bfe0-5105af318ea7\") " pod="openshift-image-registry/image-registry-66df7c8f76-74pjh" Feb 17 13:10:24 crc kubenswrapper[4955]: I0217 13:10:24.477933 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjhl5\" (UniqueName: \"kubernetes.io/projected/f5fa758b-b43b-49ea-bfe0-5105af318ea7-kube-api-access-bjhl5\") pod \"image-registry-66df7c8f76-74pjh\" (UID: \"f5fa758b-b43b-49ea-bfe0-5105af318ea7\") " pod="openshift-image-registry/image-registry-66df7c8f76-74pjh" Feb 17 13:10:24 crc kubenswrapper[4955]: I0217 13:10:24.478002 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f5fa758b-b43b-49ea-bfe0-5105af318ea7-registry-certificates\") pod \"image-registry-66df7c8f76-74pjh\" (UID: \"f5fa758b-b43b-49ea-bfe0-5105af318ea7\") " pod="openshift-image-registry/image-registry-66df7c8f76-74pjh" Feb 17 13:10:24 crc kubenswrapper[4955]: I0217 13:10:24.478046 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f5fa758b-b43b-49ea-bfe0-5105af318ea7-bound-sa-token\") pod \"image-registry-66df7c8f76-74pjh\" (UID: \"f5fa758b-b43b-49ea-bfe0-5105af318ea7\") " pod="openshift-image-registry/image-registry-66df7c8f76-74pjh" Feb 17 13:10:24 crc kubenswrapper[4955]: I0217 13:10:24.478097 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f5fa758b-b43b-49ea-bfe0-5105af318ea7-trusted-ca\") pod \"image-registry-66df7c8f76-74pjh\" (UID: \"f5fa758b-b43b-49ea-bfe0-5105af318ea7\") " pod="openshift-image-registry/image-registry-66df7c8f76-74pjh" Feb 17 13:10:24 crc kubenswrapper[4955]: I0217 13:10:24.478123 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f5fa758b-b43b-49ea-bfe0-5105af318ea7-installation-pull-secrets\") pod \"image-registry-66df7c8f76-74pjh\" (UID: \"f5fa758b-b43b-49ea-bfe0-5105af318ea7\") " pod="openshift-image-registry/image-registry-66df7c8f76-74pjh" Feb 17 13:10:24 crc kubenswrapper[4955]: I0217 13:10:24.512197 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-74pjh\" (UID: \"f5fa758b-b43b-49ea-bfe0-5105af318ea7\") " pod="openshift-image-registry/image-registry-66df7c8f76-74pjh" Feb 17 13:10:24 crc kubenswrapper[4955]: I0217 13:10:24.579203 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f5fa758b-b43b-49ea-bfe0-5105af318ea7-ca-trust-extracted\") pod \"image-registry-66df7c8f76-74pjh\" (UID: \"f5fa758b-b43b-49ea-bfe0-5105af318ea7\") " pod="openshift-image-registry/image-registry-66df7c8f76-74pjh" Feb 17 13:10:24 crc kubenswrapper[4955]: I0217 13:10:24.579286 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f5fa758b-b43b-49ea-bfe0-5105af318ea7-registry-tls\") pod \"image-registry-66df7c8f76-74pjh\" (UID: \"f5fa758b-b43b-49ea-bfe0-5105af318ea7\") " pod="openshift-image-registry/image-registry-66df7c8f76-74pjh" Feb 17 13:10:24 crc kubenswrapper[4955]: I0217 13:10:24.579309 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjhl5\" (UniqueName: \"kubernetes.io/projected/f5fa758b-b43b-49ea-bfe0-5105af318ea7-kube-api-access-bjhl5\") pod \"image-registry-66df7c8f76-74pjh\" (UID: \"f5fa758b-b43b-49ea-bfe0-5105af318ea7\") " pod="openshift-image-registry/image-registry-66df7c8f76-74pjh" Feb 17 13:10:24 crc kubenswrapper[4955]: I0217 13:10:24.579331 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f5fa758b-b43b-49ea-bfe0-5105af318ea7-registry-certificates\") pod \"image-registry-66df7c8f76-74pjh\" (UID: \"f5fa758b-b43b-49ea-bfe0-5105af318ea7\") " pod="openshift-image-registry/image-registry-66df7c8f76-74pjh" Feb 17 13:10:24 crc kubenswrapper[4955]: I0217 13:10:24.579353 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f5fa758b-b43b-49ea-bfe0-5105af318ea7-bound-sa-token\") pod \"image-registry-66df7c8f76-74pjh\" (UID: \"f5fa758b-b43b-49ea-bfe0-5105af318ea7\") " pod="openshift-image-registry/image-registry-66df7c8f76-74pjh" Feb 17 13:10:24 crc kubenswrapper[4955]: I0217 13:10:24.579375 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f5fa758b-b43b-49ea-bfe0-5105af318ea7-trusted-ca\") pod \"image-registry-66df7c8f76-74pjh\" (UID: \"f5fa758b-b43b-49ea-bfe0-5105af318ea7\") " pod="openshift-image-registry/image-registry-66df7c8f76-74pjh" Feb 17 13:10:24 crc kubenswrapper[4955]: I0217 13:10:24.579391 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f5fa758b-b43b-49ea-bfe0-5105af318ea7-installation-pull-secrets\") pod \"image-registry-66df7c8f76-74pjh\" (UID: \"f5fa758b-b43b-49ea-bfe0-5105af318ea7\") " pod="openshift-image-registry/image-registry-66df7c8f76-74pjh" Feb 17 13:10:24 crc kubenswrapper[4955]: I0217 13:10:24.580179 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f5fa758b-b43b-49ea-bfe0-5105af318ea7-ca-trust-extracted\") pod \"image-registry-66df7c8f76-74pjh\" (UID: \"f5fa758b-b43b-49ea-bfe0-5105af318ea7\") " pod="openshift-image-registry/image-registry-66df7c8f76-74pjh" Feb 17 13:10:24 crc kubenswrapper[4955]: I0217 13:10:24.582898 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f5fa758b-b43b-49ea-bfe0-5105af318ea7-registry-certificates\") pod \"image-registry-66df7c8f76-74pjh\" (UID: \"f5fa758b-b43b-49ea-bfe0-5105af318ea7\") " pod="openshift-image-registry/image-registry-66df7c8f76-74pjh" Feb 17 13:10:24 crc kubenswrapper[4955]: I0217 13:10:24.583874 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f5fa758b-b43b-49ea-bfe0-5105af318ea7-trusted-ca\") pod \"image-registry-66df7c8f76-74pjh\" (UID: \"f5fa758b-b43b-49ea-bfe0-5105af318ea7\") " pod="openshift-image-registry/image-registry-66df7c8f76-74pjh" Feb 17 13:10:24 crc kubenswrapper[4955]: I0217 13:10:24.587378 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f5fa758b-b43b-49ea-bfe0-5105af318ea7-registry-tls\") pod \"image-registry-66df7c8f76-74pjh\" (UID: \"f5fa758b-b43b-49ea-bfe0-5105af318ea7\") " pod="openshift-image-registry/image-registry-66df7c8f76-74pjh" Feb 17 13:10:24 crc kubenswrapper[4955]: I0217 13:10:24.591658 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f5fa758b-b43b-49ea-bfe0-5105af318ea7-installation-pull-secrets\") pod \"image-registry-66df7c8f76-74pjh\" (UID: \"f5fa758b-b43b-49ea-bfe0-5105af318ea7\") " pod="openshift-image-registry/image-registry-66df7c8f76-74pjh" Feb 17 13:10:24 crc kubenswrapper[4955]: I0217 13:10:24.594354 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f5fa758b-b43b-49ea-bfe0-5105af318ea7-bound-sa-token\") pod \"image-registry-66df7c8f76-74pjh\" (UID: \"f5fa758b-b43b-49ea-bfe0-5105af318ea7\") " pod="openshift-image-registry/image-registry-66df7c8f76-74pjh" Feb 17 13:10:24 crc kubenswrapper[4955]: I0217 13:10:24.606377 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjhl5\" (UniqueName: \"kubernetes.io/projected/f5fa758b-b43b-49ea-bfe0-5105af318ea7-kube-api-access-bjhl5\") pod \"image-registry-66df7c8f76-74pjh\" (UID: \"f5fa758b-b43b-49ea-bfe0-5105af318ea7\") " pod="openshift-image-registry/image-registry-66df7c8f76-74pjh" Feb 17 13:10:24 crc kubenswrapper[4955]: I0217 13:10:24.628999 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-74pjh" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.048166 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-74pjh"] Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.073474 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6xlxd"] Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.074147 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6xlxd" podUID="6d73442c-9d91-4b5b-b0dd-efc137a55959" containerName="registry-server" containerID="cri-o://a7e789ce5032ffe565391646ae6f22e89fb488f2420d1cb3130d5fba459d3241" gracePeriod=30 Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.083123 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pf44z"] Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.083380 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pf44z" podUID="3b1996f2-ac59-43dd-a18b-baeabe4f5da3" containerName="registry-server" containerID="cri-o://161b16ccd4cf9bad13d106f561a8b041a67853d47710ef575698bd11f4ce17ea" gracePeriod=30 Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.105997 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9c9zp"] Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.106243 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-9c9zp" podUID="84d8965a-d181-4f61-921f-1802347f6606" containerName="marketplace-operator" containerID="cri-o://1973286b55d00e9ed0c19af324b11bd225aa27aa5e3eebd49c9295c83656042d" gracePeriod=30 Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.117299 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mxmt9"] Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.117541 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mxmt9" podUID="700722b8-137a-4fcb-83bb-3cd0806c676c" containerName="registry-server" containerID="cri-o://250cfe09d80307d0ceef83aaf478c513dc79912f5886cb2ab9bee8e6452d58cd" gracePeriod=30 Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.123003 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7cjtv"] Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.123319 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7cjtv" podUID="3b3b3288-6b3f-4a48-ae9a-f087715c7e47" containerName="registry-server" containerID="cri-o://ef06c6e7987a2c938fe2dda1695fd2332c2134b6230674e46cd6fd0b7d28c35a" gracePeriod=30 Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.126488 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-spj8f"] Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.127375 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-spj8f" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.138121 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-spj8f"] Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.293894 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4s5tx\" (UniqueName: \"kubernetes.io/projected/f7954d5c-c985-4421-96e9-24ae232c8ba1-kube-api-access-4s5tx\") pod \"marketplace-operator-79b997595-spj8f\" (UID: \"f7954d5c-c985-4421-96e9-24ae232c8ba1\") " pod="openshift-marketplace/marketplace-operator-79b997595-spj8f" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.293979 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f7954d5c-c985-4421-96e9-24ae232c8ba1-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-spj8f\" (UID: \"f7954d5c-c985-4421-96e9-24ae232c8ba1\") " pod="openshift-marketplace/marketplace-operator-79b997595-spj8f" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.294015 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f7954d5c-c985-4421-96e9-24ae232c8ba1-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-spj8f\" (UID: \"f7954d5c-c985-4421-96e9-24ae232c8ba1\") " pod="openshift-marketplace/marketplace-operator-79b997595-spj8f" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.332445 4955 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-9c9zp container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.332506 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-9c9zp" podUID="84d8965a-d181-4f61-921f-1802347f6606" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.395417 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4s5tx\" (UniqueName: \"kubernetes.io/projected/f7954d5c-c985-4421-96e9-24ae232c8ba1-kube-api-access-4s5tx\") pod \"marketplace-operator-79b997595-spj8f\" (UID: \"f7954d5c-c985-4421-96e9-24ae232c8ba1\") " pod="openshift-marketplace/marketplace-operator-79b997595-spj8f" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.395461 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f7954d5c-c985-4421-96e9-24ae232c8ba1-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-spj8f\" (UID: \"f7954d5c-c985-4421-96e9-24ae232c8ba1\") " pod="openshift-marketplace/marketplace-operator-79b997595-spj8f" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.395508 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f7954d5c-c985-4421-96e9-24ae232c8ba1-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-spj8f\" (UID: \"f7954d5c-c985-4421-96e9-24ae232c8ba1\") " pod="openshift-marketplace/marketplace-operator-79b997595-spj8f" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.397356 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f7954d5c-c985-4421-96e9-24ae232c8ba1-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-spj8f\" (UID: \"f7954d5c-c985-4421-96e9-24ae232c8ba1\") " pod="openshift-marketplace/marketplace-operator-79b997595-spj8f" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.408434 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f7954d5c-c985-4421-96e9-24ae232c8ba1-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-spj8f\" (UID: \"f7954d5c-c985-4421-96e9-24ae232c8ba1\") " pod="openshift-marketplace/marketplace-operator-79b997595-spj8f" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.414970 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4s5tx\" (UniqueName: \"kubernetes.io/projected/f7954d5c-c985-4421-96e9-24ae232c8ba1-kube-api-access-4s5tx\") pod \"marketplace-operator-79b997595-spj8f\" (UID: \"f7954d5c-c985-4421-96e9-24ae232c8ba1\") " pod="openshift-marketplace/marketplace-operator-79b997595-spj8f" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.432809 4955 generic.go:334] "Generic (PLEG): container finished" podID="3b1996f2-ac59-43dd-a18b-baeabe4f5da3" containerID="161b16ccd4cf9bad13d106f561a8b041a67853d47710ef575698bd11f4ce17ea" exitCode=0 Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.432866 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pf44z" event={"ID":"3b1996f2-ac59-43dd-a18b-baeabe4f5da3","Type":"ContainerDied","Data":"161b16ccd4cf9bad13d106f561a8b041a67853d47710ef575698bd11f4ce17ea"} Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.434217 4955 generic.go:334] "Generic (PLEG): container finished" podID="84d8965a-d181-4f61-921f-1802347f6606" containerID="1973286b55d00e9ed0c19af324b11bd225aa27aa5e3eebd49c9295c83656042d" exitCode=0 Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.434252 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9c9zp" event={"ID":"84d8965a-d181-4f61-921f-1802347f6606","Type":"ContainerDied","Data":"1973286b55d00e9ed0c19af324b11bd225aa27aa5e3eebd49c9295c83656042d"} Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.434274 4955 scope.go:117] "RemoveContainer" containerID="11d49f6fc9e716b23a59cefaaaff05e16ef8d468c92c03e402a3224d7476fe31" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.436082 4955 generic.go:334] "Generic (PLEG): container finished" podID="6d73442c-9d91-4b5b-b0dd-efc137a55959" containerID="a7e789ce5032ffe565391646ae6f22e89fb488f2420d1cb3130d5fba459d3241" exitCode=0 Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.436121 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xlxd" event={"ID":"6d73442c-9d91-4b5b-b0dd-efc137a55959","Type":"ContainerDied","Data":"a7e789ce5032ffe565391646ae6f22e89fb488f2420d1cb3130d5fba459d3241"} Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.438287 4955 generic.go:334] "Generic (PLEG): container finished" podID="3b3b3288-6b3f-4a48-ae9a-f087715c7e47" containerID="ef06c6e7987a2c938fe2dda1695fd2332c2134b6230674e46cd6fd0b7d28c35a" exitCode=0 Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.438328 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7cjtv" event={"ID":"3b3b3288-6b3f-4a48-ae9a-f087715c7e47","Type":"ContainerDied","Data":"ef06c6e7987a2c938fe2dda1695fd2332c2134b6230674e46cd6fd0b7d28c35a"} Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.439949 4955 generic.go:334] "Generic (PLEG): container finished" podID="700722b8-137a-4fcb-83bb-3cd0806c676c" containerID="250cfe09d80307d0ceef83aaf478c513dc79912f5886cb2ab9bee8e6452d58cd" exitCode=0 Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.439987 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mxmt9" event={"ID":"700722b8-137a-4fcb-83bb-3cd0806c676c","Type":"ContainerDied","Data":"250cfe09d80307d0ceef83aaf478c513dc79912f5886cb2ab9bee8e6452d58cd"} Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.440724 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-74pjh" event={"ID":"f5fa758b-b43b-49ea-bfe0-5105af318ea7","Type":"ContainerStarted","Data":"c1b92ed9a60462975a532b27673f87aeac6cbcae62ba03b6a032458cdd18b517"} Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.490874 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-spj8f" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.498620 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6xlxd" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.597439 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vj64\" (UniqueName: \"kubernetes.io/projected/6d73442c-9d91-4b5b-b0dd-efc137a55959-kube-api-access-7vj64\") pod \"6d73442c-9d91-4b5b-b0dd-efc137a55959\" (UID: \"6d73442c-9d91-4b5b-b0dd-efc137a55959\") " Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.597822 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d73442c-9d91-4b5b-b0dd-efc137a55959-utilities\") pod \"6d73442c-9d91-4b5b-b0dd-efc137a55959\" (UID: \"6d73442c-9d91-4b5b-b0dd-efc137a55959\") " Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.597856 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d73442c-9d91-4b5b-b0dd-efc137a55959-catalog-content\") pod \"6d73442c-9d91-4b5b-b0dd-efc137a55959\" (UID: \"6d73442c-9d91-4b5b-b0dd-efc137a55959\") " Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.598839 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d73442c-9d91-4b5b-b0dd-efc137a55959-utilities" (OuterVolumeSpecName: "utilities") pod "6d73442c-9d91-4b5b-b0dd-efc137a55959" (UID: "6d73442c-9d91-4b5b-b0dd-efc137a55959"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.607127 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d73442c-9d91-4b5b-b0dd-efc137a55959-kube-api-access-7vj64" (OuterVolumeSpecName: "kube-api-access-7vj64") pod "6d73442c-9d91-4b5b-b0dd-efc137a55959" (UID: "6d73442c-9d91-4b5b-b0dd-efc137a55959"). InnerVolumeSpecName "kube-api-access-7vj64". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.629330 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pf44z" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.644602 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-9c9zp" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.657098 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7cjtv" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.661167 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mxmt9" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.684952 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d73442c-9d91-4b5b-b0dd-efc137a55959-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6d73442c-9d91-4b5b-b0dd-efc137a55959" (UID: "6d73442c-9d91-4b5b-b0dd-efc137a55959"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.698985 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vj64\" (UniqueName: \"kubernetes.io/projected/6d73442c-9d91-4b5b-b0dd-efc137a55959-kube-api-access-7vj64\") on node \"crc\" DevicePath \"\"" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.699013 4955 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d73442c-9d91-4b5b-b0dd-efc137a55959-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.699025 4955 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d73442c-9d91-4b5b-b0dd-efc137a55959-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.799667 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2hfx7\" (UniqueName: \"kubernetes.io/projected/3b3b3288-6b3f-4a48-ae9a-f087715c7e47-kube-api-access-2hfx7\") pod \"3b3b3288-6b3f-4a48-ae9a-f087715c7e47\" (UID: \"3b3b3288-6b3f-4a48-ae9a-f087715c7e47\") " Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.799707 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b3b3288-6b3f-4a48-ae9a-f087715c7e47-catalog-content\") pod \"3b3b3288-6b3f-4a48-ae9a-f087715c7e47\" (UID: \"3b3b3288-6b3f-4a48-ae9a-f087715c7e47\") " Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.799770 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b3b3288-6b3f-4a48-ae9a-f087715c7e47-utilities\") pod \"3b3b3288-6b3f-4a48-ae9a-f087715c7e47\" (UID: \"3b3b3288-6b3f-4a48-ae9a-f087715c7e47\") " Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.799802 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b1996f2-ac59-43dd-a18b-baeabe4f5da3-catalog-content\") pod \"3b1996f2-ac59-43dd-a18b-baeabe4f5da3\" (UID: \"3b1996f2-ac59-43dd-a18b-baeabe4f5da3\") " Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.799831 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2r45q\" (UniqueName: \"kubernetes.io/projected/84d8965a-d181-4f61-921f-1802347f6606-kube-api-access-2r45q\") pod \"84d8965a-d181-4f61-921f-1802347f6606\" (UID: \"84d8965a-d181-4f61-921f-1802347f6606\") " Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.799862 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnckc\" (UniqueName: \"kubernetes.io/projected/700722b8-137a-4fcb-83bb-3cd0806c676c-kube-api-access-rnckc\") pod \"700722b8-137a-4fcb-83bb-3cd0806c676c\" (UID: \"700722b8-137a-4fcb-83bb-3cd0806c676c\") " Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.799881 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/700722b8-137a-4fcb-83bb-3cd0806c676c-catalog-content\") pod \"700722b8-137a-4fcb-83bb-3cd0806c676c\" (UID: \"700722b8-137a-4fcb-83bb-3cd0806c676c\") " Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.799913 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/700722b8-137a-4fcb-83bb-3cd0806c676c-utilities\") pod \"700722b8-137a-4fcb-83bb-3cd0806c676c\" (UID: \"700722b8-137a-4fcb-83bb-3cd0806c676c\") " Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.799932 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b1996f2-ac59-43dd-a18b-baeabe4f5da3-utilities\") pod \"3b1996f2-ac59-43dd-a18b-baeabe4f5da3\" (UID: \"3b1996f2-ac59-43dd-a18b-baeabe4f5da3\") " Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.799955 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/84d8965a-d181-4f61-921f-1802347f6606-marketplace-trusted-ca\") pod \"84d8965a-d181-4f61-921f-1802347f6606\" (UID: \"84d8965a-d181-4f61-921f-1802347f6606\") " Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.799979 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vldh5\" (UniqueName: \"kubernetes.io/projected/3b1996f2-ac59-43dd-a18b-baeabe4f5da3-kube-api-access-vldh5\") pod \"3b1996f2-ac59-43dd-a18b-baeabe4f5da3\" (UID: \"3b1996f2-ac59-43dd-a18b-baeabe4f5da3\") " Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.800017 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/84d8965a-d181-4f61-921f-1802347f6606-marketplace-operator-metrics\") pod \"84d8965a-d181-4f61-921f-1802347f6606\" (UID: \"84d8965a-d181-4f61-921f-1802347f6606\") " Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.801105 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84d8965a-d181-4f61-921f-1802347f6606-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "84d8965a-d181-4f61-921f-1802347f6606" (UID: "84d8965a-d181-4f61-921f-1802347f6606"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.801170 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/700722b8-137a-4fcb-83bb-3cd0806c676c-utilities" (OuterVolumeSpecName: "utilities") pod "700722b8-137a-4fcb-83bb-3cd0806c676c" (UID: "700722b8-137a-4fcb-83bb-3cd0806c676c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.801272 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b1996f2-ac59-43dd-a18b-baeabe4f5da3-utilities" (OuterVolumeSpecName: "utilities") pod "3b1996f2-ac59-43dd-a18b-baeabe4f5da3" (UID: "3b1996f2-ac59-43dd-a18b-baeabe4f5da3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.802038 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b3b3288-6b3f-4a48-ae9a-f087715c7e47-utilities" (OuterVolumeSpecName: "utilities") pod "3b3b3288-6b3f-4a48-ae9a-f087715c7e47" (UID: "3b3b3288-6b3f-4a48-ae9a-f087715c7e47"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.803807 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/700722b8-137a-4fcb-83bb-3cd0806c676c-kube-api-access-rnckc" (OuterVolumeSpecName: "kube-api-access-rnckc") pod "700722b8-137a-4fcb-83bb-3cd0806c676c" (UID: "700722b8-137a-4fcb-83bb-3cd0806c676c"). InnerVolumeSpecName "kube-api-access-rnckc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.805173 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84d8965a-d181-4f61-921f-1802347f6606-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "84d8965a-d181-4f61-921f-1802347f6606" (UID: "84d8965a-d181-4f61-921f-1802347f6606"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.805208 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b3b3288-6b3f-4a48-ae9a-f087715c7e47-kube-api-access-2hfx7" (OuterVolumeSpecName: "kube-api-access-2hfx7") pod "3b3b3288-6b3f-4a48-ae9a-f087715c7e47" (UID: "3b3b3288-6b3f-4a48-ae9a-f087715c7e47"). InnerVolumeSpecName "kube-api-access-2hfx7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.805403 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84d8965a-d181-4f61-921f-1802347f6606-kube-api-access-2r45q" (OuterVolumeSpecName: "kube-api-access-2r45q") pod "84d8965a-d181-4f61-921f-1802347f6606" (UID: "84d8965a-d181-4f61-921f-1802347f6606"). InnerVolumeSpecName "kube-api-access-2r45q". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.805901 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b1996f2-ac59-43dd-a18b-baeabe4f5da3-kube-api-access-vldh5" (OuterVolumeSpecName: "kube-api-access-vldh5") pod "3b1996f2-ac59-43dd-a18b-baeabe4f5da3" (UID: "3b1996f2-ac59-43dd-a18b-baeabe4f5da3"). InnerVolumeSpecName "kube-api-access-vldh5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.827633 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/700722b8-137a-4fcb-83bb-3cd0806c676c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "700722b8-137a-4fcb-83bb-3cd0806c676c" (UID: "700722b8-137a-4fcb-83bb-3cd0806c676c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.858602 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b1996f2-ac59-43dd-a18b-baeabe4f5da3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3b1996f2-ac59-43dd-a18b-baeabe4f5da3" (UID: "3b1996f2-ac59-43dd-a18b-baeabe4f5da3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.901690 4955 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b3b3288-6b3f-4a48-ae9a-f087715c7e47-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.901722 4955 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b1996f2-ac59-43dd-a18b-baeabe4f5da3-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.901734 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2r45q\" (UniqueName: \"kubernetes.io/projected/84d8965a-d181-4f61-921f-1802347f6606-kube-api-access-2r45q\") on node \"crc\" DevicePath \"\"" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.901747 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnckc\" (UniqueName: \"kubernetes.io/projected/700722b8-137a-4fcb-83bb-3cd0806c676c-kube-api-access-rnckc\") on node \"crc\" DevicePath \"\"" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.901756 4955 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/700722b8-137a-4fcb-83bb-3cd0806c676c-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.901765 4955 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/700722b8-137a-4fcb-83bb-3cd0806c676c-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.901775 4955 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b1996f2-ac59-43dd-a18b-baeabe4f5da3-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.901798 4955 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/84d8965a-d181-4f61-921f-1802347f6606-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.901806 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vldh5\" (UniqueName: \"kubernetes.io/projected/3b1996f2-ac59-43dd-a18b-baeabe4f5da3-kube-api-access-vldh5\") on node \"crc\" DevicePath \"\"" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.901816 4955 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/84d8965a-d181-4f61-921f-1802347f6606-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.901825 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2hfx7\" (UniqueName: \"kubernetes.io/projected/3b3b3288-6b3f-4a48-ae9a-f087715c7e47-kube-api-access-2hfx7\") on node \"crc\" DevicePath \"\"" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.944956 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b3b3288-6b3f-4a48-ae9a-f087715c7e47-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3b3b3288-6b3f-4a48-ae9a-f087715c7e47" (UID: "3b3b3288-6b3f-4a48-ae9a-f087715c7e47"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:10:25 crc kubenswrapper[4955]: I0217 13:10:25.993428 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-spj8f"] Feb 17 13:10:26 crc kubenswrapper[4955]: I0217 13:10:26.002763 4955 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b3b3288-6b3f-4a48-ae9a-f087715c7e47-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 13:10:26 crc kubenswrapper[4955]: I0217 13:10:26.447893 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xlxd" event={"ID":"6d73442c-9d91-4b5b-b0dd-efc137a55959","Type":"ContainerDied","Data":"45fabf4277a99d2180ad6afedd56eb75fdf4de846186c002b6a9127577f0eea3"} Feb 17 13:10:26 crc kubenswrapper[4955]: I0217 13:10:26.447939 4955 scope.go:117] "RemoveContainer" containerID="a7e789ce5032ffe565391646ae6f22e89fb488f2420d1cb3130d5fba459d3241" Feb 17 13:10:26 crc kubenswrapper[4955]: I0217 13:10:26.448045 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6xlxd" Feb 17 13:10:26 crc kubenswrapper[4955]: I0217 13:10:26.450916 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7cjtv" event={"ID":"3b3b3288-6b3f-4a48-ae9a-f087715c7e47","Type":"ContainerDied","Data":"848d77e9aa3904615c7e41f94882fc61e4734894e3f404dd43a07d0d4929fd6f"} Feb 17 13:10:26 crc kubenswrapper[4955]: I0217 13:10:26.450959 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7cjtv" Feb 17 13:10:26 crc kubenswrapper[4955]: I0217 13:10:26.453384 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mxmt9" event={"ID":"700722b8-137a-4fcb-83bb-3cd0806c676c","Type":"ContainerDied","Data":"7ebd289f4ccb08edcba5ca62d06e44c968ea5e89217e8c8ab97af99902390498"} Feb 17 13:10:26 crc kubenswrapper[4955]: I0217 13:10:26.453445 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mxmt9" Feb 17 13:10:26 crc kubenswrapper[4955]: I0217 13:10:26.457934 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-74pjh" event={"ID":"f5fa758b-b43b-49ea-bfe0-5105af318ea7","Type":"ContainerStarted","Data":"36fb86e4b51761471a296abe3e5f8a448efc80084f7553ccfadfea08df3b0cd9"} Feb 17 13:10:26 crc kubenswrapper[4955]: I0217 13:10:26.458062 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-74pjh" Feb 17 13:10:26 crc kubenswrapper[4955]: I0217 13:10:26.465029 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pf44z" event={"ID":"3b1996f2-ac59-43dd-a18b-baeabe4f5da3","Type":"ContainerDied","Data":"072f6579e746217ed7519c11635ff385805ef526c07510840caa7c5b45b62b71"} Feb 17 13:10:26 crc kubenswrapper[4955]: I0217 13:10:26.465261 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pf44z" Feb 17 13:10:26 crc kubenswrapper[4955]: I0217 13:10:26.467190 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-spj8f" event={"ID":"f7954d5c-c985-4421-96e9-24ae232c8ba1","Type":"ContainerStarted","Data":"94dbbc49de348cac2e834cd220968e09000a20b0114c5cec61785427ee8fe965"} Feb 17 13:10:26 crc kubenswrapper[4955]: I0217 13:10:26.467330 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-spj8f" event={"ID":"f7954d5c-c985-4421-96e9-24ae232c8ba1","Type":"ContainerStarted","Data":"9434eb07137e070223c518fdeb3dd45e8332f7a3bd6ec15b669c5fb0beb0a450"} Feb 17 13:10:26 crc kubenswrapper[4955]: I0217 13:10:26.468978 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-spj8f" Feb 17 13:10:26 crc kubenswrapper[4955]: I0217 13:10:26.470292 4955 scope.go:117] "RemoveContainer" containerID="eecb2ead972ac212b57fbc4fa9ac053eca89ad687e169c499ea9df1790eb70d9" Feb 17 13:10:26 crc kubenswrapper[4955]: I0217 13:10:26.470943 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9c9zp" event={"ID":"84d8965a-d181-4f61-921f-1802347f6606","Type":"ContainerDied","Data":"c1840dd14017cbc9ef34d4d29a76e441fab814a35992f6aaa4a0948828651f3e"} Feb 17 13:10:26 crc kubenswrapper[4955]: I0217 13:10:26.471216 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-9c9zp" Feb 17 13:10:26 crc kubenswrapper[4955]: I0217 13:10:26.476299 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-spj8f" Feb 17 13:10:26 crc kubenswrapper[4955]: I0217 13:10:26.494969 4955 scope.go:117] "RemoveContainer" containerID="c973f82be5e4407d035f3ecc76b729686a60f36aede147e0901b6827226bc268" Feb 17 13:10:26 crc kubenswrapper[4955]: I0217 13:10:26.499734 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-74pjh" podStartSLOduration=2.499716215 podStartE2EDuration="2.499716215s" podCreationTimestamp="2026-02-17 13:10:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:10:26.483415174 +0000 UTC m=+365.006144717" watchObservedRunningTime="2026-02-17 13:10:26.499716215 +0000 UTC m=+365.022445758" Feb 17 13:10:26 crc kubenswrapper[4955]: I0217 13:10:26.516596 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6xlxd"] Feb 17 13:10:26 crc kubenswrapper[4955]: I0217 13:10:26.528121 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6xlxd"] Feb 17 13:10:26 crc kubenswrapper[4955]: I0217 13:10:26.533386 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mxmt9"] Feb 17 13:10:26 crc kubenswrapper[4955]: I0217 13:10:26.538224 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mxmt9"] Feb 17 13:10:26 crc kubenswrapper[4955]: I0217 13:10:26.544975 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7cjtv"] Feb 17 13:10:26 crc kubenswrapper[4955]: I0217 13:10:26.549453 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7cjtv"] Feb 17 13:10:26 crc kubenswrapper[4955]: I0217 13:10:26.550526 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-spj8f" podStartSLOduration=1.550513255 podStartE2EDuration="1.550513255s" podCreationTimestamp="2026-02-17 13:10:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:10:26.540246212 +0000 UTC m=+365.062975775" watchObservedRunningTime="2026-02-17 13:10:26.550513255 +0000 UTC m=+365.073242798" Feb 17 13:10:26 crc kubenswrapper[4955]: I0217 13:10:26.557535 4955 scope.go:117] "RemoveContainer" containerID="ef06c6e7987a2c938fe2dda1695fd2332c2134b6230674e46cd6fd0b7d28c35a" Feb 17 13:10:26 crc kubenswrapper[4955]: I0217 13:10:26.559215 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9c9zp"] Feb 17 13:10:26 crc kubenswrapper[4955]: I0217 13:10:26.562080 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9c9zp"] Feb 17 13:10:26 crc kubenswrapper[4955]: I0217 13:10:26.572084 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pf44z"] Feb 17 13:10:26 crc kubenswrapper[4955]: I0217 13:10:26.579019 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pf44z"] Feb 17 13:10:26 crc kubenswrapper[4955]: I0217 13:10:26.590052 4955 scope.go:117] "RemoveContainer" containerID="47b4525f063afdd024c6414642a647009881bf4f0c4cf5d2b1743d1536ab81eb" Feb 17 13:10:26 crc kubenswrapper[4955]: I0217 13:10:26.609667 4955 scope.go:117] "RemoveContainer" containerID="7d89db3dc6759750846d4e0d9a51140f2f4b1db903871367b498b94a29987160" Feb 17 13:10:26 crc kubenswrapper[4955]: I0217 13:10:26.626502 4955 scope.go:117] "RemoveContainer" containerID="250cfe09d80307d0ceef83aaf478c513dc79912f5886cb2ab9bee8e6452d58cd" Feb 17 13:10:26 crc kubenswrapper[4955]: I0217 13:10:26.640208 4955 scope.go:117] "RemoveContainer" containerID="968555c3f7f2323d0d9605a31e076b3396d018e1fee294aa23ad245cd2eac7b6" Feb 17 13:10:26 crc kubenswrapper[4955]: I0217 13:10:26.654243 4955 scope.go:117] "RemoveContainer" containerID="a876575f7716e64e2b9eb9fa534c5536269b1a2555b420ab2046d4e281b3fe80" Feb 17 13:10:26 crc kubenswrapper[4955]: I0217 13:10:26.669267 4955 scope.go:117] "RemoveContainer" containerID="161b16ccd4cf9bad13d106f561a8b041a67853d47710ef575698bd11f4ce17ea" Feb 17 13:10:26 crc kubenswrapper[4955]: I0217 13:10:26.681743 4955 scope.go:117] "RemoveContainer" containerID="91aaf6a67d05a4134b039fc74edefcae2f346e9197261e82b421cb391020989e" Feb 17 13:10:26 crc kubenswrapper[4955]: I0217 13:10:26.699977 4955 scope.go:117] "RemoveContainer" containerID="33a9b4f3a61747aceb04131eff2af6ceea8dbeb951a20741327b4b4b44057854" Feb 17 13:10:26 crc kubenswrapper[4955]: I0217 13:10:26.716893 4955 scope.go:117] "RemoveContainer" containerID="1973286b55d00e9ed0c19af324b11bd225aa27aa5e3eebd49c9295c83656042d" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.485892 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jmrnj"] Feb 17 13:10:27 crc kubenswrapper[4955]: E0217 13:10:27.486088 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b1996f2-ac59-43dd-a18b-baeabe4f5da3" containerName="extract-content" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.486100 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b1996f2-ac59-43dd-a18b-baeabe4f5da3" containerName="extract-content" Feb 17 13:10:27 crc kubenswrapper[4955]: E0217 13:10:27.486109 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d73442c-9d91-4b5b-b0dd-efc137a55959" containerName="extract-content" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.486115 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d73442c-9d91-4b5b-b0dd-efc137a55959" containerName="extract-content" Feb 17 13:10:27 crc kubenswrapper[4955]: E0217 13:10:27.486123 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b3b3288-6b3f-4a48-ae9a-f087715c7e47" containerName="extract-content" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.486130 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b3b3288-6b3f-4a48-ae9a-f087715c7e47" containerName="extract-content" Feb 17 13:10:27 crc kubenswrapper[4955]: E0217 13:10:27.486138 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="700722b8-137a-4fcb-83bb-3cd0806c676c" containerName="registry-server" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.486143 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="700722b8-137a-4fcb-83bb-3cd0806c676c" containerName="registry-server" Feb 17 13:10:27 crc kubenswrapper[4955]: E0217 13:10:27.486149 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84d8965a-d181-4f61-921f-1802347f6606" containerName="marketplace-operator" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.486154 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="84d8965a-d181-4f61-921f-1802347f6606" containerName="marketplace-operator" Feb 17 13:10:27 crc kubenswrapper[4955]: E0217 13:10:27.486163 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d73442c-9d91-4b5b-b0dd-efc137a55959" containerName="extract-utilities" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.486169 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d73442c-9d91-4b5b-b0dd-efc137a55959" containerName="extract-utilities" Feb 17 13:10:27 crc kubenswrapper[4955]: E0217 13:10:27.486179 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="700722b8-137a-4fcb-83bb-3cd0806c676c" containerName="extract-content" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.486185 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="700722b8-137a-4fcb-83bb-3cd0806c676c" containerName="extract-content" Feb 17 13:10:27 crc kubenswrapper[4955]: E0217 13:10:27.486192 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d73442c-9d91-4b5b-b0dd-efc137a55959" containerName="registry-server" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.486198 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d73442c-9d91-4b5b-b0dd-efc137a55959" containerName="registry-server" Feb 17 13:10:27 crc kubenswrapper[4955]: E0217 13:10:27.486208 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="700722b8-137a-4fcb-83bb-3cd0806c676c" containerName="extract-utilities" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.486214 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="700722b8-137a-4fcb-83bb-3cd0806c676c" containerName="extract-utilities" Feb 17 13:10:27 crc kubenswrapper[4955]: E0217 13:10:27.486223 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b1996f2-ac59-43dd-a18b-baeabe4f5da3" containerName="registry-server" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.486228 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b1996f2-ac59-43dd-a18b-baeabe4f5da3" containerName="registry-server" Feb 17 13:10:27 crc kubenswrapper[4955]: E0217 13:10:27.486235 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b1996f2-ac59-43dd-a18b-baeabe4f5da3" containerName="extract-utilities" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.486242 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b1996f2-ac59-43dd-a18b-baeabe4f5da3" containerName="extract-utilities" Feb 17 13:10:27 crc kubenswrapper[4955]: E0217 13:10:27.486252 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b3b3288-6b3f-4a48-ae9a-f087715c7e47" containerName="extract-utilities" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.486257 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b3b3288-6b3f-4a48-ae9a-f087715c7e47" containerName="extract-utilities" Feb 17 13:10:27 crc kubenswrapper[4955]: E0217 13:10:27.486267 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b3b3288-6b3f-4a48-ae9a-f087715c7e47" containerName="registry-server" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.486273 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b3b3288-6b3f-4a48-ae9a-f087715c7e47" containerName="registry-server" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.486355 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="84d8965a-d181-4f61-921f-1802347f6606" containerName="marketplace-operator" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.486364 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="84d8965a-d181-4f61-921f-1802347f6606" containerName="marketplace-operator" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.486371 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d73442c-9d91-4b5b-b0dd-efc137a55959" containerName="registry-server" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.486380 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b1996f2-ac59-43dd-a18b-baeabe4f5da3" containerName="registry-server" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.486390 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="700722b8-137a-4fcb-83bb-3cd0806c676c" containerName="registry-server" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.486398 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b3b3288-6b3f-4a48-ae9a-f087715c7e47" containerName="registry-server" Feb 17 13:10:27 crc kubenswrapper[4955]: E0217 13:10:27.486473 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84d8965a-d181-4f61-921f-1802347f6606" containerName="marketplace-operator" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.486480 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="84d8965a-d181-4f61-921f-1802347f6606" containerName="marketplace-operator" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.487049 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jmrnj" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.491427 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.502161 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jmrnj"] Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.624553 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1cf150d2-8598-45d3-b193-4d27e90ca648-catalog-content\") pod \"redhat-marketplace-jmrnj\" (UID: \"1cf150d2-8598-45d3-b193-4d27e90ca648\") " pod="openshift-marketplace/redhat-marketplace-jmrnj" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.624621 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1cf150d2-8598-45d3-b193-4d27e90ca648-utilities\") pod \"redhat-marketplace-jmrnj\" (UID: \"1cf150d2-8598-45d3-b193-4d27e90ca648\") " pod="openshift-marketplace/redhat-marketplace-jmrnj" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.624669 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbl8f\" (UniqueName: \"kubernetes.io/projected/1cf150d2-8598-45d3-b193-4d27e90ca648-kube-api-access-fbl8f\") pod \"redhat-marketplace-jmrnj\" (UID: \"1cf150d2-8598-45d3-b193-4d27e90ca648\") " pod="openshift-marketplace/redhat-marketplace-jmrnj" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.704464 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qz5rg"] Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.705609 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qz5rg" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.708621 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.718041 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qz5rg"] Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.725975 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1cf150d2-8598-45d3-b193-4d27e90ca648-catalog-content\") pod \"redhat-marketplace-jmrnj\" (UID: \"1cf150d2-8598-45d3-b193-4d27e90ca648\") " pod="openshift-marketplace/redhat-marketplace-jmrnj" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.726010 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1cf150d2-8598-45d3-b193-4d27e90ca648-utilities\") pod \"redhat-marketplace-jmrnj\" (UID: \"1cf150d2-8598-45d3-b193-4d27e90ca648\") " pod="openshift-marketplace/redhat-marketplace-jmrnj" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.726039 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbl8f\" (UniqueName: \"kubernetes.io/projected/1cf150d2-8598-45d3-b193-4d27e90ca648-kube-api-access-fbl8f\") pod \"redhat-marketplace-jmrnj\" (UID: \"1cf150d2-8598-45d3-b193-4d27e90ca648\") " pod="openshift-marketplace/redhat-marketplace-jmrnj" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.726629 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1cf150d2-8598-45d3-b193-4d27e90ca648-catalog-content\") pod \"redhat-marketplace-jmrnj\" (UID: \"1cf150d2-8598-45d3-b193-4d27e90ca648\") " pod="openshift-marketplace/redhat-marketplace-jmrnj" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.728108 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1cf150d2-8598-45d3-b193-4d27e90ca648-utilities\") pod \"redhat-marketplace-jmrnj\" (UID: \"1cf150d2-8598-45d3-b193-4d27e90ca648\") " pod="openshift-marketplace/redhat-marketplace-jmrnj" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.748547 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbl8f\" (UniqueName: \"kubernetes.io/projected/1cf150d2-8598-45d3-b193-4d27e90ca648-kube-api-access-fbl8f\") pod \"redhat-marketplace-jmrnj\" (UID: \"1cf150d2-8598-45d3-b193-4d27e90ca648\") " pod="openshift-marketplace/redhat-marketplace-jmrnj" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.819563 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jmrnj" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.827475 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zb6dh\" (UniqueName: \"kubernetes.io/projected/a5c8c29f-7770-4ba6-aad9-12604d10981f-kube-api-access-zb6dh\") pod \"redhat-operators-qz5rg\" (UID: \"a5c8c29f-7770-4ba6-aad9-12604d10981f\") " pod="openshift-marketplace/redhat-operators-qz5rg" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.827546 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5c8c29f-7770-4ba6-aad9-12604d10981f-utilities\") pod \"redhat-operators-qz5rg\" (UID: \"a5c8c29f-7770-4ba6-aad9-12604d10981f\") " pod="openshift-marketplace/redhat-operators-qz5rg" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.827619 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5c8c29f-7770-4ba6-aad9-12604d10981f-catalog-content\") pod \"redhat-operators-qz5rg\" (UID: \"a5c8c29f-7770-4ba6-aad9-12604d10981f\") " pod="openshift-marketplace/redhat-operators-qz5rg" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.928923 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5c8c29f-7770-4ba6-aad9-12604d10981f-catalog-content\") pod \"redhat-operators-qz5rg\" (UID: \"a5c8c29f-7770-4ba6-aad9-12604d10981f\") " pod="openshift-marketplace/redhat-operators-qz5rg" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.929256 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zb6dh\" (UniqueName: \"kubernetes.io/projected/a5c8c29f-7770-4ba6-aad9-12604d10981f-kube-api-access-zb6dh\") pod \"redhat-operators-qz5rg\" (UID: \"a5c8c29f-7770-4ba6-aad9-12604d10981f\") " pod="openshift-marketplace/redhat-operators-qz5rg" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.929286 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5c8c29f-7770-4ba6-aad9-12604d10981f-utilities\") pod \"redhat-operators-qz5rg\" (UID: \"a5c8c29f-7770-4ba6-aad9-12604d10981f\") " pod="openshift-marketplace/redhat-operators-qz5rg" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.929698 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5c8c29f-7770-4ba6-aad9-12604d10981f-utilities\") pod \"redhat-operators-qz5rg\" (UID: \"a5c8c29f-7770-4ba6-aad9-12604d10981f\") " pod="openshift-marketplace/redhat-operators-qz5rg" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.930082 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5c8c29f-7770-4ba6-aad9-12604d10981f-catalog-content\") pod \"redhat-operators-qz5rg\" (UID: \"a5c8c29f-7770-4ba6-aad9-12604d10981f\") " pod="openshift-marketplace/redhat-operators-qz5rg" Feb 17 13:10:27 crc kubenswrapper[4955]: I0217 13:10:27.947129 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zb6dh\" (UniqueName: \"kubernetes.io/projected/a5c8c29f-7770-4ba6-aad9-12604d10981f-kube-api-access-zb6dh\") pod \"redhat-operators-qz5rg\" (UID: \"a5c8c29f-7770-4ba6-aad9-12604d10981f\") " pod="openshift-marketplace/redhat-operators-qz5rg" Feb 17 13:10:28 crc kubenswrapper[4955]: I0217 13:10:28.019458 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qz5rg" Feb 17 13:10:28 crc kubenswrapper[4955]: I0217 13:10:28.220996 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jmrnj"] Feb 17 13:10:28 crc kubenswrapper[4955]: I0217 13:10:28.235965 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b1996f2-ac59-43dd-a18b-baeabe4f5da3" path="/var/lib/kubelet/pods/3b1996f2-ac59-43dd-a18b-baeabe4f5da3/volumes" Feb 17 13:10:28 crc kubenswrapper[4955]: I0217 13:10:28.236752 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b3b3288-6b3f-4a48-ae9a-f087715c7e47" path="/var/lib/kubelet/pods/3b3b3288-6b3f-4a48-ae9a-f087715c7e47/volumes" Feb 17 13:10:28 crc kubenswrapper[4955]: I0217 13:10:28.237328 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d73442c-9d91-4b5b-b0dd-efc137a55959" path="/var/lib/kubelet/pods/6d73442c-9d91-4b5b-b0dd-efc137a55959/volumes" Feb 17 13:10:28 crc kubenswrapper[4955]: I0217 13:10:28.238474 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="700722b8-137a-4fcb-83bb-3cd0806c676c" path="/var/lib/kubelet/pods/700722b8-137a-4fcb-83bb-3cd0806c676c/volumes" Feb 17 13:10:28 crc kubenswrapper[4955]: I0217 13:10:28.239092 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84d8965a-d181-4f61-921f-1802347f6606" path="/var/lib/kubelet/pods/84d8965a-d181-4f61-921f-1802347f6606/volumes" Feb 17 13:10:28 crc kubenswrapper[4955]: I0217 13:10:28.244517 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-74577df4c5-gw5v6"] Feb 17 13:10:28 crc kubenswrapper[4955]: I0217 13:10:28.244876 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-74577df4c5-gw5v6" podUID="ca3af582-adf9-49a2-a122-7a6d7eacb396" containerName="controller-manager" containerID="cri-o://c27ed893a2f1c6d4b4025051d9c8f7f0b2a60171322bfe21e6a480f70b73b436" gracePeriod=30 Feb 17 13:10:28 crc kubenswrapper[4955]: I0217 13:10:28.270376 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qz5rg"] Feb 17 13:10:28 crc kubenswrapper[4955]: W0217 13:10:28.276113 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5c8c29f_7770_4ba6_aad9_12604d10981f.slice/crio-7d360e3640e23f1c04dcb3364401331fc02e3b2c180902622aa50f0238c7a5f3 WatchSource:0}: Error finding container 7d360e3640e23f1c04dcb3364401331fc02e3b2c180902622aa50f0238c7a5f3: Status 404 returned error can't find the container with id 7d360e3640e23f1c04dcb3364401331fc02e3b2c180902622aa50f0238c7a5f3 Feb 17 13:10:28 crc kubenswrapper[4955]: I0217 13:10:28.512312 4955 generic.go:334] "Generic (PLEG): container finished" podID="1cf150d2-8598-45d3-b193-4d27e90ca648" containerID="3932e75631a835de356435c7b1e7145b0fad2659b9e3688fb268d56a2eeda63d" exitCode=0 Feb 17 13:10:28 crc kubenswrapper[4955]: I0217 13:10:28.512405 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jmrnj" event={"ID":"1cf150d2-8598-45d3-b193-4d27e90ca648","Type":"ContainerDied","Data":"3932e75631a835de356435c7b1e7145b0fad2659b9e3688fb268d56a2eeda63d"} Feb 17 13:10:28 crc kubenswrapper[4955]: I0217 13:10:28.512443 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jmrnj" event={"ID":"1cf150d2-8598-45d3-b193-4d27e90ca648","Type":"ContainerStarted","Data":"f9f2fac6e0f86c53c4255fb316a3154eb4f5b385984d5eab0def58df57fe08c1"} Feb 17 13:10:28 crc kubenswrapper[4955]: I0217 13:10:28.514651 4955 generic.go:334] "Generic (PLEG): container finished" podID="a5c8c29f-7770-4ba6-aad9-12604d10981f" containerID="03e8cbb19576142f6af8d55a5a11d8c681d9b142be2da52d4d5d759949b83ba2" exitCode=0 Feb 17 13:10:28 crc kubenswrapper[4955]: I0217 13:10:28.514705 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qz5rg" event={"ID":"a5c8c29f-7770-4ba6-aad9-12604d10981f","Type":"ContainerDied","Data":"03e8cbb19576142f6af8d55a5a11d8c681d9b142be2da52d4d5d759949b83ba2"} Feb 17 13:10:28 crc kubenswrapper[4955]: I0217 13:10:28.514727 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qz5rg" event={"ID":"a5c8c29f-7770-4ba6-aad9-12604d10981f","Type":"ContainerStarted","Data":"7d360e3640e23f1c04dcb3364401331fc02e3b2c180902622aa50f0238c7a5f3"} Feb 17 13:10:28 crc kubenswrapper[4955]: I0217 13:10:28.518866 4955 generic.go:334] "Generic (PLEG): container finished" podID="ca3af582-adf9-49a2-a122-7a6d7eacb396" containerID="c27ed893a2f1c6d4b4025051d9c8f7f0b2a60171322bfe21e6a480f70b73b436" exitCode=0 Feb 17 13:10:28 crc kubenswrapper[4955]: I0217 13:10:28.518921 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-74577df4c5-gw5v6" event={"ID":"ca3af582-adf9-49a2-a122-7a6d7eacb396","Type":"ContainerDied","Data":"c27ed893a2f1c6d4b4025051d9c8f7f0b2a60171322bfe21e6a480f70b73b436"} Feb 17 13:10:28 crc kubenswrapper[4955]: I0217 13:10:28.681296 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74577df4c5-gw5v6" Feb 17 13:10:28 crc kubenswrapper[4955]: I0217 13:10:28.741121 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ca3af582-adf9-49a2-a122-7a6d7eacb396-client-ca\") pod \"ca3af582-adf9-49a2-a122-7a6d7eacb396\" (UID: \"ca3af582-adf9-49a2-a122-7a6d7eacb396\") " Feb 17 13:10:28 crc kubenswrapper[4955]: I0217 13:10:28.741239 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca3af582-adf9-49a2-a122-7a6d7eacb396-config\") pod \"ca3af582-adf9-49a2-a122-7a6d7eacb396\" (UID: \"ca3af582-adf9-49a2-a122-7a6d7eacb396\") " Feb 17 13:10:28 crc kubenswrapper[4955]: I0217 13:10:28.741272 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ca3af582-adf9-49a2-a122-7a6d7eacb396-proxy-ca-bundles\") pod \"ca3af582-adf9-49a2-a122-7a6d7eacb396\" (UID: \"ca3af582-adf9-49a2-a122-7a6d7eacb396\") " Feb 17 13:10:28 crc kubenswrapper[4955]: I0217 13:10:28.741297 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ca3af582-adf9-49a2-a122-7a6d7eacb396-serving-cert\") pod \"ca3af582-adf9-49a2-a122-7a6d7eacb396\" (UID: \"ca3af582-adf9-49a2-a122-7a6d7eacb396\") " Feb 17 13:10:28 crc kubenswrapper[4955]: I0217 13:10:28.741331 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hzz4\" (UniqueName: \"kubernetes.io/projected/ca3af582-adf9-49a2-a122-7a6d7eacb396-kube-api-access-5hzz4\") pod \"ca3af582-adf9-49a2-a122-7a6d7eacb396\" (UID: \"ca3af582-adf9-49a2-a122-7a6d7eacb396\") " Feb 17 13:10:28 crc kubenswrapper[4955]: I0217 13:10:28.742024 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca3af582-adf9-49a2-a122-7a6d7eacb396-client-ca" (OuterVolumeSpecName: "client-ca") pod "ca3af582-adf9-49a2-a122-7a6d7eacb396" (UID: "ca3af582-adf9-49a2-a122-7a6d7eacb396"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:10:28 crc kubenswrapper[4955]: I0217 13:10:28.742091 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca3af582-adf9-49a2-a122-7a6d7eacb396-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "ca3af582-adf9-49a2-a122-7a6d7eacb396" (UID: "ca3af582-adf9-49a2-a122-7a6d7eacb396"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:10:28 crc kubenswrapper[4955]: I0217 13:10:28.742114 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca3af582-adf9-49a2-a122-7a6d7eacb396-config" (OuterVolumeSpecName: "config") pod "ca3af582-adf9-49a2-a122-7a6d7eacb396" (UID: "ca3af582-adf9-49a2-a122-7a6d7eacb396"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:10:28 crc kubenswrapper[4955]: I0217 13:10:28.747615 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca3af582-adf9-49a2-a122-7a6d7eacb396-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ca3af582-adf9-49a2-a122-7a6d7eacb396" (UID: "ca3af582-adf9-49a2-a122-7a6d7eacb396"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:10:28 crc kubenswrapper[4955]: I0217 13:10:28.747700 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca3af582-adf9-49a2-a122-7a6d7eacb396-kube-api-access-5hzz4" (OuterVolumeSpecName: "kube-api-access-5hzz4") pod "ca3af582-adf9-49a2-a122-7a6d7eacb396" (UID: "ca3af582-adf9-49a2-a122-7a6d7eacb396"). InnerVolumeSpecName "kube-api-access-5hzz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:10:28 crc kubenswrapper[4955]: I0217 13:10:28.843425 4955 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca3af582-adf9-49a2-a122-7a6d7eacb396-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:10:28 crc kubenswrapper[4955]: I0217 13:10:28.843466 4955 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ca3af582-adf9-49a2-a122-7a6d7eacb396-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 13:10:28 crc kubenswrapper[4955]: I0217 13:10:28.843480 4955 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ca3af582-adf9-49a2-a122-7a6d7eacb396-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 17 13:10:28 crc kubenswrapper[4955]: I0217 13:10:28.843496 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hzz4\" (UniqueName: \"kubernetes.io/projected/ca3af582-adf9-49a2-a122-7a6d7eacb396-kube-api-access-5hzz4\") on node \"crc\" DevicePath \"\"" Feb 17 13:10:28 crc kubenswrapper[4955]: I0217 13:10:28.843510 4955 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ca3af582-adf9-49a2-a122-7a6d7eacb396-client-ca\") on node \"crc\" DevicePath \"\"" Feb 17 13:10:29 crc kubenswrapper[4955]: I0217 13:10:29.526334 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-74577df4c5-gw5v6" event={"ID":"ca3af582-adf9-49a2-a122-7a6d7eacb396","Type":"ContainerDied","Data":"27b18e30f820ddda02f47f33366b3efac74ecbd1a1a89260a058c988c2a6187b"} Feb 17 13:10:29 crc kubenswrapper[4955]: I0217 13:10:29.527981 4955 scope.go:117] "RemoveContainer" containerID="c27ed893a2f1c6d4b4025051d9c8f7f0b2a60171322bfe21e6a480f70b73b436" Feb 17 13:10:29 crc kubenswrapper[4955]: I0217 13:10:29.526620 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74577df4c5-gw5v6" Feb 17 13:10:29 crc kubenswrapper[4955]: I0217 13:10:29.528669 4955 generic.go:334] "Generic (PLEG): container finished" podID="1cf150d2-8598-45d3-b193-4d27e90ca648" containerID="f8e3bda7d09678494a592ff5288b47cc1f9d86e6067d594768e1b334ff1c9b3a" exitCode=0 Feb 17 13:10:29 crc kubenswrapper[4955]: I0217 13:10:29.528725 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jmrnj" event={"ID":"1cf150d2-8598-45d3-b193-4d27e90ca648","Type":"ContainerDied","Data":"f8e3bda7d09678494a592ff5288b47cc1f9d86e6067d594768e1b334ff1c9b3a"} Feb 17 13:10:29 crc kubenswrapper[4955]: I0217 13:10:29.536417 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qz5rg" event={"ID":"a5c8c29f-7770-4ba6-aad9-12604d10981f","Type":"ContainerStarted","Data":"f2a2f30982cd1e495b6370a8de133581c5bbe8cec21887aaa6591082b5ccbcd7"} Feb 17 13:10:29 crc kubenswrapper[4955]: I0217 13:10:29.605326 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-74577df4c5-gw5v6"] Feb 17 13:10:29 crc kubenswrapper[4955]: I0217 13:10:29.609128 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-74577df4c5-gw5v6"] Feb 17 13:10:29 crc kubenswrapper[4955]: I0217 13:10:29.886183 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2jxwm"] Feb 17 13:10:29 crc kubenswrapper[4955]: E0217 13:10:29.886467 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca3af582-adf9-49a2-a122-7a6d7eacb396" containerName="controller-manager" Feb 17 13:10:29 crc kubenswrapper[4955]: I0217 13:10:29.886489 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca3af582-adf9-49a2-a122-7a6d7eacb396" containerName="controller-manager" Feb 17 13:10:29 crc kubenswrapper[4955]: I0217 13:10:29.886617 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca3af582-adf9-49a2-a122-7a6d7eacb396" containerName="controller-manager" Feb 17 13:10:29 crc kubenswrapper[4955]: I0217 13:10:29.887761 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2jxwm" Feb 17 13:10:29 crc kubenswrapper[4955]: I0217 13:10:29.892339 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 17 13:10:29 crc kubenswrapper[4955]: I0217 13:10:29.908799 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2jxwm"] Feb 17 13:10:29 crc kubenswrapper[4955]: I0217 13:10:29.957600 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59516804-ad0e-428f-bf9a-1b5561f7ec1a-catalog-content\") pod \"community-operators-2jxwm\" (UID: \"59516804-ad0e-428f-bf9a-1b5561f7ec1a\") " pod="openshift-marketplace/community-operators-2jxwm" Feb 17 13:10:29 crc kubenswrapper[4955]: I0217 13:10:29.957653 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swc46\" (UniqueName: \"kubernetes.io/projected/59516804-ad0e-428f-bf9a-1b5561f7ec1a-kube-api-access-swc46\") pod \"community-operators-2jxwm\" (UID: \"59516804-ad0e-428f-bf9a-1b5561f7ec1a\") " pod="openshift-marketplace/community-operators-2jxwm" Feb 17 13:10:29 crc kubenswrapper[4955]: I0217 13:10:29.957672 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59516804-ad0e-428f-bf9a-1b5561f7ec1a-utilities\") pod \"community-operators-2jxwm\" (UID: \"59516804-ad0e-428f-bf9a-1b5561f7ec1a\") " pod="openshift-marketplace/community-operators-2jxwm" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.001244 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-78df95dfd-2hmx7"] Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.001844 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-78df95dfd-2hmx7" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.004147 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.004375 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.004458 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.004837 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.004992 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.005544 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.012482 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.017578 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-78df95dfd-2hmx7"] Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.058739 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkftj\" (UniqueName: \"kubernetes.io/projected/55a6a9d9-d927-47d5-b44d-34fa932fa5ca-kube-api-access-lkftj\") pod \"controller-manager-78df95dfd-2hmx7\" (UID: \"55a6a9d9-d927-47d5-b44d-34fa932fa5ca\") " pod="openshift-controller-manager/controller-manager-78df95dfd-2hmx7" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.058803 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swc46\" (UniqueName: \"kubernetes.io/projected/59516804-ad0e-428f-bf9a-1b5561f7ec1a-kube-api-access-swc46\") pod \"community-operators-2jxwm\" (UID: \"59516804-ad0e-428f-bf9a-1b5561f7ec1a\") " pod="openshift-marketplace/community-operators-2jxwm" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.058873 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59516804-ad0e-428f-bf9a-1b5561f7ec1a-utilities\") pod \"community-operators-2jxwm\" (UID: \"59516804-ad0e-428f-bf9a-1b5561f7ec1a\") " pod="openshift-marketplace/community-operators-2jxwm" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.058921 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/55a6a9d9-d927-47d5-b44d-34fa932fa5ca-client-ca\") pod \"controller-manager-78df95dfd-2hmx7\" (UID: \"55a6a9d9-d927-47d5-b44d-34fa932fa5ca\") " pod="openshift-controller-manager/controller-manager-78df95dfd-2hmx7" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.059035 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/55a6a9d9-d927-47d5-b44d-34fa932fa5ca-proxy-ca-bundles\") pod \"controller-manager-78df95dfd-2hmx7\" (UID: \"55a6a9d9-d927-47d5-b44d-34fa932fa5ca\") " pod="openshift-controller-manager/controller-manager-78df95dfd-2hmx7" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.059232 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55a6a9d9-d927-47d5-b44d-34fa932fa5ca-config\") pod \"controller-manager-78df95dfd-2hmx7\" (UID: \"55a6a9d9-d927-47d5-b44d-34fa932fa5ca\") " pod="openshift-controller-manager/controller-manager-78df95dfd-2hmx7" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.059272 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55a6a9d9-d927-47d5-b44d-34fa932fa5ca-serving-cert\") pod \"controller-manager-78df95dfd-2hmx7\" (UID: \"55a6a9d9-d927-47d5-b44d-34fa932fa5ca\") " pod="openshift-controller-manager/controller-manager-78df95dfd-2hmx7" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.059297 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59516804-ad0e-428f-bf9a-1b5561f7ec1a-catalog-content\") pod \"community-operators-2jxwm\" (UID: \"59516804-ad0e-428f-bf9a-1b5561f7ec1a\") " pod="openshift-marketplace/community-operators-2jxwm" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.059308 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59516804-ad0e-428f-bf9a-1b5561f7ec1a-utilities\") pod \"community-operators-2jxwm\" (UID: \"59516804-ad0e-428f-bf9a-1b5561f7ec1a\") " pod="openshift-marketplace/community-operators-2jxwm" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.059611 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59516804-ad0e-428f-bf9a-1b5561f7ec1a-catalog-content\") pod \"community-operators-2jxwm\" (UID: \"59516804-ad0e-428f-bf9a-1b5561f7ec1a\") " pod="openshift-marketplace/community-operators-2jxwm" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.077662 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swc46\" (UniqueName: \"kubernetes.io/projected/59516804-ad0e-428f-bf9a-1b5561f7ec1a-kube-api-access-swc46\") pod \"community-operators-2jxwm\" (UID: \"59516804-ad0e-428f-bf9a-1b5561f7ec1a\") " pod="openshift-marketplace/community-operators-2jxwm" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.160907 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55a6a9d9-d927-47d5-b44d-34fa932fa5ca-config\") pod \"controller-manager-78df95dfd-2hmx7\" (UID: \"55a6a9d9-d927-47d5-b44d-34fa932fa5ca\") " pod="openshift-controller-manager/controller-manager-78df95dfd-2hmx7" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.161008 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55a6a9d9-d927-47d5-b44d-34fa932fa5ca-serving-cert\") pod \"controller-manager-78df95dfd-2hmx7\" (UID: \"55a6a9d9-d927-47d5-b44d-34fa932fa5ca\") " pod="openshift-controller-manager/controller-manager-78df95dfd-2hmx7" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.161037 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkftj\" (UniqueName: \"kubernetes.io/projected/55a6a9d9-d927-47d5-b44d-34fa932fa5ca-kube-api-access-lkftj\") pod \"controller-manager-78df95dfd-2hmx7\" (UID: \"55a6a9d9-d927-47d5-b44d-34fa932fa5ca\") " pod="openshift-controller-manager/controller-manager-78df95dfd-2hmx7" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.161060 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/55a6a9d9-d927-47d5-b44d-34fa932fa5ca-client-ca\") pod \"controller-manager-78df95dfd-2hmx7\" (UID: \"55a6a9d9-d927-47d5-b44d-34fa932fa5ca\") " pod="openshift-controller-manager/controller-manager-78df95dfd-2hmx7" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.161092 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/55a6a9d9-d927-47d5-b44d-34fa932fa5ca-proxy-ca-bundles\") pod \"controller-manager-78df95dfd-2hmx7\" (UID: \"55a6a9d9-d927-47d5-b44d-34fa932fa5ca\") " pod="openshift-controller-manager/controller-manager-78df95dfd-2hmx7" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.162676 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55a6a9d9-d927-47d5-b44d-34fa932fa5ca-config\") pod \"controller-manager-78df95dfd-2hmx7\" (UID: \"55a6a9d9-d927-47d5-b44d-34fa932fa5ca\") " pod="openshift-controller-manager/controller-manager-78df95dfd-2hmx7" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.163089 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/55a6a9d9-d927-47d5-b44d-34fa932fa5ca-client-ca\") pod \"controller-manager-78df95dfd-2hmx7\" (UID: \"55a6a9d9-d927-47d5-b44d-34fa932fa5ca\") " pod="openshift-controller-manager/controller-manager-78df95dfd-2hmx7" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.163362 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/55a6a9d9-d927-47d5-b44d-34fa932fa5ca-proxy-ca-bundles\") pod \"controller-manager-78df95dfd-2hmx7\" (UID: \"55a6a9d9-d927-47d5-b44d-34fa932fa5ca\") " pod="openshift-controller-manager/controller-manager-78df95dfd-2hmx7" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.165954 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55a6a9d9-d927-47d5-b44d-34fa932fa5ca-serving-cert\") pod \"controller-manager-78df95dfd-2hmx7\" (UID: \"55a6a9d9-d927-47d5-b44d-34fa932fa5ca\") " pod="openshift-controller-manager/controller-manager-78df95dfd-2hmx7" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.177379 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkftj\" (UniqueName: \"kubernetes.io/projected/55a6a9d9-d927-47d5-b44d-34fa932fa5ca-kube-api-access-lkftj\") pod \"controller-manager-78df95dfd-2hmx7\" (UID: \"55a6a9d9-d927-47d5-b44d-34fa932fa5ca\") " pod="openshift-controller-manager/controller-manager-78df95dfd-2hmx7" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.206417 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2jxwm" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.230001 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca3af582-adf9-49a2-a122-7a6d7eacb396" path="/var/lib/kubelet/pods/ca3af582-adf9-49a2-a122-7a6d7eacb396/volumes" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.306513 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-krpv5"] Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.310245 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-krpv5" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.314488 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.315955 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-krpv5"] Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.360903 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-78df95dfd-2hmx7" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.414858 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2jxwm"] Feb 17 13:10:30 crc kubenswrapper[4955]: W0217 13:10:30.423737 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59516804_ad0e_428f_bf9a_1b5561f7ec1a.slice/crio-3c9d6e0b8c4fd1863551eaf8cb418e6c673ca69fb2136b7ea50cf8edf499850d WatchSource:0}: Error finding container 3c9d6e0b8c4fd1863551eaf8cb418e6c673ca69fb2136b7ea50cf8edf499850d: Status 404 returned error can't find the container with id 3c9d6e0b8c4fd1863551eaf8cb418e6c673ca69fb2136b7ea50cf8edf499850d Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.466284 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47a043d1-03b2-4a46-8c24-f1d1fef7efa6-catalog-content\") pod \"certified-operators-krpv5\" (UID: \"47a043d1-03b2-4a46-8c24-f1d1fef7efa6\") " pod="openshift-marketplace/certified-operators-krpv5" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.466627 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7pdf\" (UniqueName: \"kubernetes.io/projected/47a043d1-03b2-4a46-8c24-f1d1fef7efa6-kube-api-access-m7pdf\") pod \"certified-operators-krpv5\" (UID: \"47a043d1-03b2-4a46-8c24-f1d1fef7efa6\") " pod="openshift-marketplace/certified-operators-krpv5" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.466647 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47a043d1-03b2-4a46-8c24-f1d1fef7efa6-utilities\") pod \"certified-operators-krpv5\" (UID: \"47a043d1-03b2-4a46-8c24-f1d1fef7efa6\") " pod="openshift-marketplace/certified-operators-krpv5" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.545107 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jmrnj" event={"ID":"1cf150d2-8598-45d3-b193-4d27e90ca648","Type":"ContainerStarted","Data":"0ba08dbd69c5eebcc90b191bd3ad93aa485a6cba70446896cb4aeea28081ea82"} Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.548119 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2jxwm" event={"ID":"59516804-ad0e-428f-bf9a-1b5561f7ec1a","Type":"ContainerStarted","Data":"3c9d6e0b8c4fd1863551eaf8cb418e6c673ca69fb2136b7ea50cf8edf499850d"} Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.550183 4955 generic.go:334] "Generic (PLEG): container finished" podID="a5c8c29f-7770-4ba6-aad9-12604d10981f" containerID="f2a2f30982cd1e495b6370a8de133581c5bbe8cec21887aaa6591082b5ccbcd7" exitCode=0 Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.550229 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qz5rg" event={"ID":"a5c8c29f-7770-4ba6-aad9-12604d10981f","Type":"ContainerDied","Data":"f2a2f30982cd1e495b6370a8de133581c5bbe8cec21887aaa6591082b5ccbcd7"} Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.567930 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7pdf\" (UniqueName: \"kubernetes.io/projected/47a043d1-03b2-4a46-8c24-f1d1fef7efa6-kube-api-access-m7pdf\") pod \"certified-operators-krpv5\" (UID: \"47a043d1-03b2-4a46-8c24-f1d1fef7efa6\") " pod="openshift-marketplace/certified-operators-krpv5" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.568006 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47a043d1-03b2-4a46-8c24-f1d1fef7efa6-utilities\") pod \"certified-operators-krpv5\" (UID: \"47a043d1-03b2-4a46-8c24-f1d1fef7efa6\") " pod="openshift-marketplace/certified-operators-krpv5" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.568138 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47a043d1-03b2-4a46-8c24-f1d1fef7efa6-catalog-content\") pod \"certified-operators-krpv5\" (UID: \"47a043d1-03b2-4a46-8c24-f1d1fef7efa6\") " pod="openshift-marketplace/certified-operators-krpv5" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.569121 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47a043d1-03b2-4a46-8c24-f1d1fef7efa6-utilities\") pod \"certified-operators-krpv5\" (UID: \"47a043d1-03b2-4a46-8c24-f1d1fef7efa6\") " pod="openshift-marketplace/certified-operators-krpv5" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.569334 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47a043d1-03b2-4a46-8c24-f1d1fef7efa6-catalog-content\") pod \"certified-operators-krpv5\" (UID: \"47a043d1-03b2-4a46-8c24-f1d1fef7efa6\") " pod="openshift-marketplace/certified-operators-krpv5" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.573948 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jmrnj" podStartSLOduration=2.152155864 podStartE2EDuration="3.573928076s" podCreationTimestamp="2026-02-17 13:10:27 +0000 UTC" firstStartedPulling="2026-02-17 13:10:28.514209093 +0000 UTC m=+367.036938636" lastFinishedPulling="2026-02-17 13:10:29.935981305 +0000 UTC m=+368.458710848" observedRunningTime="2026-02-17 13:10:30.571090462 +0000 UTC m=+369.093820005" watchObservedRunningTime="2026-02-17 13:10:30.573928076 +0000 UTC m=+369.096657619" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.600722 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7pdf\" (UniqueName: \"kubernetes.io/projected/47a043d1-03b2-4a46-8c24-f1d1fef7efa6-kube-api-access-m7pdf\") pod \"certified-operators-krpv5\" (UID: \"47a043d1-03b2-4a46-8c24-f1d1fef7efa6\") " pod="openshift-marketplace/certified-operators-krpv5" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.632737 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-krpv5" Feb 17 13:10:30 crc kubenswrapper[4955]: I0217 13:10:30.763320 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-78df95dfd-2hmx7"] Feb 17 13:10:31 crc kubenswrapper[4955]: I0217 13:10:31.026452 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-krpv5"] Feb 17 13:10:31 crc kubenswrapper[4955]: I0217 13:10:31.557365 4955 generic.go:334] "Generic (PLEG): container finished" podID="47a043d1-03b2-4a46-8c24-f1d1fef7efa6" containerID="cf3a277e3c0994af9787ca8f060a51abc4367a66f2b254e2f7172effb6e7fc06" exitCode=0 Feb 17 13:10:31 crc kubenswrapper[4955]: I0217 13:10:31.557677 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-krpv5" event={"ID":"47a043d1-03b2-4a46-8c24-f1d1fef7efa6","Type":"ContainerDied","Data":"cf3a277e3c0994af9787ca8f060a51abc4367a66f2b254e2f7172effb6e7fc06"} Feb 17 13:10:31 crc kubenswrapper[4955]: I0217 13:10:31.557706 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-krpv5" event={"ID":"47a043d1-03b2-4a46-8c24-f1d1fef7efa6","Type":"ContainerStarted","Data":"75f1fb4e620835c0be8a03901a5cbaf6f69534e1a1a31201fb5302b6eb846e89"} Feb 17 13:10:31 crc kubenswrapper[4955]: I0217 13:10:31.562225 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-78df95dfd-2hmx7" event={"ID":"55a6a9d9-d927-47d5-b44d-34fa932fa5ca","Type":"ContainerStarted","Data":"7aea9615f00ffcb62be0ef8d9da69356fe9644913f56afe6842ab4a73d412b03"} Feb 17 13:10:31 crc kubenswrapper[4955]: I0217 13:10:31.562280 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-78df95dfd-2hmx7" event={"ID":"55a6a9d9-d927-47d5-b44d-34fa932fa5ca","Type":"ContainerStarted","Data":"7d83ba0803cb0f73924a87e374ae12b08f187aa205ddb0530ba5708e6b2f1ae5"} Feb 17 13:10:31 crc kubenswrapper[4955]: I0217 13:10:31.562431 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-78df95dfd-2hmx7" Feb 17 13:10:31 crc kubenswrapper[4955]: I0217 13:10:31.563504 4955 generic.go:334] "Generic (PLEG): container finished" podID="59516804-ad0e-428f-bf9a-1b5561f7ec1a" containerID="d14d087b71b7c9c06bd378e39936ce2652176dc28df3c6f40fc446e8a6252587" exitCode=0 Feb 17 13:10:31 crc kubenswrapper[4955]: I0217 13:10:31.563547 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2jxwm" event={"ID":"59516804-ad0e-428f-bf9a-1b5561f7ec1a","Type":"ContainerDied","Data":"d14d087b71b7c9c06bd378e39936ce2652176dc28df3c6f40fc446e8a6252587"} Feb 17 13:10:31 crc kubenswrapper[4955]: I0217 13:10:31.568137 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qz5rg" event={"ID":"a5c8c29f-7770-4ba6-aad9-12604d10981f","Type":"ContainerStarted","Data":"76707dd52c55db25e9b11994cac136fda1893ad3a3da6a7cdba30e98053edebb"} Feb 17 13:10:31 crc kubenswrapper[4955]: I0217 13:10:31.568388 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-78df95dfd-2hmx7" Feb 17 13:10:31 crc kubenswrapper[4955]: I0217 13:10:31.635538 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qz5rg" podStartSLOduration=2.182418819 podStartE2EDuration="4.635519221s" podCreationTimestamp="2026-02-17 13:10:27 +0000 UTC" firstStartedPulling="2026-02-17 13:10:28.520570631 +0000 UTC m=+367.043300174" lastFinishedPulling="2026-02-17 13:10:30.973671033 +0000 UTC m=+369.496400576" observedRunningTime="2026-02-17 13:10:31.632529492 +0000 UTC m=+370.155259045" watchObservedRunningTime="2026-02-17 13:10:31.635519221 +0000 UTC m=+370.158248774" Feb 17 13:10:31 crc kubenswrapper[4955]: I0217 13:10:31.653300 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-78df95dfd-2hmx7" podStartSLOduration=3.653281815 podStartE2EDuration="3.653281815s" podCreationTimestamp="2026-02-17 13:10:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:10:31.647253537 +0000 UTC m=+370.169983090" watchObservedRunningTime="2026-02-17 13:10:31.653281815 +0000 UTC m=+370.176011368" Feb 17 13:10:32 crc kubenswrapper[4955]: I0217 13:10:32.573026 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2jxwm" event={"ID":"59516804-ad0e-428f-bf9a-1b5561f7ec1a","Type":"ContainerStarted","Data":"8650bee83dfa35323bcc4594a18fd0833ab6784a9db008c54d53bd7bb82e4e95"} Feb 17 13:10:33 crc kubenswrapper[4955]: I0217 13:10:33.579591 4955 generic.go:334] "Generic (PLEG): container finished" podID="59516804-ad0e-428f-bf9a-1b5561f7ec1a" containerID="8650bee83dfa35323bcc4594a18fd0833ab6784a9db008c54d53bd7bb82e4e95" exitCode=0 Feb 17 13:10:33 crc kubenswrapper[4955]: I0217 13:10:33.579643 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2jxwm" event={"ID":"59516804-ad0e-428f-bf9a-1b5561f7ec1a","Type":"ContainerDied","Data":"8650bee83dfa35323bcc4594a18fd0833ab6784a9db008c54d53bd7bb82e4e95"} Feb 17 13:10:33 crc kubenswrapper[4955]: I0217 13:10:33.582441 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-krpv5" event={"ID":"47a043d1-03b2-4a46-8c24-f1d1fef7efa6","Type":"ContainerDied","Data":"2ea9d2b6c542edbcff056197f6fe2539c2f8fa4bb64ca3616110a492cebaf5c3"} Feb 17 13:10:33 crc kubenswrapper[4955]: I0217 13:10:33.582526 4955 generic.go:334] "Generic (PLEG): container finished" podID="47a043d1-03b2-4a46-8c24-f1d1fef7efa6" containerID="2ea9d2b6c542edbcff056197f6fe2539c2f8fa4bb64ca3616110a492cebaf5c3" exitCode=0 Feb 17 13:10:34 crc kubenswrapper[4955]: I0217 13:10:34.575313 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 13:10:34 crc kubenswrapper[4955]: I0217 13:10:34.575657 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 13:10:34 crc kubenswrapper[4955]: I0217 13:10:34.589956 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-krpv5" event={"ID":"47a043d1-03b2-4a46-8c24-f1d1fef7efa6","Type":"ContainerStarted","Data":"fac3ce40086ce37bb22ae409af4101b133b4a188fe0ae5f42fe892fe0cd7b21d"} Feb 17 13:10:34 crc kubenswrapper[4955]: I0217 13:10:34.592316 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2jxwm" event={"ID":"59516804-ad0e-428f-bf9a-1b5561f7ec1a","Type":"ContainerStarted","Data":"8c321f610acdecf6425a7589aa0c53c712bafbc86f4e93d9ea22a5460fc0065b"} Feb 17 13:10:34 crc kubenswrapper[4955]: I0217 13:10:34.613718 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-krpv5" podStartSLOduration=2.05026705 podStartE2EDuration="4.613702781s" podCreationTimestamp="2026-02-17 13:10:30 +0000 UTC" firstStartedPulling="2026-02-17 13:10:31.559012541 +0000 UTC m=+370.081742084" lastFinishedPulling="2026-02-17 13:10:34.122448282 +0000 UTC m=+372.645177815" observedRunningTime="2026-02-17 13:10:34.611014962 +0000 UTC m=+373.133744495" watchObservedRunningTime="2026-02-17 13:10:34.613702781 +0000 UTC m=+373.136432324" Feb 17 13:10:34 crc kubenswrapper[4955]: I0217 13:10:34.633068 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2jxwm" podStartSLOduration=3.203984429 podStartE2EDuration="5.633048073s" podCreationTimestamp="2026-02-17 13:10:29 +0000 UTC" firstStartedPulling="2026-02-17 13:10:31.564648397 +0000 UTC m=+370.087377940" lastFinishedPulling="2026-02-17 13:10:33.993712031 +0000 UTC m=+372.516441584" observedRunningTime="2026-02-17 13:10:34.632083784 +0000 UTC m=+373.154813367" watchObservedRunningTime="2026-02-17 13:10:34.633048073 +0000 UTC m=+373.155777616" Feb 17 13:10:37 crc kubenswrapper[4955]: I0217 13:10:37.820504 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jmrnj" Feb 17 13:10:37 crc kubenswrapper[4955]: I0217 13:10:37.821030 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jmrnj" Feb 17 13:10:37 crc kubenswrapper[4955]: I0217 13:10:37.868674 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jmrnj" Feb 17 13:10:38 crc kubenswrapper[4955]: I0217 13:10:38.021034 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qz5rg" Feb 17 13:10:38 crc kubenswrapper[4955]: I0217 13:10:38.021178 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qz5rg" Feb 17 13:10:38 crc kubenswrapper[4955]: I0217 13:10:38.092124 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qz5rg" Feb 17 13:10:38 crc kubenswrapper[4955]: I0217 13:10:38.653004 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jmrnj" Feb 17 13:10:38 crc kubenswrapper[4955]: I0217 13:10:38.653400 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qz5rg" Feb 17 13:10:40 crc kubenswrapper[4955]: I0217 13:10:40.207037 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2jxwm" Feb 17 13:10:40 crc kubenswrapper[4955]: I0217 13:10:40.207130 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2jxwm" Feb 17 13:10:40 crc kubenswrapper[4955]: I0217 13:10:40.273746 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2jxwm" Feb 17 13:10:40 crc kubenswrapper[4955]: I0217 13:10:40.633231 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-krpv5" Feb 17 13:10:40 crc kubenswrapper[4955]: I0217 13:10:40.633486 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-krpv5" Feb 17 13:10:40 crc kubenswrapper[4955]: I0217 13:10:40.670581 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2jxwm" Feb 17 13:10:40 crc kubenswrapper[4955]: I0217 13:10:40.674140 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-krpv5" Feb 17 13:10:41 crc kubenswrapper[4955]: I0217 13:10:41.674544 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-krpv5" Feb 17 13:10:44 crc kubenswrapper[4955]: I0217 13:10:44.636737 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-74pjh" Feb 17 13:10:44 crc kubenswrapper[4955]: I0217 13:10:44.713822 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8mkhz"] Feb 17 13:11:04 crc kubenswrapper[4955]: I0217 13:11:04.575140 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 13:11:04 crc kubenswrapper[4955]: I0217 13:11:04.576089 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 13:11:04 crc kubenswrapper[4955]: I0217 13:11:04.576177 4955 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" Feb 17 13:11:04 crc kubenswrapper[4955]: I0217 13:11:04.577274 4955 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3d5433cfcc5f7c8d75620ee096f13026d74f5b8c91209da3a81b1363052f9063"} pod="openshift-machine-config-operator/machine-config-daemon-29qxq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 17 13:11:04 crc kubenswrapper[4955]: I0217 13:11:04.577381 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" containerID="cri-o://3d5433cfcc5f7c8d75620ee096f13026d74f5b8c91209da3a81b1363052f9063" gracePeriod=600 Feb 17 13:11:05 crc kubenswrapper[4955]: I0217 13:11:05.161696 4955 generic.go:334] "Generic (PLEG): container finished" podID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerID="3d5433cfcc5f7c8d75620ee096f13026d74f5b8c91209da3a81b1363052f9063" exitCode=0 Feb 17 13:11:05 crc kubenswrapper[4955]: I0217 13:11:05.161751 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" event={"ID":"5bf36887-2a2b-4193-abdb-a15beaf46300","Type":"ContainerDied","Data":"3d5433cfcc5f7c8d75620ee096f13026d74f5b8c91209da3a81b1363052f9063"} Feb 17 13:11:05 crc kubenswrapper[4955]: I0217 13:11:05.162117 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" event={"ID":"5bf36887-2a2b-4193-abdb-a15beaf46300","Type":"ContainerStarted","Data":"ae1cb692a0145e1501a14dd21bdbee429e632fe371cafa0c41cac4fa359ae05a"} Feb 17 13:11:05 crc kubenswrapper[4955]: I0217 13:11:05.162149 4955 scope.go:117] "RemoveContainer" containerID="9bec78f9e9f6416d6b600f6231f0d24652473b48b95bf0e0675cc85c9208d124" Feb 17 13:11:09 crc kubenswrapper[4955]: I0217 13:11:09.751936 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" podUID="e3687198-a53d-4a80-baec-bafd1c5ef14d" containerName="registry" containerID="cri-o://0cb73d5edcc4e737199a46ba7e9ae93d7ebbe1acbe6100bcafb4d31958bd04f4" gracePeriod=30 Feb 17 13:11:10 crc kubenswrapper[4955]: I0217 13:11:10.206329 4955 generic.go:334] "Generic (PLEG): container finished" podID="e3687198-a53d-4a80-baec-bafd1c5ef14d" containerID="0cb73d5edcc4e737199a46ba7e9ae93d7ebbe1acbe6100bcafb4d31958bd04f4" exitCode=0 Feb 17 13:11:10 crc kubenswrapper[4955]: I0217 13:11:10.206467 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" event={"ID":"e3687198-a53d-4a80-baec-bafd1c5ef14d","Type":"ContainerDied","Data":"0cb73d5edcc4e737199a46ba7e9ae93d7ebbe1acbe6100bcafb4d31958bd04f4"} Feb 17 13:11:10 crc kubenswrapper[4955]: I0217 13:11:10.314758 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:11:10 crc kubenswrapper[4955]: I0217 13:11:10.391935 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"e3687198-a53d-4a80-baec-bafd1c5ef14d\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " Feb 17 13:11:10 crc kubenswrapper[4955]: I0217 13:11:10.392002 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e3687198-a53d-4a80-baec-bafd1c5ef14d-installation-pull-secrets\") pod \"e3687198-a53d-4a80-baec-bafd1c5ef14d\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " Feb 17 13:11:10 crc kubenswrapper[4955]: I0217 13:11:10.392066 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e3687198-a53d-4a80-baec-bafd1c5ef14d-registry-certificates\") pod \"e3687198-a53d-4a80-baec-bafd1c5ef14d\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " Feb 17 13:11:10 crc kubenswrapper[4955]: I0217 13:11:10.392111 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e3687198-a53d-4a80-baec-bafd1c5ef14d-registry-tls\") pod \"e3687198-a53d-4a80-baec-bafd1c5ef14d\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " Feb 17 13:11:10 crc kubenswrapper[4955]: I0217 13:11:10.392171 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e3687198-a53d-4a80-baec-bafd1c5ef14d-bound-sa-token\") pod \"e3687198-a53d-4a80-baec-bafd1c5ef14d\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " Feb 17 13:11:10 crc kubenswrapper[4955]: I0217 13:11:10.392203 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-whm4d\" (UniqueName: \"kubernetes.io/projected/e3687198-a53d-4a80-baec-bafd1c5ef14d-kube-api-access-whm4d\") pod \"e3687198-a53d-4a80-baec-bafd1c5ef14d\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " Feb 17 13:11:10 crc kubenswrapper[4955]: I0217 13:11:10.392261 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e3687198-a53d-4a80-baec-bafd1c5ef14d-trusted-ca\") pod \"e3687198-a53d-4a80-baec-bafd1c5ef14d\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " Feb 17 13:11:10 crc kubenswrapper[4955]: I0217 13:11:10.392299 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e3687198-a53d-4a80-baec-bafd1c5ef14d-ca-trust-extracted\") pod \"e3687198-a53d-4a80-baec-bafd1c5ef14d\" (UID: \"e3687198-a53d-4a80-baec-bafd1c5ef14d\") " Feb 17 13:11:10 crc kubenswrapper[4955]: I0217 13:11:10.393150 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3687198-a53d-4a80-baec-bafd1c5ef14d-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "e3687198-a53d-4a80-baec-bafd1c5ef14d" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:11:10 crc kubenswrapper[4955]: I0217 13:11:10.394326 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3687198-a53d-4a80-baec-bafd1c5ef14d-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "e3687198-a53d-4a80-baec-bafd1c5ef14d" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:11:10 crc kubenswrapper[4955]: I0217 13:11:10.399038 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3687198-a53d-4a80-baec-bafd1c5ef14d-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "e3687198-a53d-4a80-baec-bafd1c5ef14d" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:11:10 crc kubenswrapper[4955]: I0217 13:11:10.400097 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3687198-a53d-4a80-baec-bafd1c5ef14d-kube-api-access-whm4d" (OuterVolumeSpecName: "kube-api-access-whm4d") pod "e3687198-a53d-4a80-baec-bafd1c5ef14d" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d"). InnerVolumeSpecName "kube-api-access-whm4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:11:10 crc kubenswrapper[4955]: I0217 13:11:10.400260 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3687198-a53d-4a80-baec-bafd1c5ef14d-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "e3687198-a53d-4a80-baec-bafd1c5ef14d" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:11:10 crc kubenswrapper[4955]: I0217 13:11:10.400455 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3687198-a53d-4a80-baec-bafd1c5ef14d-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "e3687198-a53d-4a80-baec-bafd1c5ef14d" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:11:10 crc kubenswrapper[4955]: I0217 13:11:10.408391 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3687198-a53d-4a80-baec-bafd1c5ef14d-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "e3687198-a53d-4a80-baec-bafd1c5ef14d" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:11:10 crc kubenswrapper[4955]: I0217 13:11:10.413222 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "e3687198-a53d-4a80-baec-bafd1c5ef14d" (UID: "e3687198-a53d-4a80-baec-bafd1c5ef14d"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 17 13:11:10 crc kubenswrapper[4955]: I0217 13:11:10.493744 4955 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e3687198-a53d-4a80-baec-bafd1c5ef14d-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 17 13:11:10 crc kubenswrapper[4955]: I0217 13:11:10.493881 4955 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e3687198-a53d-4a80-baec-bafd1c5ef14d-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 17 13:11:10 crc kubenswrapper[4955]: I0217 13:11:10.493906 4955 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e3687198-a53d-4a80-baec-bafd1c5ef14d-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 17 13:11:10 crc kubenswrapper[4955]: I0217 13:11:10.493926 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-whm4d\" (UniqueName: \"kubernetes.io/projected/e3687198-a53d-4a80-baec-bafd1c5ef14d-kube-api-access-whm4d\") on node \"crc\" DevicePath \"\"" Feb 17 13:11:10 crc kubenswrapper[4955]: I0217 13:11:10.493945 4955 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e3687198-a53d-4a80-baec-bafd1c5ef14d-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 17 13:11:10 crc kubenswrapper[4955]: I0217 13:11:10.493963 4955 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e3687198-a53d-4a80-baec-bafd1c5ef14d-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 17 13:11:10 crc kubenswrapper[4955]: I0217 13:11:10.493983 4955 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e3687198-a53d-4a80-baec-bafd1c5ef14d-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 17 13:11:11 crc kubenswrapper[4955]: I0217 13:11:11.216709 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" event={"ID":"e3687198-a53d-4a80-baec-bafd1c5ef14d","Type":"ContainerDied","Data":"c072344e5397278b7930e09c1b5a562c0a54495e3ac858fd306420c45604017c"} Feb 17 13:11:11 crc kubenswrapper[4955]: I0217 13:11:11.216761 4955 scope.go:117] "RemoveContainer" containerID="0cb73d5edcc4e737199a46ba7e9ae93d7ebbe1acbe6100bcafb4d31958bd04f4" Feb 17 13:11:11 crc kubenswrapper[4955]: I0217 13:11:11.216845 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8mkhz" Feb 17 13:11:11 crc kubenswrapper[4955]: I0217 13:11:11.264607 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8mkhz"] Feb 17 13:11:11 crc kubenswrapper[4955]: I0217 13:11:11.271248 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8mkhz"] Feb 17 13:11:12 crc kubenswrapper[4955]: I0217 13:11:12.229541 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3687198-a53d-4a80-baec-bafd1c5ef14d" path="/var/lib/kubelet/pods/e3687198-a53d-4a80-baec-bafd1c5ef14d/volumes" Feb 17 13:13:04 crc kubenswrapper[4955]: I0217 13:13:04.574470 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 13:13:04 crc kubenswrapper[4955]: I0217 13:13:04.575063 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 13:13:34 crc kubenswrapper[4955]: I0217 13:13:34.575167 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 13:13:34 crc kubenswrapper[4955]: I0217 13:13:34.575863 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 13:14:04 crc kubenswrapper[4955]: I0217 13:14:04.575354 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 13:14:04 crc kubenswrapper[4955]: I0217 13:14:04.576214 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 13:14:04 crc kubenswrapper[4955]: I0217 13:14:04.576283 4955 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" Feb 17 13:14:04 crc kubenswrapper[4955]: I0217 13:14:04.577119 4955 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ae1cb692a0145e1501a14dd21bdbee429e632fe371cafa0c41cac4fa359ae05a"} pod="openshift-machine-config-operator/machine-config-daemon-29qxq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 17 13:14:04 crc kubenswrapper[4955]: I0217 13:14:04.577213 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" containerID="cri-o://ae1cb692a0145e1501a14dd21bdbee429e632fe371cafa0c41cac4fa359ae05a" gracePeriod=600 Feb 17 13:14:05 crc kubenswrapper[4955]: I0217 13:14:05.315802 4955 generic.go:334] "Generic (PLEG): container finished" podID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerID="ae1cb692a0145e1501a14dd21bdbee429e632fe371cafa0c41cac4fa359ae05a" exitCode=0 Feb 17 13:14:05 crc kubenswrapper[4955]: I0217 13:14:05.315819 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" event={"ID":"5bf36887-2a2b-4193-abdb-a15beaf46300","Type":"ContainerDied","Data":"ae1cb692a0145e1501a14dd21bdbee429e632fe371cafa0c41cac4fa359ae05a"} Feb 17 13:14:05 crc kubenswrapper[4955]: I0217 13:14:05.316268 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" event={"ID":"5bf36887-2a2b-4193-abdb-a15beaf46300","Type":"ContainerStarted","Data":"817a479207945cb4a0cb4caff307634a33fb718eab5f99bb1c4a37862d5010ac"} Feb 17 13:14:05 crc kubenswrapper[4955]: I0217 13:14:05.316295 4955 scope.go:117] "RemoveContainer" containerID="3d5433cfcc5f7c8d75620ee096f13026d74f5b8c91209da3a81b1363052f9063" Feb 17 13:15:00 crc kubenswrapper[4955]: I0217 13:15:00.195575 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29522235-qssjp"] Feb 17 13:15:00 crc kubenswrapper[4955]: E0217 13:15:00.196605 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3687198-a53d-4a80-baec-bafd1c5ef14d" containerName="registry" Feb 17 13:15:00 crc kubenswrapper[4955]: I0217 13:15:00.196631 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3687198-a53d-4a80-baec-bafd1c5ef14d" containerName="registry" Feb 17 13:15:00 crc kubenswrapper[4955]: I0217 13:15:00.196884 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3687198-a53d-4a80-baec-bafd1c5ef14d" containerName="registry" Feb 17 13:15:00 crc kubenswrapper[4955]: I0217 13:15:00.197581 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29522235-qssjp" Feb 17 13:15:00 crc kubenswrapper[4955]: I0217 13:15:00.199887 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 17 13:15:00 crc kubenswrapper[4955]: I0217 13:15:00.199947 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 17 13:15:00 crc kubenswrapper[4955]: I0217 13:15:00.208942 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29522235-qssjp"] Feb 17 13:15:00 crc kubenswrapper[4955]: I0217 13:15:00.330747 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9h7np\" (UniqueName: \"kubernetes.io/projected/45631608-0d2f-4144-a96b-1a921c8f014b-kube-api-access-9h7np\") pod \"collect-profiles-29522235-qssjp\" (UID: \"45631608-0d2f-4144-a96b-1a921c8f014b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522235-qssjp" Feb 17 13:15:00 crc kubenswrapper[4955]: I0217 13:15:00.330842 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/45631608-0d2f-4144-a96b-1a921c8f014b-secret-volume\") pod \"collect-profiles-29522235-qssjp\" (UID: \"45631608-0d2f-4144-a96b-1a921c8f014b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522235-qssjp" Feb 17 13:15:00 crc kubenswrapper[4955]: I0217 13:15:00.331118 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/45631608-0d2f-4144-a96b-1a921c8f014b-config-volume\") pod \"collect-profiles-29522235-qssjp\" (UID: \"45631608-0d2f-4144-a96b-1a921c8f014b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522235-qssjp" Feb 17 13:15:00 crc kubenswrapper[4955]: I0217 13:15:00.432082 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9h7np\" (UniqueName: \"kubernetes.io/projected/45631608-0d2f-4144-a96b-1a921c8f014b-kube-api-access-9h7np\") pod \"collect-profiles-29522235-qssjp\" (UID: \"45631608-0d2f-4144-a96b-1a921c8f014b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522235-qssjp" Feb 17 13:15:00 crc kubenswrapper[4955]: I0217 13:15:00.432138 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/45631608-0d2f-4144-a96b-1a921c8f014b-secret-volume\") pod \"collect-profiles-29522235-qssjp\" (UID: \"45631608-0d2f-4144-a96b-1a921c8f014b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522235-qssjp" Feb 17 13:15:00 crc kubenswrapper[4955]: I0217 13:15:00.432200 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/45631608-0d2f-4144-a96b-1a921c8f014b-config-volume\") pod \"collect-profiles-29522235-qssjp\" (UID: \"45631608-0d2f-4144-a96b-1a921c8f014b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522235-qssjp" Feb 17 13:15:00 crc kubenswrapper[4955]: I0217 13:15:00.433589 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/45631608-0d2f-4144-a96b-1a921c8f014b-config-volume\") pod \"collect-profiles-29522235-qssjp\" (UID: \"45631608-0d2f-4144-a96b-1a921c8f014b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522235-qssjp" Feb 17 13:15:00 crc kubenswrapper[4955]: I0217 13:15:00.440372 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/45631608-0d2f-4144-a96b-1a921c8f014b-secret-volume\") pod \"collect-profiles-29522235-qssjp\" (UID: \"45631608-0d2f-4144-a96b-1a921c8f014b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522235-qssjp" Feb 17 13:15:00 crc kubenswrapper[4955]: I0217 13:15:00.454128 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9h7np\" (UniqueName: \"kubernetes.io/projected/45631608-0d2f-4144-a96b-1a921c8f014b-kube-api-access-9h7np\") pod \"collect-profiles-29522235-qssjp\" (UID: \"45631608-0d2f-4144-a96b-1a921c8f014b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522235-qssjp" Feb 17 13:15:00 crc kubenswrapper[4955]: I0217 13:15:00.517432 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29522235-qssjp" Feb 17 13:15:00 crc kubenswrapper[4955]: I0217 13:15:00.780448 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29522235-qssjp"] Feb 17 13:15:01 crc kubenswrapper[4955]: I0217 13:15:01.678021 4955 generic.go:334] "Generic (PLEG): container finished" podID="45631608-0d2f-4144-a96b-1a921c8f014b" containerID="ea3c883435d1bb5d78114d7bd4d40ad96e51f3a077b2012a807279fed60386c0" exitCode=0 Feb 17 13:15:01 crc kubenswrapper[4955]: I0217 13:15:01.678111 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29522235-qssjp" event={"ID":"45631608-0d2f-4144-a96b-1a921c8f014b","Type":"ContainerDied","Data":"ea3c883435d1bb5d78114d7bd4d40ad96e51f3a077b2012a807279fed60386c0"} Feb 17 13:15:01 crc kubenswrapper[4955]: I0217 13:15:01.678193 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29522235-qssjp" event={"ID":"45631608-0d2f-4144-a96b-1a921c8f014b","Type":"ContainerStarted","Data":"42e63a205def789290230d05a8b077017eab1d2ff67df69471379ad95da0b8c3"} Feb 17 13:15:02 crc kubenswrapper[4955]: I0217 13:15:02.969409 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29522235-qssjp" Feb 17 13:15:03 crc kubenswrapper[4955]: I0217 13:15:03.062300 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/45631608-0d2f-4144-a96b-1a921c8f014b-secret-volume\") pod \"45631608-0d2f-4144-a96b-1a921c8f014b\" (UID: \"45631608-0d2f-4144-a96b-1a921c8f014b\") " Feb 17 13:15:03 crc kubenswrapper[4955]: I0217 13:15:03.062400 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9h7np\" (UniqueName: \"kubernetes.io/projected/45631608-0d2f-4144-a96b-1a921c8f014b-kube-api-access-9h7np\") pod \"45631608-0d2f-4144-a96b-1a921c8f014b\" (UID: \"45631608-0d2f-4144-a96b-1a921c8f014b\") " Feb 17 13:15:03 crc kubenswrapper[4955]: I0217 13:15:03.062500 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/45631608-0d2f-4144-a96b-1a921c8f014b-config-volume\") pod \"45631608-0d2f-4144-a96b-1a921c8f014b\" (UID: \"45631608-0d2f-4144-a96b-1a921c8f014b\") " Feb 17 13:15:03 crc kubenswrapper[4955]: I0217 13:15:03.063657 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45631608-0d2f-4144-a96b-1a921c8f014b-config-volume" (OuterVolumeSpecName: "config-volume") pod "45631608-0d2f-4144-a96b-1a921c8f014b" (UID: "45631608-0d2f-4144-a96b-1a921c8f014b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:15:03 crc kubenswrapper[4955]: I0217 13:15:03.070596 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45631608-0d2f-4144-a96b-1a921c8f014b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "45631608-0d2f-4144-a96b-1a921c8f014b" (UID: "45631608-0d2f-4144-a96b-1a921c8f014b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:15:03 crc kubenswrapper[4955]: I0217 13:15:03.074896 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45631608-0d2f-4144-a96b-1a921c8f014b-kube-api-access-9h7np" (OuterVolumeSpecName: "kube-api-access-9h7np") pod "45631608-0d2f-4144-a96b-1a921c8f014b" (UID: "45631608-0d2f-4144-a96b-1a921c8f014b"). InnerVolumeSpecName "kube-api-access-9h7np". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:15:03 crc kubenswrapper[4955]: I0217 13:15:03.164199 4955 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/45631608-0d2f-4144-a96b-1a921c8f014b-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 17 13:15:03 crc kubenswrapper[4955]: I0217 13:15:03.164283 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9h7np\" (UniqueName: \"kubernetes.io/projected/45631608-0d2f-4144-a96b-1a921c8f014b-kube-api-access-9h7np\") on node \"crc\" DevicePath \"\"" Feb 17 13:15:03 crc kubenswrapper[4955]: I0217 13:15:03.164316 4955 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/45631608-0d2f-4144-a96b-1a921c8f014b-config-volume\") on node \"crc\" DevicePath \"\"" Feb 17 13:15:03 crc kubenswrapper[4955]: I0217 13:15:03.691721 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29522235-qssjp" event={"ID":"45631608-0d2f-4144-a96b-1a921c8f014b","Type":"ContainerDied","Data":"42e63a205def789290230d05a8b077017eab1d2ff67df69471379ad95da0b8c3"} Feb 17 13:15:03 crc kubenswrapper[4955]: I0217 13:15:03.691774 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42e63a205def789290230d05a8b077017eab1d2ff67df69471379ad95da0b8c3" Feb 17 13:15:03 crc kubenswrapper[4955]: I0217 13:15:03.691808 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29522235-qssjp" Feb 17 13:15:26 crc kubenswrapper[4955]: I0217 13:15:26.816600 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-lmhj4"] Feb 17 13:15:26 crc kubenswrapper[4955]: E0217 13:15:26.817172 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45631608-0d2f-4144-a96b-1a921c8f014b" containerName="collect-profiles" Feb 17 13:15:26 crc kubenswrapper[4955]: I0217 13:15:26.817184 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="45631608-0d2f-4144-a96b-1a921c8f014b" containerName="collect-profiles" Feb 17 13:15:26 crc kubenswrapper[4955]: I0217 13:15:26.817274 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="45631608-0d2f-4144-a96b-1a921c8f014b" containerName="collect-profiles" Feb 17 13:15:26 crc kubenswrapper[4955]: I0217 13:15:26.817624 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-lmhj4" Feb 17 13:15:26 crc kubenswrapper[4955]: I0217 13:15:26.821079 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Feb 17 13:15:26 crc kubenswrapper[4955]: I0217 13:15:26.821097 4955 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-7wdf9" Feb 17 13:15:26 crc kubenswrapper[4955]: I0217 13:15:26.832101 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Feb 17 13:15:26 crc kubenswrapper[4955]: I0217 13:15:26.845102 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-858654f9db-lg6zf"] Feb 17 13:15:26 crc kubenswrapper[4955]: I0217 13:15:26.845758 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-lg6zf" Feb 17 13:15:26 crc kubenswrapper[4955]: I0217 13:15:26.846852 4955 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-7pzpp" Feb 17 13:15:26 crc kubenswrapper[4955]: I0217 13:15:26.852849 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-lmhj4"] Feb 17 13:15:26 crc kubenswrapper[4955]: I0217 13:15:26.857311 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-69dts"] Feb 17 13:15:26 crc kubenswrapper[4955]: I0217 13:15:26.857898 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-69dts" Feb 17 13:15:26 crc kubenswrapper[4955]: I0217 13:15:26.859947 4955 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-n4446" Feb 17 13:15:26 crc kubenswrapper[4955]: I0217 13:15:26.862376 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-lg6zf"] Feb 17 13:15:26 crc kubenswrapper[4955]: I0217 13:15:26.865390 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-69dts"] Feb 17 13:15:27 crc kubenswrapper[4955]: I0217 13:15:27.003440 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzh7c\" (UniqueName: \"kubernetes.io/projected/d4a28595-c677-4966-8d3a-1ad5d299fd51-kube-api-access-lzh7c\") pod \"cert-manager-webhook-687f57d79b-69dts\" (UID: \"d4a28595-c677-4966-8d3a-1ad5d299fd51\") " pod="cert-manager/cert-manager-webhook-687f57d79b-69dts" Feb 17 13:15:27 crc kubenswrapper[4955]: I0217 13:15:27.003668 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwzdl\" (UniqueName: \"kubernetes.io/projected/54536ade-766d-4cf8-b1b5-057ccb961b13-kube-api-access-gwzdl\") pod \"cert-manager-858654f9db-lg6zf\" (UID: \"54536ade-766d-4cf8-b1b5-057ccb961b13\") " pod="cert-manager/cert-manager-858654f9db-lg6zf" Feb 17 13:15:27 crc kubenswrapper[4955]: I0217 13:15:27.003742 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lm268\" (UniqueName: \"kubernetes.io/projected/de5cd487-58d9-4c86-a4e2-f6b000a351f5-kube-api-access-lm268\") pod \"cert-manager-cainjector-cf98fcc89-lmhj4\" (UID: \"de5cd487-58d9-4c86-a4e2-f6b000a351f5\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-lmhj4" Feb 17 13:15:27 crc kubenswrapper[4955]: I0217 13:15:27.104515 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lm268\" (UniqueName: \"kubernetes.io/projected/de5cd487-58d9-4c86-a4e2-f6b000a351f5-kube-api-access-lm268\") pod \"cert-manager-cainjector-cf98fcc89-lmhj4\" (UID: \"de5cd487-58d9-4c86-a4e2-f6b000a351f5\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-lmhj4" Feb 17 13:15:27 crc kubenswrapper[4955]: I0217 13:15:27.104584 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzh7c\" (UniqueName: \"kubernetes.io/projected/d4a28595-c677-4966-8d3a-1ad5d299fd51-kube-api-access-lzh7c\") pod \"cert-manager-webhook-687f57d79b-69dts\" (UID: \"d4a28595-c677-4966-8d3a-1ad5d299fd51\") " pod="cert-manager/cert-manager-webhook-687f57d79b-69dts" Feb 17 13:15:27 crc kubenswrapper[4955]: I0217 13:15:27.104635 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwzdl\" (UniqueName: \"kubernetes.io/projected/54536ade-766d-4cf8-b1b5-057ccb961b13-kube-api-access-gwzdl\") pod \"cert-manager-858654f9db-lg6zf\" (UID: \"54536ade-766d-4cf8-b1b5-057ccb961b13\") " pod="cert-manager/cert-manager-858654f9db-lg6zf" Feb 17 13:15:27 crc kubenswrapper[4955]: I0217 13:15:27.122661 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzh7c\" (UniqueName: \"kubernetes.io/projected/d4a28595-c677-4966-8d3a-1ad5d299fd51-kube-api-access-lzh7c\") pod \"cert-manager-webhook-687f57d79b-69dts\" (UID: \"d4a28595-c677-4966-8d3a-1ad5d299fd51\") " pod="cert-manager/cert-manager-webhook-687f57d79b-69dts" Feb 17 13:15:27 crc kubenswrapper[4955]: I0217 13:15:27.122858 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwzdl\" (UniqueName: \"kubernetes.io/projected/54536ade-766d-4cf8-b1b5-057ccb961b13-kube-api-access-gwzdl\") pod \"cert-manager-858654f9db-lg6zf\" (UID: \"54536ade-766d-4cf8-b1b5-057ccb961b13\") " pod="cert-manager/cert-manager-858654f9db-lg6zf" Feb 17 13:15:27 crc kubenswrapper[4955]: I0217 13:15:27.123050 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lm268\" (UniqueName: \"kubernetes.io/projected/de5cd487-58d9-4c86-a4e2-f6b000a351f5-kube-api-access-lm268\") pod \"cert-manager-cainjector-cf98fcc89-lmhj4\" (UID: \"de5cd487-58d9-4c86-a4e2-f6b000a351f5\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-lmhj4" Feb 17 13:15:27 crc kubenswrapper[4955]: I0217 13:15:27.138868 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-lmhj4" Feb 17 13:15:27 crc kubenswrapper[4955]: I0217 13:15:27.170986 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-lg6zf" Feb 17 13:15:27 crc kubenswrapper[4955]: I0217 13:15:27.180051 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-69dts" Feb 17 13:15:27 crc kubenswrapper[4955]: I0217 13:15:27.357750 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-lmhj4"] Feb 17 13:15:27 crc kubenswrapper[4955]: I0217 13:15:27.380398 4955 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 17 13:15:27 crc kubenswrapper[4955]: I0217 13:15:27.632295 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-lg6zf"] Feb 17 13:15:27 crc kubenswrapper[4955]: I0217 13:15:27.635482 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-69dts"] Feb 17 13:15:27 crc kubenswrapper[4955]: W0217 13:15:27.643096 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4a28595_c677_4966_8d3a_1ad5d299fd51.slice/crio-5179006dd2cb8f432fd0f750dc0e1617070d7ba8b11b6c62659d861b0d57154f WatchSource:0}: Error finding container 5179006dd2cb8f432fd0f750dc0e1617070d7ba8b11b6c62659d861b0d57154f: Status 404 returned error can't find the container with id 5179006dd2cb8f432fd0f750dc0e1617070d7ba8b11b6c62659d861b0d57154f Feb 17 13:15:27 crc kubenswrapper[4955]: W0217 13:15:27.645051 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod54536ade_766d_4cf8_b1b5_057ccb961b13.slice/crio-697f1604df66921970a837740fc97381a9bfecb0170121372c8ea8ad0afd069f WatchSource:0}: Error finding container 697f1604df66921970a837740fc97381a9bfecb0170121372c8ea8ad0afd069f: Status 404 returned error can't find the container with id 697f1604df66921970a837740fc97381a9bfecb0170121372c8ea8ad0afd069f Feb 17 13:15:27 crc kubenswrapper[4955]: I0217 13:15:27.846015 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-lg6zf" event={"ID":"54536ade-766d-4cf8-b1b5-057ccb961b13","Type":"ContainerStarted","Data":"697f1604df66921970a837740fc97381a9bfecb0170121372c8ea8ad0afd069f"} Feb 17 13:15:27 crc kubenswrapper[4955]: I0217 13:15:27.847176 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-lmhj4" event={"ID":"de5cd487-58d9-4c86-a4e2-f6b000a351f5","Type":"ContainerStarted","Data":"a49a5e2664f524e78c968178bf7d3127ac118f4bd4c9418f5d0fa9afefb907a6"} Feb 17 13:15:27 crc kubenswrapper[4955]: I0217 13:15:27.849481 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-69dts" event={"ID":"d4a28595-c677-4966-8d3a-1ad5d299fd51","Type":"ContainerStarted","Data":"5179006dd2cb8f432fd0f750dc0e1617070d7ba8b11b6c62659d861b0d57154f"} Feb 17 13:15:31 crc kubenswrapper[4955]: I0217 13:15:31.876672 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-69dts" event={"ID":"d4a28595-c677-4966-8d3a-1ad5d299fd51","Type":"ContainerStarted","Data":"29ae4a0076370172d227ddd514ef358f1e72b5d9aa18b592b255889dd15e2416"} Feb 17 13:15:31 crc kubenswrapper[4955]: I0217 13:15:31.877511 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-687f57d79b-69dts" Feb 17 13:15:31 crc kubenswrapper[4955]: I0217 13:15:31.880744 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-lg6zf" event={"ID":"54536ade-766d-4cf8-b1b5-057ccb961b13","Type":"ContainerStarted","Data":"c9bfc95ff649d69d728951e0f41b00eade0131f73de252da79b3b814917d01dc"} Feb 17 13:15:31 crc kubenswrapper[4955]: I0217 13:15:31.883217 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-lmhj4" event={"ID":"de5cd487-58d9-4c86-a4e2-f6b000a351f5","Type":"ContainerStarted","Data":"01b7f1acb286a364b9f500ad7fc858e81b49c97fbf521075d426175826b5b2bf"} Feb 17 13:15:31 crc kubenswrapper[4955]: I0217 13:15:31.906709 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-687f57d79b-69dts" podStartSLOduration=2.468347497 podStartE2EDuration="5.906680881s" podCreationTimestamp="2026-02-17 13:15:26 +0000 UTC" firstStartedPulling="2026-02-17 13:15:27.645807774 +0000 UTC m=+666.168537327" lastFinishedPulling="2026-02-17 13:15:31.084141158 +0000 UTC m=+669.606870711" observedRunningTime="2026-02-17 13:15:31.899925299 +0000 UTC m=+670.422654902" watchObservedRunningTime="2026-02-17 13:15:31.906680881 +0000 UTC m=+670.429410454" Feb 17 13:15:31 crc kubenswrapper[4955]: I0217 13:15:31.930814 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-858654f9db-lg6zf" podStartSLOduration=2.406060416 podStartE2EDuration="5.930752827s" podCreationTimestamp="2026-02-17 13:15:26 +0000 UTC" firstStartedPulling="2026-02-17 13:15:27.647556834 +0000 UTC m=+666.170286377" lastFinishedPulling="2026-02-17 13:15:31.172249245 +0000 UTC m=+669.694978788" observedRunningTime="2026-02-17 13:15:31.924294423 +0000 UTC m=+670.447024026" watchObservedRunningTime="2026-02-17 13:15:31.930752827 +0000 UTC m=+670.453482410" Feb 17 13:15:31 crc kubenswrapper[4955]: I0217 13:15:31.962307 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-cf98fcc89-lmhj4" podStartSLOduration=2.338533623 podStartE2EDuration="5.962283474s" podCreationTimestamp="2026-02-17 13:15:26 +0000 UTC" firstStartedPulling="2026-02-17 13:15:27.380203245 +0000 UTC m=+665.902932788" lastFinishedPulling="2026-02-17 13:15:31.003953076 +0000 UTC m=+669.526682639" observedRunningTime="2026-02-17 13:15:31.95966256 +0000 UTC m=+670.482392173" watchObservedRunningTime="2026-02-17 13:15:31.962283474 +0000 UTC m=+670.485013027" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.082771 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-dcwgg"] Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.083837 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="ovn-controller" containerID="cri-o://89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18" gracePeriod=30 Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.083870 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff" gracePeriod=30 Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.083870 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="nbdb" containerID="cri-o://23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44" gracePeriod=30 Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.083958 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="northd" containerID="cri-o://11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4" gracePeriod=30 Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.084005 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="sbdb" containerID="cri-o://e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648" gracePeriod=30 Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.084042 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="ovn-acl-logging" containerID="cri-o://b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636" gracePeriod=30 Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.083979 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="kube-rbac-proxy-node" containerID="cri-o://d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748" gracePeriod=30 Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.129653 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="ovnkube-controller" containerID="cri-o://0bf8e3c7005480c233ef81b5a37bc7d925eadebc225ecd7d5345fa4e96383b90" gracePeriod=30 Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.183390 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-687f57d79b-69dts" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.433181 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dcwgg_dcbc7a69-58e7-4b7a-bb84-96276819ce88/ovnkube-controller/3.log" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.435503 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dcwgg_dcbc7a69-58e7-4b7a-bb84-96276819ce88/ovn-acl-logging/0.log" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.435998 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dcwgg_dcbc7a69-58e7-4b7a-bb84-96276819ce88/ovn-controller/0.log" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.436448 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.457333 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2m6wz\" (UniqueName: \"kubernetes.io/projected/dcbc7a69-58e7-4b7a-bb84-96276819ce88-kube-api-access-2m6wz\") pod \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.457647 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/dcbc7a69-58e7-4b7a-bb84-96276819ce88-ovnkube-script-lib\") pod \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.457673 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-host-cni-bin\") pod \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.457692 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-log-socket\") pod \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.457841 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "dcbc7a69-58e7-4b7a-bb84-96276819ce88" (UID: "dcbc7a69-58e7-4b7a-bb84-96276819ce88"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.457863 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-run-systemd\") pod \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.457947 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-systemd-units\") pod \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.458204 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dcbc7a69-58e7-4b7a-bb84-96276819ce88-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "dcbc7a69-58e7-4b7a-bb84-96276819ce88" (UID: "dcbc7a69-58e7-4b7a-bb84-96276819ce88"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.458256 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "dcbc7a69-58e7-4b7a-bb84-96276819ce88" (UID: "dcbc7a69-58e7-4b7a-bb84-96276819ce88"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.458298 4955 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/dcbc7a69-58e7-4b7a-bb84-96276819ce88-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.458310 4955 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-host-cni-bin\") on node \"crc\" DevicePath \"\"" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.458325 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-log-socket" (OuterVolumeSpecName: "log-socket") pod "dcbc7a69-58e7-4b7a-bb84-96276819ce88" (UID: "dcbc7a69-58e7-4b7a-bb84-96276819ce88"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.469946 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dcbc7a69-58e7-4b7a-bb84-96276819ce88-kube-api-access-2m6wz" (OuterVolumeSpecName: "kube-api-access-2m6wz") pod "dcbc7a69-58e7-4b7a-bb84-96276819ce88" (UID: "dcbc7a69-58e7-4b7a-bb84-96276819ce88"). InnerVolumeSpecName "kube-api-access-2m6wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.471738 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "dcbc7a69-58e7-4b7a-bb84-96276819ce88" (UID: "dcbc7a69-58e7-4b7a-bb84-96276819ce88"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.487145 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-t5dsn"] Feb 17 13:15:37 crc kubenswrapper[4955]: E0217 13:15:37.487407 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="sbdb" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.487428 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="sbdb" Feb 17 13:15:37 crc kubenswrapper[4955]: E0217 13:15:37.487445 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="ovnkube-controller" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.487456 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="ovnkube-controller" Feb 17 13:15:37 crc kubenswrapper[4955]: E0217 13:15:37.487476 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="ovn-controller" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.487485 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="ovn-controller" Feb 17 13:15:37 crc kubenswrapper[4955]: E0217 13:15:37.487502 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="kube-rbac-proxy-ovn-metrics" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.487511 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="kube-rbac-proxy-ovn-metrics" Feb 17 13:15:37 crc kubenswrapper[4955]: E0217 13:15:37.487526 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="kube-rbac-proxy-node" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.487535 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="kube-rbac-proxy-node" Feb 17 13:15:37 crc kubenswrapper[4955]: E0217 13:15:37.487551 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="ovnkube-controller" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.487561 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="ovnkube-controller" Feb 17 13:15:37 crc kubenswrapper[4955]: E0217 13:15:37.487575 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="kubecfg-setup" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.487585 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="kubecfg-setup" Feb 17 13:15:37 crc kubenswrapper[4955]: E0217 13:15:37.487598 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="nbdb" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.487607 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="nbdb" Feb 17 13:15:37 crc kubenswrapper[4955]: E0217 13:15:37.487622 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="northd" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.487631 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="northd" Feb 17 13:15:37 crc kubenswrapper[4955]: E0217 13:15:37.487646 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="ovnkube-controller" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.487655 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="ovnkube-controller" Feb 17 13:15:37 crc kubenswrapper[4955]: E0217 13:15:37.487670 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="ovn-acl-logging" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.487681 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="ovn-acl-logging" Feb 17 13:15:37 crc kubenswrapper[4955]: E0217 13:15:37.487696 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="ovnkube-controller" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.487708 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="ovnkube-controller" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.489233 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="nbdb" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.489272 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="ovnkube-controller" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.489301 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="ovn-acl-logging" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.489323 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="ovnkube-controller" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.489339 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="sbdb" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.489366 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="kube-rbac-proxy-node" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.489389 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="ovn-controller" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.489416 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="northd" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.489439 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="kube-rbac-proxy-ovn-metrics" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.489454 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="ovnkube-controller" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.489474 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="ovnkube-controller" Feb 17 13:15:37 crc kubenswrapper[4955]: E0217 13:15:37.490158 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="ovnkube-controller" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.490195 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="ovnkube-controller" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.491118 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerName="ovnkube-controller" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.498408 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.559157 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-host-cni-netd\") pod \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.559213 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/dcbc7a69-58e7-4b7a-bb84-96276819ce88-ovnkube-config\") pod \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.559246 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-host-var-lib-cni-networks-ovn-kubernetes\") pod \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.559295 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-var-lib-openvswitch\") pod \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.559321 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-etc-openvswitch\") pod \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.559347 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-node-log\") pod \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.559375 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-host-slash\") pod \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.559415 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "dcbc7a69-58e7-4b7a-bb84-96276819ce88" (UID: "dcbc7a69-58e7-4b7a-bb84-96276819ce88"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.559476 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "dcbc7a69-58e7-4b7a-bb84-96276819ce88" (UID: "dcbc7a69-58e7-4b7a-bb84-96276819ce88"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.559532 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/dcbc7a69-58e7-4b7a-bb84-96276819ce88-ovn-node-metrics-cert\") pod \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.559524 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-node-log" (OuterVolumeSpecName: "node-log") pod "dcbc7a69-58e7-4b7a-bb84-96276819ce88" (UID: "dcbc7a69-58e7-4b7a-bb84-96276819ce88"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.559568 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-host-slash" (OuterVolumeSpecName: "host-slash") pod "dcbc7a69-58e7-4b7a-bb84-96276819ce88" (UID: "dcbc7a69-58e7-4b7a-bb84-96276819ce88"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.559623 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "dcbc7a69-58e7-4b7a-bb84-96276819ce88" (UID: "dcbc7a69-58e7-4b7a-bb84-96276819ce88"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.559653 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dcbc7a69-58e7-4b7a-bb84-96276819ce88-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "dcbc7a69-58e7-4b7a-bb84-96276819ce88" (UID: "dcbc7a69-58e7-4b7a-bb84-96276819ce88"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.559756 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "dcbc7a69-58e7-4b7a-bb84-96276819ce88" (UID: "dcbc7a69-58e7-4b7a-bb84-96276819ce88"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.560029 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-host-run-ovn-kubernetes\") pod \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.560062 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-host-run-netns\") pod \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.560112 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/dcbc7a69-58e7-4b7a-bb84-96276819ce88-env-overrides\") pod \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.560219 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-run-ovn\") pod \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.560268 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-run-openvswitch\") pod \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.560289 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-host-kubelet\") pod \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\" (UID: \"dcbc7a69-58e7-4b7a-bb84-96276819ce88\") " Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.560422 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "dcbc7a69-58e7-4b7a-bb84-96276819ce88" (UID: "dcbc7a69-58e7-4b7a-bb84-96276819ce88"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.560449 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "dcbc7a69-58e7-4b7a-bb84-96276819ce88" (UID: "dcbc7a69-58e7-4b7a-bb84-96276819ce88"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.560465 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "dcbc7a69-58e7-4b7a-bb84-96276819ce88" (UID: "dcbc7a69-58e7-4b7a-bb84-96276819ce88"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.560463 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-systemd-units\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.560481 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "dcbc7a69-58e7-4b7a-bb84-96276819ce88" (UID: "dcbc7a69-58e7-4b7a-bb84-96276819ce88"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.560503 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bed50bf4-6d51-4be6-b446-e2835963b107-ovnkube-config\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.560515 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "dcbc7a69-58e7-4b7a-bb84-96276819ce88" (UID: "dcbc7a69-58e7-4b7a-bb84-96276819ce88"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.560619 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rx9l8\" (UniqueName: \"kubernetes.io/projected/bed50bf4-6d51-4be6-b446-e2835963b107-kube-api-access-rx9l8\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.560652 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bed50bf4-6d51-4be6-b446-e2835963b107-ovn-node-metrics-cert\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.560690 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-host-cni-netd\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.560738 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/bed50bf4-6d51-4be6-b446-e2835963b107-ovnkube-script-lib\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.560760 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-run-ovn\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.560775 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-host-run-ovn-kubernetes\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.560935 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-host-kubelet\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.560979 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-host-cni-bin\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.561045 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.561104 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-host-run-netns\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.561123 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-node-log\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.561173 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-log-socket\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.561199 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-run-systemd\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.561220 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-host-slash\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.561237 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-etc-openvswitch\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.561279 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bed50bf4-6d51-4be6-b446-e2835963b107-env-overrides\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.561294 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-run-openvswitch\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.561318 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-var-lib-openvswitch\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.561492 4955 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-host-cni-netd\") on node \"crc\" DevicePath \"\"" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.561517 4955 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/dcbc7a69-58e7-4b7a-bb84-96276819ce88-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.561530 4955 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.561544 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2m6wz\" (UniqueName: \"kubernetes.io/projected/dcbc7a69-58e7-4b7a-bb84-96276819ce88-kube-api-access-2m6wz\") on node \"crc\" DevicePath \"\"" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.561557 4955 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-log-socket\") on node \"crc\" DevicePath \"\"" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.561568 4955 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-run-systemd\") on node \"crc\" DevicePath \"\"" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.561580 4955 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-systemd-units\") on node \"crc\" DevicePath \"\"" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.561597 4955 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.561628 4955 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.561643 4955 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-node-log\") on node \"crc\" DevicePath \"\"" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.561652 4955 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-host-slash\") on node \"crc\" DevicePath \"\"" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.561662 4955 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.561671 4955 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-host-run-netns\") on node \"crc\" DevicePath \"\"" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.561680 4955 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.561689 4955 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-run-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.561698 4955 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/dcbc7a69-58e7-4b7a-bb84-96276819ce88-host-kubelet\") on node \"crc\" DevicePath \"\"" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.562909 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dcbc7a69-58e7-4b7a-bb84-96276819ce88-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "dcbc7a69-58e7-4b7a-bb84-96276819ce88" (UID: "dcbc7a69-58e7-4b7a-bb84-96276819ce88"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.565430 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dcbc7a69-58e7-4b7a-bb84-96276819ce88-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "dcbc7a69-58e7-4b7a-bb84-96276819ce88" (UID: "dcbc7a69-58e7-4b7a-bb84-96276819ce88"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.662632 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bed50bf4-6d51-4be6-b446-e2835963b107-ovn-node-metrics-cert\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.662687 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-host-cni-netd\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.662709 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/bed50bf4-6d51-4be6-b446-e2835963b107-ovnkube-script-lib\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.662732 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-run-ovn\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.662752 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-host-run-ovn-kubernetes\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.662823 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-host-kubelet\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.662849 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-host-cni-bin\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.662874 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.662908 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-run-ovn\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.662913 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-host-cni-netd\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.662927 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-host-run-netns\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.662955 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-host-run-netns\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.663011 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-host-run-ovn-kubernetes\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.663050 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-host-kubelet\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.663112 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-node-log\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.663135 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-host-cni-bin\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.663151 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-node-log\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.663142 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.663392 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-log-socket\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.663415 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-run-systemd\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.663435 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-host-slash\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.663453 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-etc-openvswitch\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.663482 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bed50bf4-6d51-4be6-b446-e2835963b107-env-overrides\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.663502 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-run-openvswitch\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.663524 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-var-lib-openvswitch\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.663556 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-systemd-units\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.663574 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bed50bf4-6d51-4be6-b446-e2835963b107-ovnkube-config\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.663598 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rx9l8\" (UniqueName: \"kubernetes.io/projected/bed50bf4-6d51-4be6-b446-e2835963b107-kube-api-access-rx9l8\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.663622 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-etc-openvswitch\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.663745 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-run-systemd\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.663771 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-host-slash\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.663846 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-var-lib-openvswitch\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.663644 4955 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/dcbc7a69-58e7-4b7a-bb84-96276819ce88-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.663888 4955 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/dcbc7a69-58e7-4b7a-bb84-96276819ce88-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.663906 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-run-openvswitch\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.663919 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-log-socket\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.664104 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/bed50bf4-6d51-4be6-b446-e2835963b107-systemd-units\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.664463 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bed50bf4-6d51-4be6-b446-e2835963b107-env-overrides\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.665166 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bed50bf4-6d51-4be6-b446-e2835963b107-ovnkube-config\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.665188 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/bed50bf4-6d51-4be6-b446-e2835963b107-ovnkube-script-lib\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.669309 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bed50bf4-6d51-4be6-b446-e2835963b107-ovn-node-metrics-cert\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.692677 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rx9l8\" (UniqueName: \"kubernetes.io/projected/bed50bf4-6d51-4be6-b446-e2835963b107-kube-api-access-rx9l8\") pod \"ovnkube-node-t5dsn\" (UID: \"bed50bf4-6d51-4be6-b446-e2835963b107\") " pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.815611 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.931144 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9hx9f_1e57bdf8-e537-402a-9815-609fecdd68ea/kube-multus/2.log" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.932063 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9hx9f_1e57bdf8-e537-402a-9815-609fecdd68ea/kube-multus/1.log" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.932225 4955 generic.go:334] "Generic (PLEG): container finished" podID="1e57bdf8-e537-402a-9815-609fecdd68ea" containerID="7db7f26cfeecd2fd564d165a58ac1b15757ebb5167a56c82a8fc8def9bdca610" exitCode=2 Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.932306 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-9hx9f" event={"ID":"1e57bdf8-e537-402a-9815-609fecdd68ea","Type":"ContainerDied","Data":"7db7f26cfeecd2fd564d165a58ac1b15757ebb5167a56c82a8fc8def9bdca610"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.932401 4955 scope.go:117] "RemoveContainer" containerID="c81948fc2806ea8718909fd269c029efc6a26675e45359e971272354eccfa74d" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.932993 4955 scope.go:117] "RemoveContainer" containerID="7db7f26cfeecd2fd564d165a58ac1b15757ebb5167a56c82a8fc8def9bdca610" Feb 17 13:15:37 crc kubenswrapper[4955]: E0217 13:15:37.933631 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-9hx9f_openshift-multus(1e57bdf8-e537-402a-9815-609fecdd68ea)\"" pod="openshift-multus/multus-9hx9f" podUID="1e57bdf8-e537-402a-9815-609fecdd68ea" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.940953 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dcwgg_dcbc7a69-58e7-4b7a-bb84-96276819ce88/ovnkube-controller/3.log" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.944142 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dcwgg_dcbc7a69-58e7-4b7a-bb84-96276819ce88/ovn-acl-logging/0.log" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.944739 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dcwgg_dcbc7a69-58e7-4b7a-bb84-96276819ce88/ovn-controller/0.log" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.945530 4955 generic.go:334] "Generic (PLEG): container finished" podID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerID="0bf8e3c7005480c233ef81b5a37bc7d925eadebc225ecd7d5345fa4e96383b90" exitCode=0 Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.945579 4955 generic.go:334] "Generic (PLEG): container finished" podID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerID="e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648" exitCode=0 Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.945599 4955 generic.go:334] "Generic (PLEG): container finished" podID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerID="23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44" exitCode=0 Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.945612 4955 generic.go:334] "Generic (PLEG): container finished" podID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerID="11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4" exitCode=0 Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.945624 4955 generic.go:334] "Generic (PLEG): container finished" podID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerID="ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff" exitCode=0 Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.945637 4955 generic.go:334] "Generic (PLEG): container finished" podID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerID="d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748" exitCode=0 Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.945649 4955 generic.go:334] "Generic (PLEG): container finished" podID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerID="b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636" exitCode=143 Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.945663 4955 generic.go:334] "Generic (PLEG): container finished" podID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" containerID="89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18" exitCode=143 Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.945644 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" event={"ID":"dcbc7a69-58e7-4b7a-bb84-96276819ce88","Type":"ContainerDied","Data":"0bf8e3c7005480c233ef81b5a37bc7d925eadebc225ecd7d5345fa4e96383b90"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.945803 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" event={"ID":"dcbc7a69-58e7-4b7a-bb84-96276819ce88","Type":"ContainerDied","Data":"e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.945830 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" event={"ID":"dcbc7a69-58e7-4b7a-bb84-96276819ce88","Type":"ContainerDied","Data":"23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.945848 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" event={"ID":"dcbc7a69-58e7-4b7a-bb84-96276819ce88","Type":"ContainerDied","Data":"11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.945865 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" event={"ID":"dcbc7a69-58e7-4b7a-bb84-96276819ce88","Type":"ContainerDied","Data":"ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.945881 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" event={"ID":"dcbc7a69-58e7-4b7a-bb84-96276819ce88","Type":"ContainerDied","Data":"d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.945898 4955 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0bf8e3c7005480c233ef81b5a37bc7d925eadebc225ecd7d5345fa4e96383b90"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.945917 4955 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.945927 4955 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.945936 4955 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.945945 4955 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.945955 4955 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.945965 4955 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.945975 4955 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.945985 4955 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.945994 4955 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.946009 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" event={"ID":"dcbc7a69-58e7-4b7a-bb84-96276819ce88","Type":"ContainerDied","Data":"b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.946026 4955 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0bf8e3c7005480c233ef81b5a37bc7d925eadebc225ecd7d5345fa4e96383b90"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.946040 4955 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.946050 4955 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.946059 4955 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.946068 4955 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.946077 4955 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.946085 4955 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.946094 4955 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.946104 4955 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.946113 4955 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.946126 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" event={"ID":"dcbc7a69-58e7-4b7a-bb84-96276819ce88","Type":"ContainerDied","Data":"89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.946141 4955 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0bf8e3c7005480c233ef81b5a37bc7d925eadebc225ecd7d5345fa4e96383b90"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.946152 4955 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.946162 4955 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.946171 4955 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.946180 4955 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.946189 4955 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.946199 4955 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.946208 4955 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.946216 4955 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.946225 4955 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.946237 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" event={"ID":"dcbc7a69-58e7-4b7a-bb84-96276819ce88","Type":"ContainerDied","Data":"4d43867b072716a140df4d0f846947f25ee3e32f7c51e8569e825b40bbc38dda"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.946253 4955 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0bf8e3c7005480c233ef81b5a37bc7d925eadebc225ecd7d5345fa4e96383b90"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.946264 4955 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.946273 4955 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.946282 4955 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.946291 4955 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.946300 4955 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.946309 4955 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.946318 4955 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.946329 4955 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.946338 4955 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.947140 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dcwgg" Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.949211 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" event={"ID":"bed50bf4-6d51-4be6-b446-e2835963b107","Type":"ContainerStarted","Data":"574294cae5703fdbf915e6944520b5d6889a248677397c08174f048646c4b6c8"} Feb 17 13:15:37 crc kubenswrapper[4955]: I0217 13:15:37.985658 4955 scope.go:117] "RemoveContainer" containerID="0bf8e3c7005480c233ef81b5a37bc7d925eadebc225ecd7d5345fa4e96383b90" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.018392 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-dcwgg"] Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.019161 4955 scope.go:117] "RemoveContainer" containerID="528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.027373 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-dcwgg"] Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.058047 4955 scope.go:117] "RemoveContainer" containerID="e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.104690 4955 scope.go:117] "RemoveContainer" containerID="23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.122259 4955 scope.go:117] "RemoveContainer" containerID="11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.138765 4955 scope.go:117] "RemoveContainer" containerID="ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.156177 4955 scope.go:117] "RemoveContainer" containerID="d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.175936 4955 scope.go:117] "RemoveContainer" containerID="b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.192559 4955 scope.go:117] "RemoveContainer" containerID="89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.235499 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dcbc7a69-58e7-4b7a-bb84-96276819ce88" path="/var/lib/kubelet/pods/dcbc7a69-58e7-4b7a-bb84-96276819ce88/volumes" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.243094 4955 scope.go:117] "RemoveContainer" containerID="e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.261936 4955 scope.go:117] "RemoveContainer" containerID="0bf8e3c7005480c233ef81b5a37bc7d925eadebc225ecd7d5345fa4e96383b90" Feb 17 13:15:38 crc kubenswrapper[4955]: E0217 13:15:38.262465 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0bf8e3c7005480c233ef81b5a37bc7d925eadebc225ecd7d5345fa4e96383b90\": container with ID starting with 0bf8e3c7005480c233ef81b5a37bc7d925eadebc225ecd7d5345fa4e96383b90 not found: ID does not exist" containerID="0bf8e3c7005480c233ef81b5a37bc7d925eadebc225ecd7d5345fa4e96383b90" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.262519 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bf8e3c7005480c233ef81b5a37bc7d925eadebc225ecd7d5345fa4e96383b90"} err="failed to get container status \"0bf8e3c7005480c233ef81b5a37bc7d925eadebc225ecd7d5345fa4e96383b90\": rpc error: code = NotFound desc = could not find container \"0bf8e3c7005480c233ef81b5a37bc7d925eadebc225ecd7d5345fa4e96383b90\": container with ID starting with 0bf8e3c7005480c233ef81b5a37bc7d925eadebc225ecd7d5345fa4e96383b90 not found: ID does not exist" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.262551 4955 scope.go:117] "RemoveContainer" containerID="528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6" Feb 17 13:15:38 crc kubenswrapper[4955]: E0217 13:15:38.262884 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6\": container with ID starting with 528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6 not found: ID does not exist" containerID="528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.262921 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6"} err="failed to get container status \"528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6\": rpc error: code = NotFound desc = could not find container \"528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6\": container with ID starting with 528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6 not found: ID does not exist" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.262947 4955 scope.go:117] "RemoveContainer" containerID="e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648" Feb 17 13:15:38 crc kubenswrapper[4955]: E0217 13:15:38.263228 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648\": container with ID starting with e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648 not found: ID does not exist" containerID="e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.263258 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648"} err="failed to get container status \"e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648\": rpc error: code = NotFound desc = could not find container \"e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648\": container with ID starting with e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648 not found: ID does not exist" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.263279 4955 scope.go:117] "RemoveContainer" containerID="23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44" Feb 17 13:15:38 crc kubenswrapper[4955]: E0217 13:15:38.263516 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44\": container with ID starting with 23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44 not found: ID does not exist" containerID="23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.263540 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44"} err="failed to get container status \"23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44\": rpc error: code = NotFound desc = could not find container \"23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44\": container with ID starting with 23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44 not found: ID does not exist" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.263555 4955 scope.go:117] "RemoveContainer" containerID="11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4" Feb 17 13:15:38 crc kubenswrapper[4955]: E0217 13:15:38.263891 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4\": container with ID starting with 11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4 not found: ID does not exist" containerID="11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.263934 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4"} err="failed to get container status \"11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4\": rpc error: code = NotFound desc = could not find container \"11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4\": container with ID starting with 11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4 not found: ID does not exist" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.263966 4955 scope.go:117] "RemoveContainer" containerID="ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff" Feb 17 13:15:38 crc kubenswrapper[4955]: E0217 13:15:38.264498 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff\": container with ID starting with ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff not found: ID does not exist" containerID="ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.264529 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff"} err="failed to get container status \"ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff\": rpc error: code = NotFound desc = could not find container \"ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff\": container with ID starting with ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff not found: ID does not exist" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.264548 4955 scope.go:117] "RemoveContainer" containerID="d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748" Feb 17 13:15:38 crc kubenswrapper[4955]: E0217 13:15:38.265060 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748\": container with ID starting with d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748 not found: ID does not exist" containerID="d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.265093 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748"} err="failed to get container status \"d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748\": rpc error: code = NotFound desc = could not find container \"d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748\": container with ID starting with d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748 not found: ID does not exist" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.265111 4955 scope.go:117] "RemoveContainer" containerID="b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636" Feb 17 13:15:38 crc kubenswrapper[4955]: E0217 13:15:38.265744 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636\": container with ID starting with b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636 not found: ID does not exist" containerID="b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.265799 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636"} err="failed to get container status \"b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636\": rpc error: code = NotFound desc = could not find container \"b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636\": container with ID starting with b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636 not found: ID does not exist" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.265820 4955 scope.go:117] "RemoveContainer" containerID="89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18" Feb 17 13:15:38 crc kubenswrapper[4955]: E0217 13:15:38.266268 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18\": container with ID starting with 89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18 not found: ID does not exist" containerID="89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.266298 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18"} err="failed to get container status \"89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18\": rpc error: code = NotFound desc = could not find container \"89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18\": container with ID starting with 89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18 not found: ID does not exist" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.266322 4955 scope.go:117] "RemoveContainer" containerID="e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0" Feb 17 13:15:38 crc kubenswrapper[4955]: E0217 13:15:38.266695 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\": container with ID starting with e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0 not found: ID does not exist" containerID="e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.266732 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0"} err="failed to get container status \"e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\": rpc error: code = NotFound desc = could not find container \"e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\": container with ID starting with e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0 not found: ID does not exist" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.266763 4955 scope.go:117] "RemoveContainer" containerID="0bf8e3c7005480c233ef81b5a37bc7d925eadebc225ecd7d5345fa4e96383b90" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.267324 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bf8e3c7005480c233ef81b5a37bc7d925eadebc225ecd7d5345fa4e96383b90"} err="failed to get container status \"0bf8e3c7005480c233ef81b5a37bc7d925eadebc225ecd7d5345fa4e96383b90\": rpc error: code = NotFound desc = could not find container \"0bf8e3c7005480c233ef81b5a37bc7d925eadebc225ecd7d5345fa4e96383b90\": container with ID starting with 0bf8e3c7005480c233ef81b5a37bc7d925eadebc225ecd7d5345fa4e96383b90 not found: ID does not exist" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.267352 4955 scope.go:117] "RemoveContainer" containerID="528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.268149 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6"} err="failed to get container status \"528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6\": rpc error: code = NotFound desc = could not find container \"528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6\": container with ID starting with 528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6 not found: ID does not exist" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.268445 4955 scope.go:117] "RemoveContainer" containerID="e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.268936 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648"} err="failed to get container status \"e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648\": rpc error: code = NotFound desc = could not find container \"e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648\": container with ID starting with e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648 not found: ID does not exist" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.268969 4955 scope.go:117] "RemoveContainer" containerID="23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.269287 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44"} err="failed to get container status \"23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44\": rpc error: code = NotFound desc = could not find container \"23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44\": container with ID starting with 23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44 not found: ID does not exist" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.269314 4955 scope.go:117] "RemoveContainer" containerID="11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.269751 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4"} err="failed to get container status \"11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4\": rpc error: code = NotFound desc = could not find container \"11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4\": container with ID starting with 11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4 not found: ID does not exist" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.269817 4955 scope.go:117] "RemoveContainer" containerID="ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.270318 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff"} err="failed to get container status \"ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff\": rpc error: code = NotFound desc = could not find container \"ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff\": container with ID starting with ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff not found: ID does not exist" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.270346 4955 scope.go:117] "RemoveContainer" containerID="d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.270805 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748"} err="failed to get container status \"d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748\": rpc error: code = NotFound desc = could not find container \"d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748\": container with ID starting with d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748 not found: ID does not exist" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.270849 4955 scope.go:117] "RemoveContainer" containerID="b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.271335 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636"} err="failed to get container status \"b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636\": rpc error: code = NotFound desc = could not find container \"b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636\": container with ID starting with b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636 not found: ID does not exist" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.271363 4955 scope.go:117] "RemoveContainer" containerID="89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.271664 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18"} err="failed to get container status \"89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18\": rpc error: code = NotFound desc = could not find container \"89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18\": container with ID starting with 89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18 not found: ID does not exist" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.271695 4955 scope.go:117] "RemoveContainer" containerID="e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.272950 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0"} err="failed to get container status \"e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\": rpc error: code = NotFound desc = could not find container \"e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\": container with ID starting with e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0 not found: ID does not exist" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.272979 4955 scope.go:117] "RemoveContainer" containerID="0bf8e3c7005480c233ef81b5a37bc7d925eadebc225ecd7d5345fa4e96383b90" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.273627 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bf8e3c7005480c233ef81b5a37bc7d925eadebc225ecd7d5345fa4e96383b90"} err="failed to get container status \"0bf8e3c7005480c233ef81b5a37bc7d925eadebc225ecd7d5345fa4e96383b90\": rpc error: code = NotFound desc = could not find container \"0bf8e3c7005480c233ef81b5a37bc7d925eadebc225ecd7d5345fa4e96383b90\": container with ID starting with 0bf8e3c7005480c233ef81b5a37bc7d925eadebc225ecd7d5345fa4e96383b90 not found: ID does not exist" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.273679 4955 scope.go:117] "RemoveContainer" containerID="528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.274084 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6"} err="failed to get container status \"528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6\": rpc error: code = NotFound desc = could not find container \"528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6\": container with ID starting with 528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6 not found: ID does not exist" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.274109 4955 scope.go:117] "RemoveContainer" containerID="e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.274501 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648"} err="failed to get container status \"e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648\": rpc error: code = NotFound desc = could not find container \"e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648\": container with ID starting with e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648 not found: ID does not exist" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.274530 4955 scope.go:117] "RemoveContainer" containerID="23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.274864 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44"} err="failed to get container status \"23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44\": rpc error: code = NotFound desc = could not find container \"23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44\": container with ID starting with 23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44 not found: ID does not exist" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.274910 4955 scope.go:117] "RemoveContainer" containerID="11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.275275 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4"} err="failed to get container status \"11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4\": rpc error: code = NotFound desc = could not find container \"11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4\": container with ID starting with 11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4 not found: ID does not exist" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.275299 4955 scope.go:117] "RemoveContainer" containerID="ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.275805 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff"} err="failed to get container status \"ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff\": rpc error: code = NotFound desc = could not find container \"ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff\": container with ID starting with ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff not found: ID does not exist" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.275844 4955 scope.go:117] "RemoveContainer" containerID="d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.276207 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748"} err="failed to get container status \"d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748\": rpc error: code = NotFound desc = could not find container \"d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748\": container with ID starting with d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748 not found: ID does not exist" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.276248 4955 scope.go:117] "RemoveContainer" containerID="b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.276771 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636"} err="failed to get container status \"b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636\": rpc error: code = NotFound desc = could not find container \"b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636\": container with ID starting with b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636 not found: ID does not exist" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.276845 4955 scope.go:117] "RemoveContainer" containerID="89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.277399 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18"} err="failed to get container status \"89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18\": rpc error: code = NotFound desc = could not find container \"89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18\": container with ID starting with 89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18 not found: ID does not exist" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.277437 4955 scope.go:117] "RemoveContainer" containerID="e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.277847 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0"} err="failed to get container status \"e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\": rpc error: code = NotFound desc = could not find container \"e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\": container with ID starting with e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0 not found: ID does not exist" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.277888 4955 scope.go:117] "RemoveContainer" containerID="0bf8e3c7005480c233ef81b5a37bc7d925eadebc225ecd7d5345fa4e96383b90" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.278285 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bf8e3c7005480c233ef81b5a37bc7d925eadebc225ecd7d5345fa4e96383b90"} err="failed to get container status \"0bf8e3c7005480c233ef81b5a37bc7d925eadebc225ecd7d5345fa4e96383b90\": rpc error: code = NotFound desc = could not find container \"0bf8e3c7005480c233ef81b5a37bc7d925eadebc225ecd7d5345fa4e96383b90\": container with ID starting with 0bf8e3c7005480c233ef81b5a37bc7d925eadebc225ecd7d5345fa4e96383b90 not found: ID does not exist" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.278323 4955 scope.go:117] "RemoveContainer" containerID="528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.278884 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6"} err="failed to get container status \"528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6\": rpc error: code = NotFound desc = could not find container \"528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6\": container with ID starting with 528a137c21197e3e77f96fde81eec39e134a8136c92881f96d6a50f96c8989d6 not found: ID does not exist" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.278916 4955 scope.go:117] "RemoveContainer" containerID="e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.279455 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648"} err="failed to get container status \"e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648\": rpc error: code = NotFound desc = could not find container \"e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648\": container with ID starting with e0e6cb330520063f4fae4fb074c5e307b14e631525d7f234d37bc9cf1a945648 not found: ID does not exist" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.279508 4955 scope.go:117] "RemoveContainer" containerID="23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.280020 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44"} err="failed to get container status \"23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44\": rpc error: code = NotFound desc = could not find container \"23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44\": container with ID starting with 23fabee62b2eaac4f96e05b6a190a97b5581e7153a3dfbe4a49db184dd6c4a44 not found: ID does not exist" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.280057 4955 scope.go:117] "RemoveContainer" containerID="11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.280391 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4"} err="failed to get container status \"11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4\": rpc error: code = NotFound desc = could not find container \"11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4\": container with ID starting with 11913c8e607355e0081a11b27491a4e344126325fb212068449095dd1368d3a4 not found: ID does not exist" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.280438 4955 scope.go:117] "RemoveContainer" containerID="ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.280733 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff"} err="failed to get container status \"ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff\": rpc error: code = NotFound desc = could not find container \"ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff\": container with ID starting with ef4972c7c532c4d4d294f11473c0b421ec363c9a81e237c7687050b4436a86ff not found: ID does not exist" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.280768 4955 scope.go:117] "RemoveContainer" containerID="d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.281271 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748"} err="failed to get container status \"d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748\": rpc error: code = NotFound desc = could not find container \"d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748\": container with ID starting with d62f36205a4f721d2b70bfe7d7df2c7f7c480e322334d894fd0562f817ae1748 not found: ID does not exist" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.281315 4955 scope.go:117] "RemoveContainer" containerID="b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.281636 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636"} err="failed to get container status \"b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636\": rpc error: code = NotFound desc = could not find container \"b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636\": container with ID starting with b38db0f8ef18847384fb1f782994667ca04ca0e2685e4e43d58fa50b2b0ee636 not found: ID does not exist" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.281665 4955 scope.go:117] "RemoveContainer" containerID="89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.282171 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18"} err="failed to get container status \"89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18\": rpc error: code = NotFound desc = could not find container \"89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18\": container with ID starting with 89684c054c1fe500cc3474aeab607acc98dbee9b3d2b3a1f5e9b0641ce490a18 not found: ID does not exist" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.282206 4955 scope.go:117] "RemoveContainer" containerID="e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.282540 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0"} err="failed to get container status \"e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\": rpc error: code = NotFound desc = could not find container \"e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0\": container with ID starting with e4193e27d07b7d1ee62897baa4ae6bef8bfab536c1cac4a79f09edd3cc97b5f0 not found: ID does not exist" Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.962399 4955 generic.go:334] "Generic (PLEG): container finished" podID="bed50bf4-6d51-4be6-b446-e2835963b107" containerID="f63a3d5963f2561bada6c22d0e08cb42587f87481800abe30b41747aa2b455b4" exitCode=0 Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.962494 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" event={"ID":"bed50bf4-6d51-4be6-b446-e2835963b107","Type":"ContainerDied","Data":"f63a3d5963f2561bada6c22d0e08cb42587f87481800abe30b41747aa2b455b4"} Feb 17 13:15:38 crc kubenswrapper[4955]: I0217 13:15:38.966477 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9hx9f_1e57bdf8-e537-402a-9815-609fecdd68ea/kube-multus/2.log" Feb 17 13:15:39 crc kubenswrapper[4955]: I0217 13:15:39.979937 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" event={"ID":"bed50bf4-6d51-4be6-b446-e2835963b107","Type":"ContainerStarted","Data":"72e7862c10579f05e3892ffee1b4e332f0fa096140735f131998272343363d83"} Feb 17 13:15:39 crc kubenswrapper[4955]: I0217 13:15:39.980338 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" event={"ID":"bed50bf4-6d51-4be6-b446-e2835963b107","Type":"ContainerStarted","Data":"5fee13df9e5633a7025ce69677a903146fb48c3da07ce7e9192c5af522bc0a77"} Feb 17 13:15:39 crc kubenswrapper[4955]: I0217 13:15:39.980362 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" event={"ID":"bed50bf4-6d51-4be6-b446-e2835963b107","Type":"ContainerStarted","Data":"603673137a1794af66df037b076a503d1a673db3d5a954aded0bf1d3349da54f"} Feb 17 13:15:39 crc kubenswrapper[4955]: I0217 13:15:39.980382 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" event={"ID":"bed50bf4-6d51-4be6-b446-e2835963b107","Type":"ContainerStarted","Data":"f6717ef2031a543704dedea3e5a0625575b67952d32c231fe60e2adbef316314"} Feb 17 13:15:39 crc kubenswrapper[4955]: I0217 13:15:39.980396 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" event={"ID":"bed50bf4-6d51-4be6-b446-e2835963b107","Type":"ContainerStarted","Data":"710c96f914ed6802c8e41f5e31f1e51a96b859c9461ff976fa13f945d06fe60d"} Feb 17 13:15:39 crc kubenswrapper[4955]: I0217 13:15:39.980411 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" event={"ID":"bed50bf4-6d51-4be6-b446-e2835963b107","Type":"ContainerStarted","Data":"6c98812813851b5db98019552ef091072de2e4ec0d8f4b9c9b065a29c746520a"} Feb 17 13:15:43 crc kubenswrapper[4955]: I0217 13:15:43.010752 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" event={"ID":"bed50bf4-6d51-4be6-b446-e2835963b107","Type":"ContainerStarted","Data":"acccbef694dec172541808743f681e83c431b6ce28cc615283af0083eb18344d"} Feb 17 13:15:45 crc kubenswrapper[4955]: I0217 13:15:45.027605 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" event={"ID":"bed50bf4-6d51-4be6-b446-e2835963b107","Type":"ContainerStarted","Data":"524a6b3efbd10ff41dd450b59c3135055ca5236f1efc0301e48e57e2433e29e2"} Feb 17 13:15:45 crc kubenswrapper[4955]: I0217 13:15:45.027955 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:45 crc kubenswrapper[4955]: I0217 13:15:45.027972 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:45 crc kubenswrapper[4955]: I0217 13:15:45.027983 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:45 crc kubenswrapper[4955]: I0217 13:15:45.068248 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:45 crc kubenswrapper[4955]: I0217 13:15:45.072654 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" podStartSLOduration=8.072632873 podStartE2EDuration="8.072632873s" podCreationTimestamp="2026-02-17 13:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:15:45.068437974 +0000 UTC m=+683.591167517" watchObservedRunningTime="2026-02-17 13:15:45.072632873 +0000 UTC m=+683.595362416" Feb 17 13:15:45 crc kubenswrapper[4955]: I0217 13:15:45.080349 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:15:53 crc kubenswrapper[4955]: I0217 13:15:53.222992 4955 scope.go:117] "RemoveContainer" containerID="7db7f26cfeecd2fd564d165a58ac1b15757ebb5167a56c82a8fc8def9bdca610" Feb 17 13:15:53 crc kubenswrapper[4955]: E0217 13:15:53.224401 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-9hx9f_openshift-multus(1e57bdf8-e537-402a-9815-609fecdd68ea)\"" pod="openshift-multus/multus-9hx9f" podUID="1e57bdf8-e537-402a-9815-609fecdd68ea" Feb 17 13:16:04 crc kubenswrapper[4955]: I0217 13:16:04.575191 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 13:16:04 crc kubenswrapper[4955]: I0217 13:16:04.575916 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 13:16:06 crc kubenswrapper[4955]: I0217 13:16:06.223653 4955 scope.go:117] "RemoveContainer" containerID="7db7f26cfeecd2fd564d165a58ac1b15757ebb5167a56c82a8fc8def9bdca610" Feb 17 13:16:07 crc kubenswrapper[4955]: I0217 13:16:07.180635 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9hx9f_1e57bdf8-e537-402a-9815-609fecdd68ea/kube-multus/2.log" Feb 17 13:16:07 crc kubenswrapper[4955]: I0217 13:16:07.181197 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-9hx9f" event={"ID":"1e57bdf8-e537-402a-9815-609fecdd68ea","Type":"ContainerStarted","Data":"567d80857d582ac1fc3834f9dd908ca78d1cdefb2dbe42454866a89cd38087d9"} Feb 17 13:16:07 crc kubenswrapper[4955]: I0217 13:16:07.859315 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-t5dsn" Feb 17 13:16:14 crc kubenswrapper[4955]: I0217 13:16:14.451838 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j"] Feb 17 13:16:14 crc kubenswrapper[4955]: I0217 13:16:14.455200 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j" Feb 17 13:16:14 crc kubenswrapper[4955]: I0217 13:16:14.459739 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 17 13:16:14 crc kubenswrapper[4955]: I0217 13:16:14.475119 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j"] Feb 17 13:16:14 crc kubenswrapper[4955]: I0217 13:16:14.515969 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dtxq\" (UniqueName: \"kubernetes.io/projected/c5bff73f-3f9e-4313-93ac-6276ceaf16d7-kube-api-access-2dtxq\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j\" (UID: \"c5bff73f-3f9e-4313-93ac-6276ceaf16d7\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j" Feb 17 13:16:14 crc kubenswrapper[4955]: I0217 13:16:14.516123 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c5bff73f-3f9e-4313-93ac-6276ceaf16d7-util\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j\" (UID: \"c5bff73f-3f9e-4313-93ac-6276ceaf16d7\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j" Feb 17 13:16:14 crc kubenswrapper[4955]: I0217 13:16:14.516161 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c5bff73f-3f9e-4313-93ac-6276ceaf16d7-bundle\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j\" (UID: \"c5bff73f-3f9e-4313-93ac-6276ceaf16d7\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j" Feb 17 13:16:14 crc kubenswrapper[4955]: I0217 13:16:14.617281 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dtxq\" (UniqueName: \"kubernetes.io/projected/c5bff73f-3f9e-4313-93ac-6276ceaf16d7-kube-api-access-2dtxq\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j\" (UID: \"c5bff73f-3f9e-4313-93ac-6276ceaf16d7\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j" Feb 17 13:16:14 crc kubenswrapper[4955]: I0217 13:16:14.617440 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c5bff73f-3f9e-4313-93ac-6276ceaf16d7-util\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j\" (UID: \"c5bff73f-3f9e-4313-93ac-6276ceaf16d7\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j" Feb 17 13:16:14 crc kubenswrapper[4955]: I0217 13:16:14.617483 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c5bff73f-3f9e-4313-93ac-6276ceaf16d7-bundle\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j\" (UID: \"c5bff73f-3f9e-4313-93ac-6276ceaf16d7\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j" Feb 17 13:16:14 crc kubenswrapper[4955]: I0217 13:16:14.621136 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c5bff73f-3f9e-4313-93ac-6276ceaf16d7-util\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j\" (UID: \"c5bff73f-3f9e-4313-93ac-6276ceaf16d7\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j" Feb 17 13:16:14 crc kubenswrapper[4955]: I0217 13:16:14.621289 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c5bff73f-3f9e-4313-93ac-6276ceaf16d7-bundle\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j\" (UID: \"c5bff73f-3f9e-4313-93ac-6276ceaf16d7\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j" Feb 17 13:16:14 crc kubenswrapper[4955]: I0217 13:16:14.646257 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dtxq\" (UniqueName: \"kubernetes.io/projected/c5bff73f-3f9e-4313-93ac-6276ceaf16d7-kube-api-access-2dtxq\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j\" (UID: \"c5bff73f-3f9e-4313-93ac-6276ceaf16d7\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j" Feb 17 13:16:14 crc kubenswrapper[4955]: I0217 13:16:14.783900 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j" Feb 17 13:16:15 crc kubenswrapper[4955]: I0217 13:16:15.092129 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j"] Feb 17 13:16:15 crc kubenswrapper[4955]: I0217 13:16:15.237064 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j" event={"ID":"c5bff73f-3f9e-4313-93ac-6276ceaf16d7","Type":"ContainerStarted","Data":"93a724356d14abbcd991a8479885bc5ea4bf963df1420b6a51794841ee035caa"} Feb 17 13:16:16 crc kubenswrapper[4955]: I0217 13:16:16.247307 4955 generic.go:334] "Generic (PLEG): container finished" podID="c5bff73f-3f9e-4313-93ac-6276ceaf16d7" containerID="dea8377e1971c3a91412dc87783e3b8ea8782a8b36e2e4463a8d1613cd51954e" exitCode=0 Feb 17 13:16:16 crc kubenswrapper[4955]: I0217 13:16:16.247372 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j" event={"ID":"c5bff73f-3f9e-4313-93ac-6276ceaf16d7","Type":"ContainerDied","Data":"dea8377e1971c3a91412dc87783e3b8ea8782a8b36e2e4463a8d1613cd51954e"} Feb 17 13:16:19 crc kubenswrapper[4955]: I0217 13:16:19.267946 4955 generic.go:334] "Generic (PLEG): container finished" podID="c5bff73f-3f9e-4313-93ac-6276ceaf16d7" containerID="d76901adc5cbdebc3b91f887219f7271f786ada239e36eaf4a532ca7da94e6c4" exitCode=0 Feb 17 13:16:19 crc kubenswrapper[4955]: I0217 13:16:19.268021 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j" event={"ID":"c5bff73f-3f9e-4313-93ac-6276ceaf16d7","Type":"ContainerDied","Data":"d76901adc5cbdebc3b91f887219f7271f786ada239e36eaf4a532ca7da94e6c4"} Feb 17 13:16:20 crc kubenswrapper[4955]: I0217 13:16:20.282571 4955 generic.go:334] "Generic (PLEG): container finished" podID="c5bff73f-3f9e-4313-93ac-6276ceaf16d7" containerID="324d6654adc0e977ad6be5ffa1663266275867773bf643c55169422f73bf749d" exitCode=0 Feb 17 13:16:20 crc kubenswrapper[4955]: I0217 13:16:20.282663 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j" event={"ID":"c5bff73f-3f9e-4313-93ac-6276ceaf16d7","Type":"ContainerDied","Data":"324d6654adc0e977ad6be5ffa1663266275867773bf643c55169422f73bf749d"} Feb 17 13:16:21 crc kubenswrapper[4955]: I0217 13:16:21.625551 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j" Feb 17 13:16:21 crc kubenswrapper[4955]: I0217 13:16:21.817208 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c5bff73f-3f9e-4313-93ac-6276ceaf16d7-bundle\") pod \"c5bff73f-3f9e-4313-93ac-6276ceaf16d7\" (UID: \"c5bff73f-3f9e-4313-93ac-6276ceaf16d7\") " Feb 17 13:16:21 crc kubenswrapper[4955]: I0217 13:16:21.817295 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2dtxq\" (UniqueName: \"kubernetes.io/projected/c5bff73f-3f9e-4313-93ac-6276ceaf16d7-kube-api-access-2dtxq\") pod \"c5bff73f-3f9e-4313-93ac-6276ceaf16d7\" (UID: \"c5bff73f-3f9e-4313-93ac-6276ceaf16d7\") " Feb 17 13:16:21 crc kubenswrapper[4955]: I0217 13:16:21.817376 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c5bff73f-3f9e-4313-93ac-6276ceaf16d7-util\") pod \"c5bff73f-3f9e-4313-93ac-6276ceaf16d7\" (UID: \"c5bff73f-3f9e-4313-93ac-6276ceaf16d7\") " Feb 17 13:16:21 crc kubenswrapper[4955]: I0217 13:16:21.817996 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5bff73f-3f9e-4313-93ac-6276ceaf16d7-bundle" (OuterVolumeSpecName: "bundle") pod "c5bff73f-3f9e-4313-93ac-6276ceaf16d7" (UID: "c5bff73f-3f9e-4313-93ac-6276ceaf16d7"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:16:21 crc kubenswrapper[4955]: I0217 13:16:21.825414 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5bff73f-3f9e-4313-93ac-6276ceaf16d7-kube-api-access-2dtxq" (OuterVolumeSpecName: "kube-api-access-2dtxq") pod "c5bff73f-3f9e-4313-93ac-6276ceaf16d7" (UID: "c5bff73f-3f9e-4313-93ac-6276ceaf16d7"). InnerVolumeSpecName "kube-api-access-2dtxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:16:21 crc kubenswrapper[4955]: I0217 13:16:21.839580 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5bff73f-3f9e-4313-93ac-6276ceaf16d7-util" (OuterVolumeSpecName: "util") pod "c5bff73f-3f9e-4313-93ac-6276ceaf16d7" (UID: "c5bff73f-3f9e-4313-93ac-6276ceaf16d7"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:16:21 crc kubenswrapper[4955]: I0217 13:16:21.918764 4955 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c5bff73f-3f9e-4313-93ac-6276ceaf16d7-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:16:21 crc kubenswrapper[4955]: I0217 13:16:21.918869 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2dtxq\" (UniqueName: \"kubernetes.io/projected/c5bff73f-3f9e-4313-93ac-6276ceaf16d7-kube-api-access-2dtxq\") on node \"crc\" DevicePath \"\"" Feb 17 13:16:21 crc kubenswrapper[4955]: I0217 13:16:21.918898 4955 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c5bff73f-3f9e-4313-93ac-6276ceaf16d7-util\") on node \"crc\" DevicePath \"\"" Feb 17 13:16:22 crc kubenswrapper[4955]: I0217 13:16:22.309597 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j" event={"ID":"c5bff73f-3f9e-4313-93ac-6276ceaf16d7","Type":"ContainerDied","Data":"93a724356d14abbcd991a8479885bc5ea4bf963df1420b6a51794841ee035caa"} Feb 17 13:16:22 crc kubenswrapper[4955]: I0217 13:16:22.309688 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="93a724356d14abbcd991a8479885bc5ea4bf963df1420b6a51794841ee035caa" Feb 17 13:16:22 crc kubenswrapper[4955]: I0217 13:16:22.309876 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j" Feb 17 13:16:24 crc kubenswrapper[4955]: I0217 13:16:24.543728 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-694c9596b7-j9h2x"] Feb 17 13:16:24 crc kubenswrapper[4955]: E0217 13:16:24.544167 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5bff73f-3f9e-4313-93ac-6276ceaf16d7" containerName="util" Feb 17 13:16:24 crc kubenswrapper[4955]: I0217 13:16:24.544179 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5bff73f-3f9e-4313-93ac-6276ceaf16d7" containerName="util" Feb 17 13:16:24 crc kubenswrapper[4955]: E0217 13:16:24.544188 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5bff73f-3f9e-4313-93ac-6276ceaf16d7" containerName="pull" Feb 17 13:16:24 crc kubenswrapper[4955]: I0217 13:16:24.544194 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5bff73f-3f9e-4313-93ac-6276ceaf16d7" containerName="pull" Feb 17 13:16:24 crc kubenswrapper[4955]: E0217 13:16:24.544202 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5bff73f-3f9e-4313-93ac-6276ceaf16d7" containerName="extract" Feb 17 13:16:24 crc kubenswrapper[4955]: I0217 13:16:24.544208 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5bff73f-3f9e-4313-93ac-6276ceaf16d7" containerName="extract" Feb 17 13:16:24 crc kubenswrapper[4955]: I0217 13:16:24.544318 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5bff73f-3f9e-4313-93ac-6276ceaf16d7" containerName="extract" Feb 17 13:16:24 crc kubenswrapper[4955]: I0217 13:16:24.544650 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-694c9596b7-j9h2x" Feb 17 13:16:24 crc kubenswrapper[4955]: I0217 13:16:24.545839 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Feb 17 13:16:24 crc kubenswrapper[4955]: I0217 13:16:24.546435 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-rw9qt" Feb 17 13:16:24 crc kubenswrapper[4955]: I0217 13:16:24.546583 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Feb 17 13:16:24 crc kubenswrapper[4955]: I0217 13:16:24.556545 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-694c9596b7-j9h2x"] Feb 17 13:16:24 crc kubenswrapper[4955]: I0217 13:16:24.655461 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bb2x\" (UniqueName: \"kubernetes.io/projected/4bd50965-e917-4e34-80b2-4e54e1d67feb-kube-api-access-4bb2x\") pod \"nmstate-operator-694c9596b7-j9h2x\" (UID: \"4bd50965-e917-4e34-80b2-4e54e1d67feb\") " pod="openshift-nmstate/nmstate-operator-694c9596b7-j9h2x" Feb 17 13:16:24 crc kubenswrapper[4955]: I0217 13:16:24.756977 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bb2x\" (UniqueName: \"kubernetes.io/projected/4bd50965-e917-4e34-80b2-4e54e1d67feb-kube-api-access-4bb2x\") pod \"nmstate-operator-694c9596b7-j9h2x\" (UID: \"4bd50965-e917-4e34-80b2-4e54e1d67feb\") " pod="openshift-nmstate/nmstate-operator-694c9596b7-j9h2x" Feb 17 13:16:24 crc kubenswrapper[4955]: I0217 13:16:24.775915 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bb2x\" (UniqueName: \"kubernetes.io/projected/4bd50965-e917-4e34-80b2-4e54e1d67feb-kube-api-access-4bb2x\") pod \"nmstate-operator-694c9596b7-j9h2x\" (UID: \"4bd50965-e917-4e34-80b2-4e54e1d67feb\") " pod="openshift-nmstate/nmstate-operator-694c9596b7-j9h2x" Feb 17 13:16:24 crc kubenswrapper[4955]: I0217 13:16:24.858562 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-694c9596b7-j9h2x" Feb 17 13:16:25 crc kubenswrapper[4955]: I0217 13:16:25.169631 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-694c9596b7-j9h2x"] Feb 17 13:16:25 crc kubenswrapper[4955]: W0217 13:16:25.188739 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4bd50965_e917_4e34_80b2_4e54e1d67feb.slice/crio-6388836bf32af5c191ff08d6d0f691c593559209e4bc2129d8ddc5a6d63f1bfa WatchSource:0}: Error finding container 6388836bf32af5c191ff08d6d0f691c593559209e4bc2129d8ddc5a6d63f1bfa: Status 404 returned error can't find the container with id 6388836bf32af5c191ff08d6d0f691c593559209e4bc2129d8ddc5a6d63f1bfa Feb 17 13:16:25 crc kubenswrapper[4955]: I0217 13:16:25.326576 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-694c9596b7-j9h2x" event={"ID":"4bd50965-e917-4e34-80b2-4e54e1d67feb","Type":"ContainerStarted","Data":"6388836bf32af5c191ff08d6d0f691c593559209e4bc2129d8ddc5a6d63f1bfa"} Feb 17 13:16:29 crc kubenswrapper[4955]: I0217 13:16:29.349621 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-694c9596b7-j9h2x" event={"ID":"4bd50965-e917-4e34-80b2-4e54e1d67feb","Type":"ContainerStarted","Data":"accab7422f790f9b911caeeb0744b2bee365c6d1bfa3c80581937e6add7e1809"} Feb 17 13:16:29 crc kubenswrapper[4955]: I0217 13:16:29.364927 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-694c9596b7-j9h2x" podStartSLOduration=2.026662036 podStartE2EDuration="5.364905602s" podCreationTimestamp="2026-02-17 13:16:24 +0000 UTC" firstStartedPulling="2026-02-17 13:16:25.194922037 +0000 UTC m=+723.717651590" lastFinishedPulling="2026-02-17 13:16:28.533165613 +0000 UTC m=+727.055895156" observedRunningTime="2026-02-17 13:16:29.362210684 +0000 UTC m=+727.884940297" watchObservedRunningTime="2026-02-17 13:16:29.364905602 +0000 UTC m=+727.887635145" Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.575146 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.575380 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.661850 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-58c85c668d-b4ngx"] Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.663150 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58c85c668d-b4ngx" Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.665240 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-nph5b" Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.681416 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58c85c668d-b4ngx"] Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.693388 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-866bcb46dc-d7hqh"] Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.694279 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-d7hqh" Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.695743 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.704525 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-9vdw4"] Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.705413 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-9vdw4" Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.722374 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-866bcb46dc-d7hqh"] Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.792846 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqwmm\" (UniqueName: \"kubernetes.io/projected/3aca323f-af5c-452c-8a4e-8c2b55e39dd5-kube-api-access-jqwmm\") pod \"nmstate-webhook-866bcb46dc-d7hqh\" (UID: \"3aca323f-af5c-452c-8a4e-8c2b55e39dd5\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-d7hqh" Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.792930 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zj46v\" (UniqueName: \"kubernetes.io/projected/bcf404eb-4186-40c9-ba70-768b78c8a60d-kube-api-access-zj46v\") pod \"nmstate-metrics-58c85c668d-b4ngx\" (UID: \"bcf404eb-4186-40c9-ba70-768b78c8a60d\") " pod="openshift-nmstate/nmstate-metrics-58c85c668d-b4ngx" Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.792954 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/3aca323f-af5c-452c-8a4e-8c2b55e39dd5-tls-key-pair\") pod \"nmstate-webhook-866bcb46dc-d7hqh\" (UID: \"3aca323f-af5c-452c-8a4e-8c2b55e39dd5\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-d7hqh" Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.806318 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5c78fc5d65-bdkvr"] Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.806921 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-bdkvr" Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.819461 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-pgxxk" Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.819518 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.820409 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5c78fc5d65-bdkvr"] Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.824427 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.894063 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/bb2572e6-ed65-4a6c-a3ca-f72c6cb4b5c0-nmstate-lock\") pod \"nmstate-handler-9vdw4\" (UID: \"bb2572e6-ed65-4a6c-a3ca-f72c6cb4b5c0\") " pod="openshift-nmstate/nmstate-handler-9vdw4" Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.894108 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zj46v\" (UniqueName: \"kubernetes.io/projected/bcf404eb-4186-40c9-ba70-768b78c8a60d-kube-api-access-zj46v\") pod \"nmstate-metrics-58c85c668d-b4ngx\" (UID: \"bcf404eb-4186-40c9-ba70-768b78c8a60d\") " pod="openshift-nmstate/nmstate-metrics-58c85c668d-b4ngx" Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.894130 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/3aca323f-af5c-452c-8a4e-8c2b55e39dd5-tls-key-pair\") pod \"nmstate-webhook-866bcb46dc-d7hqh\" (UID: \"3aca323f-af5c-452c-8a4e-8c2b55e39dd5\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-d7hqh" Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.894148 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l976l\" (UniqueName: \"kubernetes.io/projected/bb2572e6-ed65-4a6c-a3ca-f72c6cb4b5c0-kube-api-access-l976l\") pod \"nmstate-handler-9vdw4\" (UID: \"bb2572e6-ed65-4a6c-a3ca-f72c6cb4b5c0\") " pod="openshift-nmstate/nmstate-handler-9vdw4" Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.894185 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/bb2572e6-ed65-4a6c-a3ca-f72c6cb4b5c0-ovs-socket\") pod \"nmstate-handler-9vdw4\" (UID: \"bb2572e6-ed65-4a6c-a3ca-f72c6cb4b5c0\") " pod="openshift-nmstate/nmstate-handler-9vdw4" Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.894200 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/bb2572e6-ed65-4a6c-a3ca-f72c6cb4b5c0-dbus-socket\") pod \"nmstate-handler-9vdw4\" (UID: \"bb2572e6-ed65-4a6c-a3ca-f72c6cb4b5c0\") " pod="openshift-nmstate/nmstate-handler-9vdw4" Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.894215 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqwmm\" (UniqueName: \"kubernetes.io/projected/3aca323f-af5c-452c-8a4e-8c2b55e39dd5-kube-api-access-jqwmm\") pod \"nmstate-webhook-866bcb46dc-d7hqh\" (UID: \"3aca323f-af5c-452c-8a4e-8c2b55e39dd5\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-d7hqh" Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.914823 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/3aca323f-af5c-452c-8a4e-8c2b55e39dd5-tls-key-pair\") pod \"nmstate-webhook-866bcb46dc-d7hqh\" (UID: \"3aca323f-af5c-452c-8a4e-8c2b55e39dd5\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-d7hqh" Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.921327 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zj46v\" (UniqueName: \"kubernetes.io/projected/bcf404eb-4186-40c9-ba70-768b78c8a60d-kube-api-access-zj46v\") pod \"nmstate-metrics-58c85c668d-b4ngx\" (UID: \"bcf404eb-4186-40c9-ba70-768b78c8a60d\") " pod="openshift-nmstate/nmstate-metrics-58c85c668d-b4ngx" Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.922018 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqwmm\" (UniqueName: \"kubernetes.io/projected/3aca323f-af5c-452c-8a4e-8c2b55e39dd5-kube-api-access-jqwmm\") pod \"nmstate-webhook-866bcb46dc-d7hqh\" (UID: \"3aca323f-af5c-452c-8a4e-8c2b55e39dd5\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-d7hqh" Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.986651 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-659cdcfdb9-jbnw5"] Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.987325 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-659cdcfdb9-jbnw5" Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.989520 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58c85c668d-b4ngx" Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.995322 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/0afd3165-a15b-406f-88fb-4ce82ab17355-plugin-serving-cert\") pod \"nmstate-console-plugin-5c78fc5d65-bdkvr\" (UID: \"0afd3165-a15b-406f-88fb-4ce82ab17355\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-bdkvr" Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.995366 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/bb2572e6-ed65-4a6c-a3ca-f72c6cb4b5c0-nmstate-lock\") pod \"nmstate-handler-9vdw4\" (UID: \"bb2572e6-ed65-4a6c-a3ca-f72c6cb4b5c0\") " pod="openshift-nmstate/nmstate-handler-9vdw4" Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.995387 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/0afd3165-a15b-406f-88fb-4ce82ab17355-nginx-conf\") pod \"nmstate-console-plugin-5c78fc5d65-bdkvr\" (UID: \"0afd3165-a15b-406f-88fb-4ce82ab17355\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-bdkvr" Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.995412 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l976l\" (UniqueName: \"kubernetes.io/projected/bb2572e6-ed65-4a6c-a3ca-f72c6cb4b5c0-kube-api-access-l976l\") pod \"nmstate-handler-9vdw4\" (UID: \"bb2572e6-ed65-4a6c-a3ca-f72c6cb4b5c0\") " pod="openshift-nmstate/nmstate-handler-9vdw4" Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.995498 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m28w2\" (UniqueName: \"kubernetes.io/projected/0afd3165-a15b-406f-88fb-4ce82ab17355-kube-api-access-m28w2\") pod \"nmstate-console-plugin-5c78fc5d65-bdkvr\" (UID: \"0afd3165-a15b-406f-88fb-4ce82ab17355\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-bdkvr" Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.995502 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/bb2572e6-ed65-4a6c-a3ca-f72c6cb4b5c0-nmstate-lock\") pod \"nmstate-handler-9vdw4\" (UID: \"bb2572e6-ed65-4a6c-a3ca-f72c6cb4b5c0\") " pod="openshift-nmstate/nmstate-handler-9vdw4" Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.995609 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/bb2572e6-ed65-4a6c-a3ca-f72c6cb4b5c0-ovs-socket\") pod \"nmstate-handler-9vdw4\" (UID: \"bb2572e6-ed65-4a6c-a3ca-f72c6cb4b5c0\") " pod="openshift-nmstate/nmstate-handler-9vdw4" Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.995636 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/bb2572e6-ed65-4a6c-a3ca-f72c6cb4b5c0-dbus-socket\") pod \"nmstate-handler-9vdw4\" (UID: \"bb2572e6-ed65-4a6c-a3ca-f72c6cb4b5c0\") " pod="openshift-nmstate/nmstate-handler-9vdw4" Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.995656 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/bb2572e6-ed65-4a6c-a3ca-f72c6cb4b5c0-ovs-socket\") pod \"nmstate-handler-9vdw4\" (UID: \"bb2572e6-ed65-4a6c-a3ca-f72c6cb4b5c0\") " pod="openshift-nmstate/nmstate-handler-9vdw4" Feb 17 13:16:34 crc kubenswrapper[4955]: I0217 13:16:34.995888 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/bb2572e6-ed65-4a6c-a3ca-f72c6cb4b5c0-dbus-socket\") pod \"nmstate-handler-9vdw4\" (UID: \"bb2572e6-ed65-4a6c-a3ca-f72c6cb4b5c0\") " pod="openshift-nmstate/nmstate-handler-9vdw4" Feb 17 13:16:35 crc kubenswrapper[4955]: I0217 13:16:35.003733 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-659cdcfdb9-jbnw5"] Feb 17 13:16:35 crc kubenswrapper[4955]: I0217 13:16:35.023087 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-d7hqh" Feb 17 13:16:35 crc kubenswrapper[4955]: I0217 13:16:35.024042 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l976l\" (UniqueName: \"kubernetes.io/projected/bb2572e6-ed65-4a6c-a3ca-f72c6cb4b5c0-kube-api-access-l976l\") pod \"nmstate-handler-9vdw4\" (UID: \"bb2572e6-ed65-4a6c-a3ca-f72c6cb4b5c0\") " pod="openshift-nmstate/nmstate-handler-9vdw4" Feb 17 13:16:35 crc kubenswrapper[4955]: I0217 13:16:35.033065 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-9vdw4" Feb 17 13:16:35 crc kubenswrapper[4955]: I0217 13:16:35.097253 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/0afd3165-a15b-406f-88fb-4ce82ab17355-nginx-conf\") pod \"nmstate-console-plugin-5c78fc5d65-bdkvr\" (UID: \"0afd3165-a15b-406f-88fb-4ce82ab17355\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-bdkvr" Feb 17 13:16:35 crc kubenswrapper[4955]: I0217 13:16:35.097552 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d24e34f4-6873-40ce-abf1-3d1786fb87e5-console-serving-cert\") pod \"console-659cdcfdb9-jbnw5\" (UID: \"d24e34f4-6873-40ce-abf1-3d1786fb87e5\") " pod="openshift-console/console-659cdcfdb9-jbnw5" Feb 17 13:16:35 crc kubenswrapper[4955]: I0217 13:16:35.097570 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d24e34f4-6873-40ce-abf1-3d1786fb87e5-service-ca\") pod \"console-659cdcfdb9-jbnw5\" (UID: \"d24e34f4-6873-40ce-abf1-3d1786fb87e5\") " pod="openshift-console/console-659cdcfdb9-jbnw5" Feb 17 13:16:35 crc kubenswrapper[4955]: I0217 13:16:35.097618 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m28w2\" (UniqueName: \"kubernetes.io/projected/0afd3165-a15b-406f-88fb-4ce82ab17355-kube-api-access-m28w2\") pod \"nmstate-console-plugin-5c78fc5d65-bdkvr\" (UID: \"0afd3165-a15b-406f-88fb-4ce82ab17355\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-bdkvr" Feb 17 13:16:35 crc kubenswrapper[4955]: I0217 13:16:35.097640 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d24e34f4-6873-40ce-abf1-3d1786fb87e5-console-oauth-config\") pod \"console-659cdcfdb9-jbnw5\" (UID: \"d24e34f4-6873-40ce-abf1-3d1786fb87e5\") " pod="openshift-console/console-659cdcfdb9-jbnw5" Feb 17 13:16:35 crc kubenswrapper[4955]: I0217 13:16:35.097676 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99k7p\" (UniqueName: \"kubernetes.io/projected/d24e34f4-6873-40ce-abf1-3d1786fb87e5-kube-api-access-99k7p\") pod \"console-659cdcfdb9-jbnw5\" (UID: \"d24e34f4-6873-40ce-abf1-3d1786fb87e5\") " pod="openshift-console/console-659cdcfdb9-jbnw5" Feb 17 13:16:35 crc kubenswrapper[4955]: I0217 13:16:35.097716 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d24e34f4-6873-40ce-abf1-3d1786fb87e5-trusted-ca-bundle\") pod \"console-659cdcfdb9-jbnw5\" (UID: \"d24e34f4-6873-40ce-abf1-3d1786fb87e5\") " pod="openshift-console/console-659cdcfdb9-jbnw5" Feb 17 13:16:35 crc kubenswrapper[4955]: I0217 13:16:35.097767 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d24e34f4-6873-40ce-abf1-3d1786fb87e5-oauth-serving-cert\") pod \"console-659cdcfdb9-jbnw5\" (UID: \"d24e34f4-6873-40ce-abf1-3d1786fb87e5\") " pod="openshift-console/console-659cdcfdb9-jbnw5" Feb 17 13:16:35 crc kubenswrapper[4955]: I0217 13:16:35.097813 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/0afd3165-a15b-406f-88fb-4ce82ab17355-plugin-serving-cert\") pod \"nmstate-console-plugin-5c78fc5d65-bdkvr\" (UID: \"0afd3165-a15b-406f-88fb-4ce82ab17355\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-bdkvr" Feb 17 13:16:35 crc kubenswrapper[4955]: I0217 13:16:35.097830 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d24e34f4-6873-40ce-abf1-3d1786fb87e5-console-config\") pod \"console-659cdcfdb9-jbnw5\" (UID: \"d24e34f4-6873-40ce-abf1-3d1786fb87e5\") " pod="openshift-console/console-659cdcfdb9-jbnw5" Feb 17 13:16:35 crc kubenswrapper[4955]: I0217 13:16:35.100610 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/0afd3165-a15b-406f-88fb-4ce82ab17355-nginx-conf\") pod \"nmstate-console-plugin-5c78fc5d65-bdkvr\" (UID: \"0afd3165-a15b-406f-88fb-4ce82ab17355\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-bdkvr" Feb 17 13:16:35 crc kubenswrapper[4955]: I0217 13:16:35.104357 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/0afd3165-a15b-406f-88fb-4ce82ab17355-plugin-serving-cert\") pod \"nmstate-console-plugin-5c78fc5d65-bdkvr\" (UID: \"0afd3165-a15b-406f-88fb-4ce82ab17355\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-bdkvr" Feb 17 13:16:35 crc kubenswrapper[4955]: I0217 13:16:35.117979 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m28w2\" (UniqueName: \"kubernetes.io/projected/0afd3165-a15b-406f-88fb-4ce82ab17355-kube-api-access-m28w2\") pod \"nmstate-console-plugin-5c78fc5d65-bdkvr\" (UID: \"0afd3165-a15b-406f-88fb-4ce82ab17355\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-bdkvr" Feb 17 13:16:35 crc kubenswrapper[4955]: I0217 13:16:35.152144 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-bdkvr" Feb 17 13:16:35 crc kubenswrapper[4955]: I0217 13:16:35.192102 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58c85c668d-b4ngx"] Feb 17 13:16:35 crc kubenswrapper[4955]: W0217 13:16:35.197816 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbcf404eb_4186_40c9_ba70_768b78c8a60d.slice/crio-14449ba53827abd8681d3fd8931610874780f6929326040f70f845df3128600e WatchSource:0}: Error finding container 14449ba53827abd8681d3fd8931610874780f6929326040f70f845df3128600e: Status 404 returned error can't find the container with id 14449ba53827abd8681d3fd8931610874780f6929326040f70f845df3128600e Feb 17 13:16:35 crc kubenswrapper[4955]: I0217 13:16:35.199366 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d24e34f4-6873-40ce-abf1-3d1786fb87e5-console-config\") pod \"console-659cdcfdb9-jbnw5\" (UID: \"d24e34f4-6873-40ce-abf1-3d1786fb87e5\") " pod="openshift-console/console-659cdcfdb9-jbnw5" Feb 17 13:16:35 crc kubenswrapper[4955]: I0217 13:16:35.199443 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d24e34f4-6873-40ce-abf1-3d1786fb87e5-console-serving-cert\") pod \"console-659cdcfdb9-jbnw5\" (UID: \"d24e34f4-6873-40ce-abf1-3d1786fb87e5\") " pod="openshift-console/console-659cdcfdb9-jbnw5" Feb 17 13:16:35 crc kubenswrapper[4955]: I0217 13:16:35.199465 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d24e34f4-6873-40ce-abf1-3d1786fb87e5-service-ca\") pod \"console-659cdcfdb9-jbnw5\" (UID: \"d24e34f4-6873-40ce-abf1-3d1786fb87e5\") " pod="openshift-console/console-659cdcfdb9-jbnw5" Feb 17 13:16:35 crc kubenswrapper[4955]: I0217 13:16:35.199511 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d24e34f4-6873-40ce-abf1-3d1786fb87e5-console-oauth-config\") pod \"console-659cdcfdb9-jbnw5\" (UID: \"d24e34f4-6873-40ce-abf1-3d1786fb87e5\") " pod="openshift-console/console-659cdcfdb9-jbnw5" Feb 17 13:16:35 crc kubenswrapper[4955]: I0217 13:16:35.199532 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99k7p\" (UniqueName: \"kubernetes.io/projected/d24e34f4-6873-40ce-abf1-3d1786fb87e5-kube-api-access-99k7p\") pod \"console-659cdcfdb9-jbnw5\" (UID: \"d24e34f4-6873-40ce-abf1-3d1786fb87e5\") " pod="openshift-console/console-659cdcfdb9-jbnw5" Feb 17 13:16:35 crc kubenswrapper[4955]: I0217 13:16:35.199590 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d24e34f4-6873-40ce-abf1-3d1786fb87e5-trusted-ca-bundle\") pod \"console-659cdcfdb9-jbnw5\" (UID: \"d24e34f4-6873-40ce-abf1-3d1786fb87e5\") " pod="openshift-console/console-659cdcfdb9-jbnw5" Feb 17 13:16:35 crc kubenswrapper[4955]: I0217 13:16:35.199655 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d24e34f4-6873-40ce-abf1-3d1786fb87e5-oauth-serving-cert\") pod \"console-659cdcfdb9-jbnw5\" (UID: \"d24e34f4-6873-40ce-abf1-3d1786fb87e5\") " pod="openshift-console/console-659cdcfdb9-jbnw5" Feb 17 13:16:35 crc kubenswrapper[4955]: I0217 13:16:35.200563 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d24e34f4-6873-40ce-abf1-3d1786fb87e5-oauth-serving-cert\") pod \"console-659cdcfdb9-jbnw5\" (UID: \"d24e34f4-6873-40ce-abf1-3d1786fb87e5\") " pod="openshift-console/console-659cdcfdb9-jbnw5" Feb 17 13:16:35 crc kubenswrapper[4955]: I0217 13:16:35.200868 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d24e34f4-6873-40ce-abf1-3d1786fb87e5-console-config\") pod \"console-659cdcfdb9-jbnw5\" (UID: \"d24e34f4-6873-40ce-abf1-3d1786fb87e5\") " pod="openshift-console/console-659cdcfdb9-jbnw5" Feb 17 13:16:35 crc kubenswrapper[4955]: I0217 13:16:35.201122 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d24e34f4-6873-40ce-abf1-3d1786fb87e5-service-ca\") pod \"console-659cdcfdb9-jbnw5\" (UID: \"d24e34f4-6873-40ce-abf1-3d1786fb87e5\") " pod="openshift-console/console-659cdcfdb9-jbnw5" Feb 17 13:16:35 crc kubenswrapper[4955]: I0217 13:16:35.202125 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d24e34f4-6873-40ce-abf1-3d1786fb87e5-trusted-ca-bundle\") pod \"console-659cdcfdb9-jbnw5\" (UID: \"d24e34f4-6873-40ce-abf1-3d1786fb87e5\") " pod="openshift-console/console-659cdcfdb9-jbnw5" Feb 17 13:16:35 crc kubenswrapper[4955]: I0217 13:16:35.202507 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d24e34f4-6873-40ce-abf1-3d1786fb87e5-console-serving-cert\") pod \"console-659cdcfdb9-jbnw5\" (UID: \"d24e34f4-6873-40ce-abf1-3d1786fb87e5\") " pod="openshift-console/console-659cdcfdb9-jbnw5" Feb 17 13:16:35 crc kubenswrapper[4955]: I0217 13:16:35.204599 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d24e34f4-6873-40ce-abf1-3d1786fb87e5-console-oauth-config\") pod \"console-659cdcfdb9-jbnw5\" (UID: \"d24e34f4-6873-40ce-abf1-3d1786fb87e5\") " pod="openshift-console/console-659cdcfdb9-jbnw5" Feb 17 13:16:35 crc kubenswrapper[4955]: I0217 13:16:35.218703 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99k7p\" (UniqueName: \"kubernetes.io/projected/d24e34f4-6873-40ce-abf1-3d1786fb87e5-kube-api-access-99k7p\") pod \"console-659cdcfdb9-jbnw5\" (UID: \"d24e34f4-6873-40ce-abf1-3d1786fb87e5\") " pod="openshift-console/console-659cdcfdb9-jbnw5" Feb 17 13:16:35 crc kubenswrapper[4955]: I0217 13:16:35.243351 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-866bcb46dc-d7hqh"] Feb 17 13:16:35 crc kubenswrapper[4955]: W0217 13:16:35.247537 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3aca323f_af5c_452c_8a4e_8c2b55e39dd5.slice/crio-e4a850628fd78bfc6b042459e03fe1665598c76b9d328ff662e11422873e8280 WatchSource:0}: Error finding container e4a850628fd78bfc6b042459e03fe1665598c76b9d328ff662e11422873e8280: Status 404 returned error can't find the container with id e4a850628fd78bfc6b042459e03fe1665598c76b9d328ff662e11422873e8280 Feb 17 13:16:35 crc kubenswrapper[4955]: I0217 13:16:35.338851 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5c78fc5d65-bdkvr"] Feb 17 13:16:35 crc kubenswrapper[4955]: W0217 13:16:35.344222 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0afd3165_a15b_406f_88fb_4ce82ab17355.slice/crio-3e87fc086055ce3e735e27255dab53cb31da268f21b742aa92d6f40e1ed33d08 WatchSource:0}: Error finding container 3e87fc086055ce3e735e27255dab53cb31da268f21b742aa92d6f40e1ed33d08: Status 404 returned error can't find the container with id 3e87fc086055ce3e735e27255dab53cb31da268f21b742aa92d6f40e1ed33d08 Feb 17 13:16:35 crc kubenswrapper[4955]: I0217 13:16:35.366921 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-659cdcfdb9-jbnw5" Feb 17 13:16:35 crc kubenswrapper[4955]: I0217 13:16:35.390341 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-d7hqh" event={"ID":"3aca323f-af5c-452c-8a4e-8c2b55e39dd5","Type":"ContainerStarted","Data":"e4a850628fd78bfc6b042459e03fe1665598c76b9d328ff662e11422873e8280"} Feb 17 13:16:35 crc kubenswrapper[4955]: I0217 13:16:35.391804 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58c85c668d-b4ngx" event={"ID":"bcf404eb-4186-40c9-ba70-768b78c8a60d","Type":"ContainerStarted","Data":"14449ba53827abd8681d3fd8931610874780f6929326040f70f845df3128600e"} Feb 17 13:16:35 crc kubenswrapper[4955]: I0217 13:16:35.393518 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-9vdw4" event={"ID":"bb2572e6-ed65-4a6c-a3ca-f72c6cb4b5c0","Type":"ContainerStarted","Data":"8d228b2cbf5ddb030534d751b9c7e62a71ce6223ec74cae5ca334a1cf9d5d894"} Feb 17 13:16:35 crc kubenswrapper[4955]: I0217 13:16:35.394599 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-bdkvr" event={"ID":"0afd3165-a15b-406f-88fb-4ce82ab17355","Type":"ContainerStarted","Data":"3e87fc086055ce3e735e27255dab53cb31da268f21b742aa92d6f40e1ed33d08"} Feb 17 13:16:35 crc kubenswrapper[4955]: I0217 13:16:35.599824 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-659cdcfdb9-jbnw5"] Feb 17 13:16:35 crc kubenswrapper[4955]: W0217 13:16:35.606213 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd24e34f4_6873_40ce_abf1_3d1786fb87e5.slice/crio-2acc44498ef967d5009e345fcd185e201780089da641f19ceb416e79edabdfe2 WatchSource:0}: Error finding container 2acc44498ef967d5009e345fcd185e201780089da641f19ceb416e79edabdfe2: Status 404 returned error can't find the container with id 2acc44498ef967d5009e345fcd185e201780089da641f19ceb416e79edabdfe2 Feb 17 13:16:36 crc kubenswrapper[4955]: I0217 13:16:36.404351 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-659cdcfdb9-jbnw5" event={"ID":"d24e34f4-6873-40ce-abf1-3d1786fb87e5","Type":"ContainerStarted","Data":"15a9695903a55bbb00f14a2d9d1de8c08a0e420832689a51f2a9ae3520aa166b"} Feb 17 13:16:36 crc kubenswrapper[4955]: I0217 13:16:36.404905 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-659cdcfdb9-jbnw5" event={"ID":"d24e34f4-6873-40ce-abf1-3d1786fb87e5","Type":"ContainerStarted","Data":"2acc44498ef967d5009e345fcd185e201780089da641f19ceb416e79edabdfe2"} Feb 17 13:16:36 crc kubenswrapper[4955]: I0217 13:16:36.438274 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-659cdcfdb9-jbnw5" podStartSLOduration=2.438248404 podStartE2EDuration="2.438248404s" podCreationTimestamp="2026-02-17 13:16:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:16:36.431139097 +0000 UTC m=+734.953868680" watchObservedRunningTime="2026-02-17 13:16:36.438248404 +0000 UTC m=+734.960977957" Feb 17 13:16:38 crc kubenswrapper[4955]: I0217 13:16:38.416393 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-bdkvr" event={"ID":"0afd3165-a15b-406f-88fb-4ce82ab17355","Type":"ContainerStarted","Data":"79ba3d2840b811e790ea2b5d94f82312f001b9c9f3a9be4d5c6b46ffd715ea67"} Feb 17 13:16:38 crc kubenswrapper[4955]: I0217 13:16:38.428197 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-bdkvr" podStartSLOduration=2.325650536 podStartE2EDuration="4.428177463s" podCreationTimestamp="2026-02-17 13:16:34 +0000 UTC" firstStartedPulling="2026-02-17 13:16:35.347185262 +0000 UTC m=+733.869914805" lastFinishedPulling="2026-02-17 13:16:37.449712179 +0000 UTC m=+735.972441732" observedRunningTime="2026-02-17 13:16:38.427295447 +0000 UTC m=+736.950024990" watchObservedRunningTime="2026-02-17 13:16:38.428177463 +0000 UTC m=+736.950907016" Feb 17 13:16:39 crc kubenswrapper[4955]: I0217 13:16:39.423675 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-d7hqh" event={"ID":"3aca323f-af5c-452c-8a4e-8c2b55e39dd5","Type":"ContainerStarted","Data":"2e8e87162e60abb4905c5b9c7a49f713dfc3ff13bd3f0fa33acaaa07fdf09b93"} Feb 17 13:16:39 crc kubenswrapper[4955]: I0217 13:16:39.424153 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-d7hqh" Feb 17 13:16:39 crc kubenswrapper[4955]: I0217 13:16:39.426120 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58c85c668d-b4ngx" event={"ID":"bcf404eb-4186-40c9-ba70-768b78c8a60d","Type":"ContainerStarted","Data":"80135640d0a6ed86f6b4786f2b10f525e46f5d3f58492e328493bcc13d8eaf21"} Feb 17 13:16:39 crc kubenswrapper[4955]: I0217 13:16:39.427685 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-9vdw4" event={"ID":"bb2572e6-ed65-4a6c-a3ca-f72c6cb4b5c0","Type":"ContainerStarted","Data":"8251f28fd383a8832a91cf4845ede34a068f278516abee32fa258ce2cf72a112"} Feb 17 13:16:39 crc kubenswrapper[4955]: I0217 13:16:39.442104 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-d7hqh" podStartSLOduration=1.777553087 podStartE2EDuration="5.442085348s" podCreationTimestamp="2026-02-17 13:16:34 +0000 UTC" firstStartedPulling="2026-02-17 13:16:35.253931445 +0000 UTC m=+733.776660988" lastFinishedPulling="2026-02-17 13:16:38.918463666 +0000 UTC m=+737.441193249" observedRunningTime="2026-02-17 13:16:39.441042948 +0000 UTC m=+737.963772501" watchObservedRunningTime="2026-02-17 13:16:39.442085348 +0000 UTC m=+737.964814901" Feb 17 13:16:39 crc kubenswrapper[4955]: I0217 13:16:39.466109 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-9vdw4" podStartSLOduration=1.62347047 podStartE2EDuration="5.466086208s" podCreationTimestamp="2026-02-17 13:16:34 +0000 UTC" firstStartedPulling="2026-02-17 13:16:35.066733182 +0000 UTC m=+733.589462725" lastFinishedPulling="2026-02-17 13:16:38.90934888 +0000 UTC m=+737.432078463" observedRunningTime="2026-02-17 13:16:39.463544663 +0000 UTC m=+737.986274246" watchObservedRunningTime="2026-02-17 13:16:39.466086208 +0000 UTC m=+737.988815761" Feb 17 13:16:40 crc kubenswrapper[4955]: I0217 13:16:40.034063 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-9vdw4" Feb 17 13:16:43 crc kubenswrapper[4955]: I0217 13:16:43.459889 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58c85c668d-b4ngx" event={"ID":"bcf404eb-4186-40c9-ba70-768b78c8a60d","Type":"ContainerStarted","Data":"46d962efca3b69b16e61d30d4ced1f101a4e6658cce4ce69eca028f0008c590e"} Feb 17 13:16:43 crc kubenswrapper[4955]: I0217 13:16:43.504127 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-58c85c668d-b4ngx" podStartSLOduration=1.650278432 podStartE2EDuration="9.504091988s" podCreationTimestamp="2026-02-17 13:16:34 +0000 UTC" firstStartedPulling="2026-02-17 13:16:35.209050108 +0000 UTC m=+733.731779651" lastFinishedPulling="2026-02-17 13:16:43.062863644 +0000 UTC m=+741.585593207" observedRunningTime="2026-02-17 13:16:43.494959632 +0000 UTC m=+742.017689215" watchObservedRunningTime="2026-02-17 13:16:43.504091988 +0000 UTC m=+742.026821571" Feb 17 13:16:45 crc kubenswrapper[4955]: I0217 13:16:45.071212 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-9vdw4" Feb 17 13:16:45 crc kubenswrapper[4955]: I0217 13:16:45.368251 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-659cdcfdb9-jbnw5" Feb 17 13:16:45 crc kubenswrapper[4955]: I0217 13:16:45.369433 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-659cdcfdb9-jbnw5" Feb 17 13:16:45 crc kubenswrapper[4955]: I0217 13:16:45.376286 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-659cdcfdb9-jbnw5" Feb 17 13:16:45 crc kubenswrapper[4955]: I0217 13:16:45.482533 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-659cdcfdb9-jbnw5" Feb 17 13:16:45 crc kubenswrapper[4955]: I0217 13:16:45.563660 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-npb7n"] Feb 17 13:16:55 crc kubenswrapper[4955]: I0217 13:16:55.031236 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-d7hqh" Feb 17 13:16:58 crc kubenswrapper[4955]: I0217 13:16:58.274534 4955 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 17 13:17:03 crc kubenswrapper[4955]: I0217 13:17:03.149493 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wl2jr"] Feb 17 13:17:03 crc kubenswrapper[4955]: I0217 13:17:03.151041 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wl2jr" Feb 17 13:17:03 crc kubenswrapper[4955]: I0217 13:17:03.162023 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wl2jr"] Feb 17 13:17:03 crc kubenswrapper[4955]: I0217 13:17:03.330188 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d226a93-aca2-4e11-9a77-6159b3ca087b-utilities\") pod \"redhat-operators-wl2jr\" (UID: \"5d226a93-aca2-4e11-9a77-6159b3ca087b\") " pod="openshift-marketplace/redhat-operators-wl2jr" Feb 17 13:17:03 crc kubenswrapper[4955]: I0217 13:17:03.330348 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d226a93-aca2-4e11-9a77-6159b3ca087b-catalog-content\") pod \"redhat-operators-wl2jr\" (UID: \"5d226a93-aca2-4e11-9a77-6159b3ca087b\") " pod="openshift-marketplace/redhat-operators-wl2jr" Feb 17 13:17:03 crc kubenswrapper[4955]: I0217 13:17:03.330506 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78pvf\" (UniqueName: \"kubernetes.io/projected/5d226a93-aca2-4e11-9a77-6159b3ca087b-kube-api-access-78pvf\") pod \"redhat-operators-wl2jr\" (UID: \"5d226a93-aca2-4e11-9a77-6159b3ca087b\") " pod="openshift-marketplace/redhat-operators-wl2jr" Feb 17 13:17:03 crc kubenswrapper[4955]: I0217 13:17:03.431910 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d226a93-aca2-4e11-9a77-6159b3ca087b-utilities\") pod \"redhat-operators-wl2jr\" (UID: \"5d226a93-aca2-4e11-9a77-6159b3ca087b\") " pod="openshift-marketplace/redhat-operators-wl2jr" Feb 17 13:17:03 crc kubenswrapper[4955]: I0217 13:17:03.431977 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d226a93-aca2-4e11-9a77-6159b3ca087b-catalog-content\") pod \"redhat-operators-wl2jr\" (UID: \"5d226a93-aca2-4e11-9a77-6159b3ca087b\") " pod="openshift-marketplace/redhat-operators-wl2jr" Feb 17 13:17:03 crc kubenswrapper[4955]: I0217 13:17:03.432033 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78pvf\" (UniqueName: \"kubernetes.io/projected/5d226a93-aca2-4e11-9a77-6159b3ca087b-kube-api-access-78pvf\") pod \"redhat-operators-wl2jr\" (UID: \"5d226a93-aca2-4e11-9a77-6159b3ca087b\") " pod="openshift-marketplace/redhat-operators-wl2jr" Feb 17 13:17:03 crc kubenswrapper[4955]: I0217 13:17:03.432461 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d226a93-aca2-4e11-9a77-6159b3ca087b-utilities\") pod \"redhat-operators-wl2jr\" (UID: \"5d226a93-aca2-4e11-9a77-6159b3ca087b\") " pod="openshift-marketplace/redhat-operators-wl2jr" Feb 17 13:17:03 crc kubenswrapper[4955]: I0217 13:17:03.432767 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d226a93-aca2-4e11-9a77-6159b3ca087b-catalog-content\") pod \"redhat-operators-wl2jr\" (UID: \"5d226a93-aca2-4e11-9a77-6159b3ca087b\") " pod="openshift-marketplace/redhat-operators-wl2jr" Feb 17 13:17:03 crc kubenswrapper[4955]: I0217 13:17:03.452172 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78pvf\" (UniqueName: \"kubernetes.io/projected/5d226a93-aca2-4e11-9a77-6159b3ca087b-kube-api-access-78pvf\") pod \"redhat-operators-wl2jr\" (UID: \"5d226a93-aca2-4e11-9a77-6159b3ca087b\") " pod="openshift-marketplace/redhat-operators-wl2jr" Feb 17 13:17:03 crc kubenswrapper[4955]: I0217 13:17:03.495422 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wl2jr" Feb 17 13:17:03 crc kubenswrapper[4955]: I0217 13:17:03.931050 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wl2jr"] Feb 17 13:17:03 crc kubenswrapper[4955]: W0217 13:17:03.933754 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5d226a93_aca2_4e11_9a77_6159b3ca087b.slice/crio-b20d4d84ec4a7e251577e17661022d5029fbcb0ec17256d7394203b5673bfabb WatchSource:0}: Error finding container b20d4d84ec4a7e251577e17661022d5029fbcb0ec17256d7394203b5673bfabb: Status 404 returned error can't find the container with id b20d4d84ec4a7e251577e17661022d5029fbcb0ec17256d7394203b5673bfabb Feb 17 13:17:04 crc kubenswrapper[4955]: I0217 13:17:04.574532 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 13:17:04 crc kubenswrapper[4955]: I0217 13:17:04.574900 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 13:17:04 crc kubenswrapper[4955]: I0217 13:17:04.574957 4955 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" Feb 17 13:17:04 crc kubenswrapper[4955]: I0217 13:17:04.575582 4955 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"817a479207945cb4a0cb4caff307634a33fb718eab5f99bb1c4a37862d5010ac"} pod="openshift-machine-config-operator/machine-config-daemon-29qxq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 17 13:17:04 crc kubenswrapper[4955]: I0217 13:17:04.575632 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" containerID="cri-o://817a479207945cb4a0cb4caff307634a33fb718eab5f99bb1c4a37862d5010ac" gracePeriod=600 Feb 17 13:17:04 crc kubenswrapper[4955]: I0217 13:17:04.620606 4955 generic.go:334] "Generic (PLEG): container finished" podID="5d226a93-aca2-4e11-9a77-6159b3ca087b" containerID="e9c15e95a677567793bc31417f9c8a220a357d7ec072566d0d85c4d5f93acd4c" exitCode=0 Feb 17 13:17:04 crc kubenswrapper[4955]: I0217 13:17:04.620656 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wl2jr" event={"ID":"5d226a93-aca2-4e11-9a77-6159b3ca087b","Type":"ContainerDied","Data":"e9c15e95a677567793bc31417f9c8a220a357d7ec072566d0d85c4d5f93acd4c"} Feb 17 13:17:04 crc kubenswrapper[4955]: I0217 13:17:04.620684 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wl2jr" event={"ID":"5d226a93-aca2-4e11-9a77-6159b3ca087b","Type":"ContainerStarted","Data":"b20d4d84ec4a7e251577e17661022d5029fbcb0ec17256d7394203b5673bfabb"} Feb 17 13:17:05 crc kubenswrapper[4955]: I0217 13:17:05.633185 4955 generic.go:334] "Generic (PLEG): container finished" podID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerID="817a479207945cb4a0cb4caff307634a33fb718eab5f99bb1c4a37862d5010ac" exitCode=0 Feb 17 13:17:05 crc kubenswrapper[4955]: I0217 13:17:05.633233 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" event={"ID":"5bf36887-2a2b-4193-abdb-a15beaf46300","Type":"ContainerDied","Data":"817a479207945cb4a0cb4caff307634a33fb718eab5f99bb1c4a37862d5010ac"} Feb 17 13:17:05 crc kubenswrapper[4955]: I0217 13:17:05.633831 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" event={"ID":"5bf36887-2a2b-4193-abdb-a15beaf46300","Type":"ContainerStarted","Data":"a72972c52870687e44697a4fc3675d78bf2389f390e43593c993f04221b0f584"} Feb 17 13:17:05 crc kubenswrapper[4955]: I0217 13:17:05.633866 4955 scope.go:117] "RemoveContainer" containerID="ae1cb692a0145e1501a14dd21bdbee429e632fe371cafa0c41cac4fa359ae05a" Feb 17 13:17:06 crc kubenswrapper[4955]: I0217 13:17:06.645818 4955 generic.go:334] "Generic (PLEG): container finished" podID="5d226a93-aca2-4e11-9a77-6159b3ca087b" containerID="5f2d65d1b698a552cbe04bf7aee28ccafa6430c34fd28ae624c6d8eab6716e7c" exitCode=0 Feb 17 13:17:06 crc kubenswrapper[4955]: I0217 13:17:06.645869 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wl2jr" event={"ID":"5d226a93-aca2-4e11-9a77-6159b3ca087b","Type":"ContainerDied","Data":"5f2d65d1b698a552cbe04bf7aee28ccafa6430c34fd28ae624c6d8eab6716e7c"} Feb 17 13:17:07 crc kubenswrapper[4955]: I0217 13:17:07.653584 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wl2jr" event={"ID":"5d226a93-aca2-4e11-9a77-6159b3ca087b","Type":"ContainerStarted","Data":"7cafb2e453e05faf238c715aa28370a4d54bae35af8a5b5c05a1e630343ed317"} Feb 17 13:17:07 crc kubenswrapper[4955]: I0217 13:17:07.676648 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wl2jr" podStartSLOduration=1.988923928 podStartE2EDuration="4.676625244s" podCreationTimestamp="2026-02-17 13:17:03 +0000 UTC" firstStartedPulling="2026-02-17 13:17:04.621890746 +0000 UTC m=+763.144620289" lastFinishedPulling="2026-02-17 13:17:07.309592062 +0000 UTC m=+765.832321605" observedRunningTime="2026-02-17 13:17:07.672479643 +0000 UTC m=+766.195209226" watchObservedRunningTime="2026-02-17 13:17:07.676625244 +0000 UTC m=+766.199354827" Feb 17 13:17:09 crc kubenswrapper[4955]: I0217 13:17:09.519354 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jprkf"] Feb 17 13:17:09 crc kubenswrapper[4955]: I0217 13:17:09.521914 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jprkf" Feb 17 13:17:09 crc kubenswrapper[4955]: I0217 13:17:09.527560 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jprkf"] Feb 17 13:17:09 crc kubenswrapper[4955]: I0217 13:17:09.609744 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvrpg\" (UniqueName: \"kubernetes.io/projected/5dc80fd1-674b-4442-90f4-10da0796e4fc-kube-api-access-lvrpg\") pod \"certified-operators-jprkf\" (UID: \"5dc80fd1-674b-4442-90f4-10da0796e4fc\") " pod="openshift-marketplace/certified-operators-jprkf" Feb 17 13:17:09 crc kubenswrapper[4955]: I0217 13:17:09.609841 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5dc80fd1-674b-4442-90f4-10da0796e4fc-catalog-content\") pod \"certified-operators-jprkf\" (UID: \"5dc80fd1-674b-4442-90f4-10da0796e4fc\") " pod="openshift-marketplace/certified-operators-jprkf" Feb 17 13:17:09 crc kubenswrapper[4955]: I0217 13:17:09.610282 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5dc80fd1-674b-4442-90f4-10da0796e4fc-utilities\") pod \"certified-operators-jprkf\" (UID: \"5dc80fd1-674b-4442-90f4-10da0796e4fc\") " pod="openshift-marketplace/certified-operators-jprkf" Feb 17 13:17:09 crc kubenswrapper[4955]: I0217 13:17:09.711968 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvrpg\" (UniqueName: \"kubernetes.io/projected/5dc80fd1-674b-4442-90f4-10da0796e4fc-kube-api-access-lvrpg\") pod \"certified-operators-jprkf\" (UID: \"5dc80fd1-674b-4442-90f4-10da0796e4fc\") " pod="openshift-marketplace/certified-operators-jprkf" Feb 17 13:17:09 crc kubenswrapper[4955]: I0217 13:17:09.712049 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5dc80fd1-674b-4442-90f4-10da0796e4fc-catalog-content\") pod \"certified-operators-jprkf\" (UID: \"5dc80fd1-674b-4442-90f4-10da0796e4fc\") " pod="openshift-marketplace/certified-operators-jprkf" Feb 17 13:17:09 crc kubenswrapper[4955]: I0217 13:17:09.712129 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5dc80fd1-674b-4442-90f4-10da0796e4fc-utilities\") pod \"certified-operators-jprkf\" (UID: \"5dc80fd1-674b-4442-90f4-10da0796e4fc\") " pod="openshift-marketplace/certified-operators-jprkf" Feb 17 13:17:09 crc kubenswrapper[4955]: I0217 13:17:09.712755 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5dc80fd1-674b-4442-90f4-10da0796e4fc-utilities\") pod \"certified-operators-jprkf\" (UID: \"5dc80fd1-674b-4442-90f4-10da0796e4fc\") " pod="openshift-marketplace/certified-operators-jprkf" Feb 17 13:17:09 crc kubenswrapper[4955]: I0217 13:17:09.712855 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5dc80fd1-674b-4442-90f4-10da0796e4fc-catalog-content\") pod \"certified-operators-jprkf\" (UID: \"5dc80fd1-674b-4442-90f4-10da0796e4fc\") " pod="openshift-marketplace/certified-operators-jprkf" Feb 17 13:17:09 crc kubenswrapper[4955]: I0217 13:17:09.734172 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvrpg\" (UniqueName: \"kubernetes.io/projected/5dc80fd1-674b-4442-90f4-10da0796e4fc-kube-api-access-lvrpg\") pod \"certified-operators-jprkf\" (UID: \"5dc80fd1-674b-4442-90f4-10da0796e4fc\") " pod="openshift-marketplace/certified-operators-jprkf" Feb 17 13:17:09 crc kubenswrapper[4955]: I0217 13:17:09.850158 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jprkf" Feb 17 13:17:10 crc kubenswrapper[4955]: I0217 13:17:10.118389 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jprkf"] Feb 17 13:17:10 crc kubenswrapper[4955]: I0217 13:17:10.615665 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-npb7n" podUID="e690bd42-0ab1-4b3a-802b-4511117b6fe0" containerName="console" containerID="cri-o://541ff6f58738d8ceca7382578b38728152686198ff183271bc41d9f2d35e933f" gracePeriod=15 Feb 17 13:17:10 crc kubenswrapper[4955]: I0217 13:17:10.674467 4955 generic.go:334] "Generic (PLEG): container finished" podID="5dc80fd1-674b-4442-90f4-10da0796e4fc" containerID="1dd886d04cead8fe44ea298047debf672ae809651f0b21c88eab244186e68c6a" exitCode=0 Feb 17 13:17:10 crc kubenswrapper[4955]: I0217 13:17:10.674509 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jprkf" event={"ID":"5dc80fd1-674b-4442-90f4-10da0796e4fc","Type":"ContainerDied","Data":"1dd886d04cead8fe44ea298047debf672ae809651f0b21c88eab244186e68c6a"} Feb 17 13:17:10 crc kubenswrapper[4955]: I0217 13:17:10.674536 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jprkf" event={"ID":"5dc80fd1-674b-4442-90f4-10da0796e4fc","Type":"ContainerStarted","Data":"a8b31833842d7296995204acc31ff10586708908364cb6950dc2bd8b2a85372c"} Feb 17 13:17:10 crc kubenswrapper[4955]: I0217 13:17:10.965315 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-npb7n_e690bd42-0ab1-4b3a-802b-4511117b6fe0/console/0.log" Feb 17 13:17:10 crc kubenswrapper[4955]: I0217 13:17:10.965569 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-npb7n" Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.133008 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b9bjs\" (UniqueName: \"kubernetes.io/projected/e690bd42-0ab1-4b3a-802b-4511117b6fe0-kube-api-access-b9bjs\") pod \"e690bd42-0ab1-4b3a-802b-4511117b6fe0\" (UID: \"e690bd42-0ab1-4b3a-802b-4511117b6fe0\") " Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.133068 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e690bd42-0ab1-4b3a-802b-4511117b6fe0-console-config\") pod \"e690bd42-0ab1-4b3a-802b-4511117b6fe0\" (UID: \"e690bd42-0ab1-4b3a-802b-4511117b6fe0\") " Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.133121 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e690bd42-0ab1-4b3a-802b-4511117b6fe0-service-ca\") pod \"e690bd42-0ab1-4b3a-802b-4511117b6fe0\" (UID: \"e690bd42-0ab1-4b3a-802b-4511117b6fe0\") " Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.133152 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e690bd42-0ab1-4b3a-802b-4511117b6fe0-console-serving-cert\") pod \"e690bd42-0ab1-4b3a-802b-4511117b6fe0\" (UID: \"e690bd42-0ab1-4b3a-802b-4511117b6fe0\") " Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.133193 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e690bd42-0ab1-4b3a-802b-4511117b6fe0-oauth-serving-cert\") pod \"e690bd42-0ab1-4b3a-802b-4511117b6fe0\" (UID: \"e690bd42-0ab1-4b3a-802b-4511117b6fe0\") " Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.133221 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e690bd42-0ab1-4b3a-802b-4511117b6fe0-trusted-ca-bundle\") pod \"e690bd42-0ab1-4b3a-802b-4511117b6fe0\" (UID: \"e690bd42-0ab1-4b3a-802b-4511117b6fe0\") " Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.133238 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e690bd42-0ab1-4b3a-802b-4511117b6fe0-console-oauth-config\") pod \"e690bd42-0ab1-4b3a-802b-4511117b6fe0\" (UID: \"e690bd42-0ab1-4b3a-802b-4511117b6fe0\") " Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.134057 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e690bd42-0ab1-4b3a-802b-4511117b6fe0-console-config" (OuterVolumeSpecName: "console-config") pod "e690bd42-0ab1-4b3a-802b-4511117b6fe0" (UID: "e690bd42-0ab1-4b3a-802b-4511117b6fe0"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.134216 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e690bd42-0ab1-4b3a-802b-4511117b6fe0-service-ca" (OuterVolumeSpecName: "service-ca") pod "e690bd42-0ab1-4b3a-802b-4511117b6fe0" (UID: "e690bd42-0ab1-4b3a-802b-4511117b6fe0"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.134292 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e690bd42-0ab1-4b3a-802b-4511117b6fe0-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "e690bd42-0ab1-4b3a-802b-4511117b6fe0" (UID: "e690bd42-0ab1-4b3a-802b-4511117b6fe0"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.134226 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e690bd42-0ab1-4b3a-802b-4511117b6fe0-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "e690bd42-0ab1-4b3a-802b-4511117b6fe0" (UID: "e690bd42-0ab1-4b3a-802b-4511117b6fe0"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.141391 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e690bd42-0ab1-4b3a-802b-4511117b6fe0-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "e690bd42-0ab1-4b3a-802b-4511117b6fe0" (UID: "e690bd42-0ab1-4b3a-802b-4511117b6fe0"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.141760 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e690bd42-0ab1-4b3a-802b-4511117b6fe0-kube-api-access-b9bjs" (OuterVolumeSpecName: "kube-api-access-b9bjs") pod "e690bd42-0ab1-4b3a-802b-4511117b6fe0" (UID: "e690bd42-0ab1-4b3a-802b-4511117b6fe0"). InnerVolumeSpecName "kube-api-access-b9bjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.141839 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e690bd42-0ab1-4b3a-802b-4511117b6fe0-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "e690bd42-0ab1-4b3a-802b-4511117b6fe0" (UID: "e690bd42-0ab1-4b3a-802b-4511117b6fe0"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.160047 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l"] Feb 17 13:17:11 crc kubenswrapper[4955]: E0217 13:17:11.160374 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e690bd42-0ab1-4b3a-802b-4511117b6fe0" containerName="console" Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.160404 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="e690bd42-0ab1-4b3a-802b-4511117b6fe0" containerName="console" Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.160626 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="e690bd42-0ab1-4b3a-802b-4511117b6fe0" containerName="console" Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.161990 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l" Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.164071 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.178566 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l"] Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.234883 4955 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e690bd42-0ab1-4b3a-802b-4511117b6fe0-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.234912 4955 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e690bd42-0ab1-4b3a-802b-4511117b6fe0-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.234922 4955 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e690bd42-0ab1-4b3a-802b-4511117b6fe0-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.234930 4955 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e690bd42-0ab1-4b3a-802b-4511117b6fe0-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.234939 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b9bjs\" (UniqueName: \"kubernetes.io/projected/e690bd42-0ab1-4b3a-802b-4511117b6fe0-kube-api-access-b9bjs\") on node \"crc\" DevicePath \"\"" Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.234947 4955 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e690bd42-0ab1-4b3a-802b-4511117b6fe0-console-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.234957 4955 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e690bd42-0ab1-4b3a-802b-4511117b6fe0-service-ca\") on node \"crc\" DevicePath \"\"" Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.336286 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/102e3d48-f56c-4e7c-956e-23484bf45776-bundle\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l\" (UID: \"102e3d48-f56c-4e7c-956e-23484bf45776\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l" Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.336506 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5r8mb\" (UniqueName: \"kubernetes.io/projected/102e3d48-f56c-4e7c-956e-23484bf45776-kube-api-access-5r8mb\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l\" (UID: \"102e3d48-f56c-4e7c-956e-23484bf45776\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l" Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.336583 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/102e3d48-f56c-4e7c-956e-23484bf45776-util\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l\" (UID: \"102e3d48-f56c-4e7c-956e-23484bf45776\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l" Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.438095 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/102e3d48-f56c-4e7c-956e-23484bf45776-util\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l\" (UID: \"102e3d48-f56c-4e7c-956e-23484bf45776\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l" Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.438308 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/102e3d48-f56c-4e7c-956e-23484bf45776-bundle\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l\" (UID: \"102e3d48-f56c-4e7c-956e-23484bf45776\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l" Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.438442 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5r8mb\" (UniqueName: \"kubernetes.io/projected/102e3d48-f56c-4e7c-956e-23484bf45776-kube-api-access-5r8mb\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l\" (UID: \"102e3d48-f56c-4e7c-956e-23484bf45776\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l" Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.438771 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/102e3d48-f56c-4e7c-956e-23484bf45776-util\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l\" (UID: \"102e3d48-f56c-4e7c-956e-23484bf45776\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l" Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.438998 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/102e3d48-f56c-4e7c-956e-23484bf45776-bundle\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l\" (UID: \"102e3d48-f56c-4e7c-956e-23484bf45776\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l" Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.468343 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5r8mb\" (UniqueName: \"kubernetes.io/projected/102e3d48-f56c-4e7c-956e-23484bf45776-kube-api-access-5r8mb\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l\" (UID: \"102e3d48-f56c-4e7c-956e-23484bf45776\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l" Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.555182 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l" Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.685531 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-npb7n_e690bd42-0ab1-4b3a-802b-4511117b6fe0/console/0.log" Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.685592 4955 generic.go:334] "Generic (PLEG): container finished" podID="e690bd42-0ab1-4b3a-802b-4511117b6fe0" containerID="541ff6f58738d8ceca7382578b38728152686198ff183271bc41d9f2d35e933f" exitCode=2 Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.685630 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-npb7n" event={"ID":"e690bd42-0ab1-4b3a-802b-4511117b6fe0","Type":"ContainerDied","Data":"541ff6f58738d8ceca7382578b38728152686198ff183271bc41d9f2d35e933f"} Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.685672 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-npb7n" event={"ID":"e690bd42-0ab1-4b3a-802b-4511117b6fe0","Type":"ContainerDied","Data":"c0e9980ef890eb038f5e55419af044fc0ace9c568ae670bb96a43c23a77fb651"} Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.685697 4955 scope.go:117] "RemoveContainer" containerID="541ff6f58738d8ceca7382578b38728152686198ff183271bc41d9f2d35e933f" Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.685719 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-npb7n" Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.712647 4955 scope.go:117] "RemoveContainer" containerID="541ff6f58738d8ceca7382578b38728152686198ff183271bc41d9f2d35e933f" Feb 17 13:17:11 crc kubenswrapper[4955]: E0217 13:17:11.713896 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"541ff6f58738d8ceca7382578b38728152686198ff183271bc41d9f2d35e933f\": container with ID starting with 541ff6f58738d8ceca7382578b38728152686198ff183271bc41d9f2d35e933f not found: ID does not exist" containerID="541ff6f58738d8ceca7382578b38728152686198ff183271bc41d9f2d35e933f" Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.713958 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"541ff6f58738d8ceca7382578b38728152686198ff183271bc41d9f2d35e933f"} err="failed to get container status \"541ff6f58738d8ceca7382578b38728152686198ff183271bc41d9f2d35e933f\": rpc error: code = NotFound desc = could not find container \"541ff6f58738d8ceca7382578b38728152686198ff183271bc41d9f2d35e933f\": container with ID starting with 541ff6f58738d8ceca7382578b38728152686198ff183271bc41d9f2d35e933f not found: ID does not exist" Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.729054 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-npb7n"] Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.736822 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-npb7n"] Feb 17 13:17:11 crc kubenswrapper[4955]: I0217 13:17:11.895049 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l"] Feb 17 13:17:12 crc kubenswrapper[4955]: I0217 13:17:12.232164 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e690bd42-0ab1-4b3a-802b-4511117b6fe0" path="/var/lib/kubelet/pods/e690bd42-0ab1-4b3a-802b-4511117b6fe0/volumes" Feb 17 13:17:12 crc kubenswrapper[4955]: I0217 13:17:12.691492 4955 generic.go:334] "Generic (PLEG): container finished" podID="5dc80fd1-674b-4442-90f4-10da0796e4fc" containerID="3926052d617dc9ed98152ba808ac9c1020fe756329269f685ee109da9e4c11b1" exitCode=0 Feb 17 13:17:12 crc kubenswrapper[4955]: I0217 13:17:12.691552 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jprkf" event={"ID":"5dc80fd1-674b-4442-90f4-10da0796e4fc","Type":"ContainerDied","Data":"3926052d617dc9ed98152ba808ac9c1020fe756329269f685ee109da9e4c11b1"} Feb 17 13:17:12 crc kubenswrapper[4955]: I0217 13:17:12.699223 4955 generic.go:334] "Generic (PLEG): container finished" podID="102e3d48-f56c-4e7c-956e-23484bf45776" containerID="368b9ba7c845806bc25740e638518a5a47db6712c297dcf5a5a73ec18507df0c" exitCode=0 Feb 17 13:17:12 crc kubenswrapper[4955]: I0217 13:17:12.699303 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l" event={"ID":"102e3d48-f56c-4e7c-956e-23484bf45776","Type":"ContainerDied","Data":"368b9ba7c845806bc25740e638518a5a47db6712c297dcf5a5a73ec18507df0c"} Feb 17 13:17:12 crc kubenswrapper[4955]: I0217 13:17:12.699327 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l" event={"ID":"102e3d48-f56c-4e7c-956e-23484bf45776","Type":"ContainerStarted","Data":"2d509412573e42d88522dce0fc0858a849fb8146da2f834d11360616376285df"} Feb 17 13:17:13 crc kubenswrapper[4955]: I0217 13:17:13.496488 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wl2jr" Feb 17 13:17:13 crc kubenswrapper[4955]: I0217 13:17:13.496878 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wl2jr" Feb 17 13:17:13 crc kubenswrapper[4955]: I0217 13:17:13.709734 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jprkf" event={"ID":"5dc80fd1-674b-4442-90f4-10da0796e4fc","Type":"ContainerStarted","Data":"a367bafa929dd7c6310fd49b2ef42a73b33f6fcc4e45782b3809336c8b317b3a"} Feb 17 13:17:14 crc kubenswrapper[4955]: I0217 13:17:14.537492 4955 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wl2jr" podUID="5d226a93-aca2-4e11-9a77-6159b3ca087b" containerName="registry-server" probeResult="failure" output=< Feb 17 13:17:14 crc kubenswrapper[4955]: timeout: failed to connect service ":50051" within 1s Feb 17 13:17:14 crc kubenswrapper[4955]: > Feb 17 13:17:15 crc kubenswrapper[4955]: I0217 13:17:15.729164 4955 generic.go:334] "Generic (PLEG): container finished" podID="102e3d48-f56c-4e7c-956e-23484bf45776" containerID="3a02478eccf845c2aa47b3f56796f5684827dff192fe92bb014cf292a2bc114f" exitCode=0 Feb 17 13:17:15 crc kubenswrapper[4955]: I0217 13:17:15.729243 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l" event={"ID":"102e3d48-f56c-4e7c-956e-23484bf45776","Type":"ContainerDied","Data":"3a02478eccf845c2aa47b3f56796f5684827dff192fe92bb014cf292a2bc114f"} Feb 17 13:17:15 crc kubenswrapper[4955]: I0217 13:17:15.761230 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jprkf" podStartSLOduration=4.359935063 podStartE2EDuration="6.761206923s" podCreationTimestamp="2026-02-17 13:17:09 +0000 UTC" firstStartedPulling="2026-02-17 13:17:10.676450031 +0000 UTC m=+769.199179574" lastFinishedPulling="2026-02-17 13:17:13.077721881 +0000 UTC m=+771.600451434" observedRunningTime="2026-02-17 13:17:13.737055169 +0000 UTC m=+772.259784752" watchObservedRunningTime="2026-02-17 13:17:15.761206923 +0000 UTC m=+774.283936506" Feb 17 13:17:16 crc kubenswrapper[4955]: I0217 13:17:16.741859 4955 generic.go:334] "Generic (PLEG): container finished" podID="102e3d48-f56c-4e7c-956e-23484bf45776" containerID="a836f87bcddad3b14f59d58132c0a5f9584759cfd197916fb93c1f387d0ae017" exitCode=0 Feb 17 13:17:16 crc kubenswrapper[4955]: I0217 13:17:16.741935 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l" event={"ID":"102e3d48-f56c-4e7c-956e-23484bf45776","Type":"ContainerDied","Data":"a836f87bcddad3b14f59d58132c0a5f9584759cfd197916fb93c1f387d0ae017"} Feb 17 13:17:18 crc kubenswrapper[4955]: I0217 13:17:18.060499 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l" Feb 17 13:17:18 crc kubenswrapper[4955]: I0217 13:17:18.247611 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/102e3d48-f56c-4e7c-956e-23484bf45776-util\") pod \"102e3d48-f56c-4e7c-956e-23484bf45776\" (UID: \"102e3d48-f56c-4e7c-956e-23484bf45776\") " Feb 17 13:17:18 crc kubenswrapper[4955]: I0217 13:17:18.247698 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5r8mb\" (UniqueName: \"kubernetes.io/projected/102e3d48-f56c-4e7c-956e-23484bf45776-kube-api-access-5r8mb\") pod \"102e3d48-f56c-4e7c-956e-23484bf45776\" (UID: \"102e3d48-f56c-4e7c-956e-23484bf45776\") " Feb 17 13:17:18 crc kubenswrapper[4955]: I0217 13:17:18.247731 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/102e3d48-f56c-4e7c-956e-23484bf45776-bundle\") pod \"102e3d48-f56c-4e7c-956e-23484bf45776\" (UID: \"102e3d48-f56c-4e7c-956e-23484bf45776\") " Feb 17 13:17:18 crc kubenswrapper[4955]: I0217 13:17:18.250531 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/102e3d48-f56c-4e7c-956e-23484bf45776-bundle" (OuterVolumeSpecName: "bundle") pod "102e3d48-f56c-4e7c-956e-23484bf45776" (UID: "102e3d48-f56c-4e7c-956e-23484bf45776"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:17:18 crc kubenswrapper[4955]: I0217 13:17:18.257077 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/102e3d48-f56c-4e7c-956e-23484bf45776-kube-api-access-5r8mb" (OuterVolumeSpecName: "kube-api-access-5r8mb") pod "102e3d48-f56c-4e7c-956e-23484bf45776" (UID: "102e3d48-f56c-4e7c-956e-23484bf45776"). InnerVolumeSpecName "kube-api-access-5r8mb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:17:18 crc kubenswrapper[4955]: I0217 13:17:18.349868 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5r8mb\" (UniqueName: \"kubernetes.io/projected/102e3d48-f56c-4e7c-956e-23484bf45776-kube-api-access-5r8mb\") on node \"crc\" DevicePath \"\"" Feb 17 13:17:18 crc kubenswrapper[4955]: I0217 13:17:18.349936 4955 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/102e3d48-f56c-4e7c-956e-23484bf45776-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:17:18 crc kubenswrapper[4955]: I0217 13:17:18.761932 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l" event={"ID":"102e3d48-f56c-4e7c-956e-23484bf45776","Type":"ContainerDied","Data":"2d509412573e42d88522dce0fc0858a849fb8146da2f834d11360616376285df"} Feb 17 13:17:18 crc kubenswrapper[4955]: I0217 13:17:18.761991 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d509412573e42d88522dce0fc0858a849fb8146da2f834d11360616376285df" Feb 17 13:17:18 crc kubenswrapper[4955]: I0217 13:17:18.762017 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l" Feb 17 13:17:19 crc kubenswrapper[4955]: I0217 13:17:19.851259 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jprkf" Feb 17 13:17:19 crc kubenswrapper[4955]: I0217 13:17:19.851670 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jprkf" Feb 17 13:17:19 crc kubenswrapper[4955]: I0217 13:17:19.920597 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jprkf" Feb 17 13:17:20 crc kubenswrapper[4955]: I0217 13:17:20.526426 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/102e3d48-f56c-4e7c-956e-23484bf45776-util" (OuterVolumeSpecName: "util") pod "102e3d48-f56c-4e7c-956e-23484bf45776" (UID: "102e3d48-f56c-4e7c-956e-23484bf45776"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:17:20 crc kubenswrapper[4955]: I0217 13:17:20.585692 4955 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/102e3d48-f56c-4e7c-956e-23484bf45776-util\") on node \"crc\" DevicePath \"\"" Feb 17 13:17:20 crc kubenswrapper[4955]: I0217 13:17:20.829016 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jprkf" Feb 17 13:17:22 crc kubenswrapper[4955]: I0217 13:17:22.303769 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jprkf"] Feb 17 13:17:23 crc kubenswrapper[4955]: I0217 13:17:23.565214 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wl2jr" Feb 17 13:17:23 crc kubenswrapper[4955]: I0217 13:17:23.624535 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wl2jr" Feb 17 13:17:23 crc kubenswrapper[4955]: I0217 13:17:23.795420 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jprkf" podUID="5dc80fd1-674b-4442-90f4-10da0796e4fc" containerName="registry-server" containerID="cri-o://a367bafa929dd7c6310fd49b2ef42a73b33f6fcc4e45782b3809336c8b317b3a" gracePeriod=2 Feb 17 13:17:24 crc kubenswrapper[4955]: I0217 13:17:24.222838 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jprkf" Feb 17 13:17:24 crc kubenswrapper[4955]: I0217 13:17:24.336587 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5dc80fd1-674b-4442-90f4-10da0796e4fc-catalog-content\") pod \"5dc80fd1-674b-4442-90f4-10da0796e4fc\" (UID: \"5dc80fd1-674b-4442-90f4-10da0796e4fc\") " Feb 17 13:17:24 crc kubenswrapper[4955]: I0217 13:17:24.336659 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvrpg\" (UniqueName: \"kubernetes.io/projected/5dc80fd1-674b-4442-90f4-10da0796e4fc-kube-api-access-lvrpg\") pod \"5dc80fd1-674b-4442-90f4-10da0796e4fc\" (UID: \"5dc80fd1-674b-4442-90f4-10da0796e4fc\") " Feb 17 13:17:24 crc kubenswrapper[4955]: I0217 13:17:24.336761 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5dc80fd1-674b-4442-90f4-10da0796e4fc-utilities\") pod \"5dc80fd1-674b-4442-90f4-10da0796e4fc\" (UID: \"5dc80fd1-674b-4442-90f4-10da0796e4fc\") " Feb 17 13:17:24 crc kubenswrapper[4955]: I0217 13:17:24.338203 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5dc80fd1-674b-4442-90f4-10da0796e4fc-utilities" (OuterVolumeSpecName: "utilities") pod "5dc80fd1-674b-4442-90f4-10da0796e4fc" (UID: "5dc80fd1-674b-4442-90f4-10da0796e4fc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:17:24 crc kubenswrapper[4955]: I0217 13:17:24.342161 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5dc80fd1-674b-4442-90f4-10da0796e4fc-kube-api-access-lvrpg" (OuterVolumeSpecName: "kube-api-access-lvrpg") pod "5dc80fd1-674b-4442-90f4-10da0796e4fc" (UID: "5dc80fd1-674b-4442-90f4-10da0796e4fc"). InnerVolumeSpecName "kube-api-access-lvrpg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:17:24 crc kubenswrapper[4955]: I0217 13:17:24.389340 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5dc80fd1-674b-4442-90f4-10da0796e4fc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5dc80fd1-674b-4442-90f4-10da0796e4fc" (UID: "5dc80fd1-674b-4442-90f4-10da0796e4fc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:17:24 crc kubenswrapper[4955]: I0217 13:17:24.437946 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvrpg\" (UniqueName: \"kubernetes.io/projected/5dc80fd1-674b-4442-90f4-10da0796e4fc-kube-api-access-lvrpg\") on node \"crc\" DevicePath \"\"" Feb 17 13:17:24 crc kubenswrapper[4955]: I0217 13:17:24.437980 4955 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5dc80fd1-674b-4442-90f4-10da0796e4fc-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 13:17:24 crc kubenswrapper[4955]: I0217 13:17:24.437990 4955 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5dc80fd1-674b-4442-90f4-10da0796e4fc-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 13:17:24 crc kubenswrapper[4955]: I0217 13:17:24.806637 4955 generic.go:334] "Generic (PLEG): container finished" podID="5dc80fd1-674b-4442-90f4-10da0796e4fc" containerID="a367bafa929dd7c6310fd49b2ef42a73b33f6fcc4e45782b3809336c8b317b3a" exitCode=0 Feb 17 13:17:24 crc kubenswrapper[4955]: I0217 13:17:24.806707 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jprkf" event={"ID":"5dc80fd1-674b-4442-90f4-10da0796e4fc","Type":"ContainerDied","Data":"a367bafa929dd7c6310fd49b2ef42a73b33f6fcc4e45782b3809336c8b317b3a"} Feb 17 13:17:24 crc kubenswrapper[4955]: I0217 13:17:24.807167 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jprkf" event={"ID":"5dc80fd1-674b-4442-90f4-10da0796e4fc","Type":"ContainerDied","Data":"a8b31833842d7296995204acc31ff10586708908364cb6950dc2bd8b2a85372c"} Feb 17 13:17:24 crc kubenswrapper[4955]: I0217 13:17:24.807215 4955 scope.go:117] "RemoveContainer" containerID="a367bafa929dd7c6310fd49b2ef42a73b33f6fcc4e45782b3809336c8b317b3a" Feb 17 13:17:24 crc kubenswrapper[4955]: I0217 13:17:24.806735 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jprkf" Feb 17 13:17:24 crc kubenswrapper[4955]: I0217 13:17:24.834520 4955 scope.go:117] "RemoveContainer" containerID="3926052d617dc9ed98152ba808ac9c1020fe756329269f685ee109da9e4c11b1" Feb 17 13:17:24 crc kubenswrapper[4955]: I0217 13:17:24.852651 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jprkf"] Feb 17 13:17:24 crc kubenswrapper[4955]: I0217 13:17:24.857582 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jprkf"] Feb 17 13:17:24 crc kubenswrapper[4955]: I0217 13:17:24.869887 4955 scope.go:117] "RemoveContainer" containerID="1dd886d04cead8fe44ea298047debf672ae809651f0b21c88eab244186e68c6a" Feb 17 13:17:24 crc kubenswrapper[4955]: I0217 13:17:24.888639 4955 scope.go:117] "RemoveContainer" containerID="a367bafa929dd7c6310fd49b2ef42a73b33f6fcc4e45782b3809336c8b317b3a" Feb 17 13:17:24 crc kubenswrapper[4955]: E0217 13:17:24.889153 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a367bafa929dd7c6310fd49b2ef42a73b33f6fcc4e45782b3809336c8b317b3a\": container with ID starting with a367bafa929dd7c6310fd49b2ef42a73b33f6fcc4e45782b3809336c8b317b3a not found: ID does not exist" containerID="a367bafa929dd7c6310fd49b2ef42a73b33f6fcc4e45782b3809336c8b317b3a" Feb 17 13:17:24 crc kubenswrapper[4955]: I0217 13:17:24.889202 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a367bafa929dd7c6310fd49b2ef42a73b33f6fcc4e45782b3809336c8b317b3a"} err="failed to get container status \"a367bafa929dd7c6310fd49b2ef42a73b33f6fcc4e45782b3809336c8b317b3a\": rpc error: code = NotFound desc = could not find container \"a367bafa929dd7c6310fd49b2ef42a73b33f6fcc4e45782b3809336c8b317b3a\": container with ID starting with a367bafa929dd7c6310fd49b2ef42a73b33f6fcc4e45782b3809336c8b317b3a not found: ID does not exist" Feb 17 13:17:24 crc kubenswrapper[4955]: I0217 13:17:24.889228 4955 scope.go:117] "RemoveContainer" containerID="3926052d617dc9ed98152ba808ac9c1020fe756329269f685ee109da9e4c11b1" Feb 17 13:17:24 crc kubenswrapper[4955]: E0217 13:17:24.889594 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3926052d617dc9ed98152ba808ac9c1020fe756329269f685ee109da9e4c11b1\": container with ID starting with 3926052d617dc9ed98152ba808ac9c1020fe756329269f685ee109da9e4c11b1 not found: ID does not exist" containerID="3926052d617dc9ed98152ba808ac9c1020fe756329269f685ee109da9e4c11b1" Feb 17 13:17:24 crc kubenswrapper[4955]: I0217 13:17:24.889611 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3926052d617dc9ed98152ba808ac9c1020fe756329269f685ee109da9e4c11b1"} err="failed to get container status \"3926052d617dc9ed98152ba808ac9c1020fe756329269f685ee109da9e4c11b1\": rpc error: code = NotFound desc = could not find container \"3926052d617dc9ed98152ba808ac9c1020fe756329269f685ee109da9e4c11b1\": container with ID starting with 3926052d617dc9ed98152ba808ac9c1020fe756329269f685ee109da9e4c11b1 not found: ID does not exist" Feb 17 13:17:24 crc kubenswrapper[4955]: I0217 13:17:24.889623 4955 scope.go:117] "RemoveContainer" containerID="1dd886d04cead8fe44ea298047debf672ae809651f0b21c88eab244186e68c6a" Feb 17 13:17:24 crc kubenswrapper[4955]: E0217 13:17:24.889849 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1dd886d04cead8fe44ea298047debf672ae809651f0b21c88eab244186e68c6a\": container with ID starting with 1dd886d04cead8fe44ea298047debf672ae809651f0b21c88eab244186e68c6a not found: ID does not exist" containerID="1dd886d04cead8fe44ea298047debf672ae809651f0b21c88eab244186e68c6a" Feb 17 13:17:24 crc kubenswrapper[4955]: I0217 13:17:24.889865 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1dd886d04cead8fe44ea298047debf672ae809651f0b21c88eab244186e68c6a"} err="failed to get container status \"1dd886d04cead8fe44ea298047debf672ae809651f0b21c88eab244186e68c6a\": rpc error: code = NotFound desc = could not find container \"1dd886d04cead8fe44ea298047debf672ae809651f0b21c88eab244186e68c6a\": container with ID starting with 1dd886d04cead8fe44ea298047debf672ae809651f0b21c88eab244186e68c6a not found: ID does not exist" Feb 17 13:17:26 crc kubenswrapper[4955]: I0217 13:17:26.235399 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5dc80fd1-674b-4442-90f4-10da0796e4fc" path="/var/lib/kubelet/pods/5dc80fd1-674b-4442-90f4-10da0796e4fc/volumes" Feb 17 13:17:26 crc kubenswrapper[4955]: I0217 13:17:26.902145 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wl2jr"] Feb 17 13:17:26 crc kubenswrapper[4955]: I0217 13:17:26.902405 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wl2jr" podUID="5d226a93-aca2-4e11-9a77-6159b3ca087b" containerName="registry-server" containerID="cri-o://7cafb2e453e05faf238c715aa28370a4d54bae35af8a5b5c05a1e630343ed317" gracePeriod=2 Feb 17 13:17:27 crc kubenswrapper[4955]: I0217 13:17:27.233573 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wl2jr" Feb 17 13:17:27 crc kubenswrapper[4955]: I0217 13:17:27.379495 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78pvf\" (UniqueName: \"kubernetes.io/projected/5d226a93-aca2-4e11-9a77-6159b3ca087b-kube-api-access-78pvf\") pod \"5d226a93-aca2-4e11-9a77-6159b3ca087b\" (UID: \"5d226a93-aca2-4e11-9a77-6159b3ca087b\") " Feb 17 13:17:27 crc kubenswrapper[4955]: I0217 13:17:27.379544 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d226a93-aca2-4e11-9a77-6159b3ca087b-utilities\") pod \"5d226a93-aca2-4e11-9a77-6159b3ca087b\" (UID: \"5d226a93-aca2-4e11-9a77-6159b3ca087b\") " Feb 17 13:17:27 crc kubenswrapper[4955]: I0217 13:17:27.379645 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d226a93-aca2-4e11-9a77-6159b3ca087b-catalog-content\") pod \"5d226a93-aca2-4e11-9a77-6159b3ca087b\" (UID: \"5d226a93-aca2-4e11-9a77-6159b3ca087b\") " Feb 17 13:17:27 crc kubenswrapper[4955]: I0217 13:17:27.380948 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d226a93-aca2-4e11-9a77-6159b3ca087b-utilities" (OuterVolumeSpecName: "utilities") pod "5d226a93-aca2-4e11-9a77-6159b3ca087b" (UID: "5d226a93-aca2-4e11-9a77-6159b3ca087b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:17:27 crc kubenswrapper[4955]: I0217 13:17:27.385360 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d226a93-aca2-4e11-9a77-6159b3ca087b-kube-api-access-78pvf" (OuterVolumeSpecName: "kube-api-access-78pvf") pod "5d226a93-aca2-4e11-9a77-6159b3ca087b" (UID: "5d226a93-aca2-4e11-9a77-6159b3ca087b"). InnerVolumeSpecName "kube-api-access-78pvf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:17:27 crc kubenswrapper[4955]: I0217 13:17:27.481453 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78pvf\" (UniqueName: \"kubernetes.io/projected/5d226a93-aca2-4e11-9a77-6159b3ca087b-kube-api-access-78pvf\") on node \"crc\" DevicePath \"\"" Feb 17 13:17:27 crc kubenswrapper[4955]: I0217 13:17:27.481493 4955 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d226a93-aca2-4e11-9a77-6159b3ca087b-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 13:17:27 crc kubenswrapper[4955]: I0217 13:17:27.522675 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d226a93-aca2-4e11-9a77-6159b3ca087b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5d226a93-aca2-4e11-9a77-6159b3ca087b" (UID: "5d226a93-aca2-4e11-9a77-6159b3ca087b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:17:27 crc kubenswrapper[4955]: I0217 13:17:27.582936 4955 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d226a93-aca2-4e11-9a77-6159b3ca087b-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 13:17:27 crc kubenswrapper[4955]: I0217 13:17:27.832121 4955 generic.go:334] "Generic (PLEG): container finished" podID="5d226a93-aca2-4e11-9a77-6159b3ca087b" containerID="7cafb2e453e05faf238c715aa28370a4d54bae35af8a5b5c05a1e630343ed317" exitCode=0 Feb 17 13:17:27 crc kubenswrapper[4955]: I0217 13:17:27.832626 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wl2jr" event={"ID":"5d226a93-aca2-4e11-9a77-6159b3ca087b","Type":"ContainerDied","Data":"7cafb2e453e05faf238c715aa28370a4d54bae35af8a5b5c05a1e630343ed317"} Feb 17 13:17:27 crc kubenswrapper[4955]: I0217 13:17:27.832807 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wl2jr" event={"ID":"5d226a93-aca2-4e11-9a77-6159b3ca087b","Type":"ContainerDied","Data":"b20d4d84ec4a7e251577e17661022d5029fbcb0ec17256d7394203b5673bfabb"} Feb 17 13:17:27 crc kubenswrapper[4955]: I0217 13:17:27.832939 4955 scope.go:117] "RemoveContainer" containerID="7cafb2e453e05faf238c715aa28370a4d54bae35af8a5b5c05a1e630343ed317" Feb 17 13:17:27 crc kubenswrapper[4955]: I0217 13:17:27.833196 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wl2jr" Feb 17 13:17:27 crc kubenswrapper[4955]: I0217 13:17:27.848850 4955 scope.go:117] "RemoveContainer" containerID="5f2d65d1b698a552cbe04bf7aee28ccafa6430c34fd28ae624c6d8eab6716e7c" Feb 17 13:17:27 crc kubenswrapper[4955]: I0217 13:17:27.872029 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wl2jr"] Feb 17 13:17:27 crc kubenswrapper[4955]: I0217 13:17:27.879383 4955 scope.go:117] "RemoveContainer" containerID="e9c15e95a677567793bc31417f9c8a220a357d7ec072566d0d85c4d5f93acd4c" Feb 17 13:17:27 crc kubenswrapper[4955]: I0217 13:17:27.885141 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wl2jr"] Feb 17 13:17:27 crc kubenswrapper[4955]: I0217 13:17:27.896453 4955 scope.go:117] "RemoveContainer" containerID="7cafb2e453e05faf238c715aa28370a4d54bae35af8a5b5c05a1e630343ed317" Feb 17 13:17:27 crc kubenswrapper[4955]: E0217 13:17:27.896957 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7cafb2e453e05faf238c715aa28370a4d54bae35af8a5b5c05a1e630343ed317\": container with ID starting with 7cafb2e453e05faf238c715aa28370a4d54bae35af8a5b5c05a1e630343ed317 not found: ID does not exist" containerID="7cafb2e453e05faf238c715aa28370a4d54bae35af8a5b5c05a1e630343ed317" Feb 17 13:17:27 crc kubenswrapper[4955]: I0217 13:17:27.897009 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7cafb2e453e05faf238c715aa28370a4d54bae35af8a5b5c05a1e630343ed317"} err="failed to get container status \"7cafb2e453e05faf238c715aa28370a4d54bae35af8a5b5c05a1e630343ed317\": rpc error: code = NotFound desc = could not find container \"7cafb2e453e05faf238c715aa28370a4d54bae35af8a5b5c05a1e630343ed317\": container with ID starting with 7cafb2e453e05faf238c715aa28370a4d54bae35af8a5b5c05a1e630343ed317 not found: ID does not exist" Feb 17 13:17:27 crc kubenswrapper[4955]: I0217 13:17:27.897040 4955 scope.go:117] "RemoveContainer" containerID="5f2d65d1b698a552cbe04bf7aee28ccafa6430c34fd28ae624c6d8eab6716e7c" Feb 17 13:17:27 crc kubenswrapper[4955]: E0217 13:17:27.897436 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f2d65d1b698a552cbe04bf7aee28ccafa6430c34fd28ae624c6d8eab6716e7c\": container with ID starting with 5f2d65d1b698a552cbe04bf7aee28ccafa6430c34fd28ae624c6d8eab6716e7c not found: ID does not exist" containerID="5f2d65d1b698a552cbe04bf7aee28ccafa6430c34fd28ae624c6d8eab6716e7c" Feb 17 13:17:27 crc kubenswrapper[4955]: I0217 13:17:27.897466 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f2d65d1b698a552cbe04bf7aee28ccafa6430c34fd28ae624c6d8eab6716e7c"} err="failed to get container status \"5f2d65d1b698a552cbe04bf7aee28ccafa6430c34fd28ae624c6d8eab6716e7c\": rpc error: code = NotFound desc = could not find container \"5f2d65d1b698a552cbe04bf7aee28ccafa6430c34fd28ae624c6d8eab6716e7c\": container with ID starting with 5f2d65d1b698a552cbe04bf7aee28ccafa6430c34fd28ae624c6d8eab6716e7c not found: ID does not exist" Feb 17 13:17:27 crc kubenswrapper[4955]: I0217 13:17:27.897485 4955 scope.go:117] "RemoveContainer" containerID="e9c15e95a677567793bc31417f9c8a220a357d7ec072566d0d85c4d5f93acd4c" Feb 17 13:17:27 crc kubenswrapper[4955]: E0217 13:17:27.897728 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9c15e95a677567793bc31417f9c8a220a357d7ec072566d0d85c4d5f93acd4c\": container with ID starting with e9c15e95a677567793bc31417f9c8a220a357d7ec072566d0d85c4d5f93acd4c not found: ID does not exist" containerID="e9c15e95a677567793bc31417f9c8a220a357d7ec072566d0d85c4d5f93acd4c" Feb 17 13:17:27 crc kubenswrapper[4955]: I0217 13:17:27.897759 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9c15e95a677567793bc31417f9c8a220a357d7ec072566d0d85c4d5f93acd4c"} err="failed to get container status \"e9c15e95a677567793bc31417f9c8a220a357d7ec072566d0d85c4d5f93acd4c\": rpc error: code = NotFound desc = could not find container \"e9c15e95a677567793bc31417f9c8a220a357d7ec072566d0d85c4d5f93acd4c\": container with ID starting with e9c15e95a677567793bc31417f9c8a220a357d7ec072566d0d85c4d5f93acd4c not found: ID does not exist" Feb 17 13:17:28 crc kubenswrapper[4955]: I0217 13:17:28.231963 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d226a93-aca2-4e11-9a77-6159b3ca087b" path="/var/lib/kubelet/pods/5d226a93-aca2-4e11-9a77-6159b3ca087b/volumes" Feb 17 13:17:28 crc kubenswrapper[4955]: I0217 13:17:28.942299 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-754bdc4867-j8r95"] Feb 17 13:17:28 crc kubenswrapper[4955]: E0217 13:17:28.942519 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d226a93-aca2-4e11-9a77-6159b3ca087b" containerName="extract-content" Feb 17 13:17:28 crc kubenswrapper[4955]: I0217 13:17:28.942533 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d226a93-aca2-4e11-9a77-6159b3ca087b" containerName="extract-content" Feb 17 13:17:28 crc kubenswrapper[4955]: E0217 13:17:28.942552 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dc80fd1-674b-4442-90f4-10da0796e4fc" containerName="extract-content" Feb 17 13:17:28 crc kubenswrapper[4955]: I0217 13:17:28.942560 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dc80fd1-674b-4442-90f4-10da0796e4fc" containerName="extract-content" Feb 17 13:17:28 crc kubenswrapper[4955]: E0217 13:17:28.942571 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d226a93-aca2-4e11-9a77-6159b3ca087b" containerName="registry-server" Feb 17 13:17:28 crc kubenswrapper[4955]: I0217 13:17:28.942579 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d226a93-aca2-4e11-9a77-6159b3ca087b" containerName="registry-server" Feb 17 13:17:28 crc kubenswrapper[4955]: E0217 13:17:28.942589 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d226a93-aca2-4e11-9a77-6159b3ca087b" containerName="extract-utilities" Feb 17 13:17:28 crc kubenswrapper[4955]: I0217 13:17:28.942596 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d226a93-aca2-4e11-9a77-6159b3ca087b" containerName="extract-utilities" Feb 17 13:17:28 crc kubenswrapper[4955]: E0217 13:17:28.942608 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="102e3d48-f56c-4e7c-956e-23484bf45776" containerName="util" Feb 17 13:17:28 crc kubenswrapper[4955]: I0217 13:17:28.942615 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="102e3d48-f56c-4e7c-956e-23484bf45776" containerName="util" Feb 17 13:17:28 crc kubenswrapper[4955]: E0217 13:17:28.942628 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="102e3d48-f56c-4e7c-956e-23484bf45776" containerName="pull" Feb 17 13:17:28 crc kubenswrapper[4955]: I0217 13:17:28.942635 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="102e3d48-f56c-4e7c-956e-23484bf45776" containerName="pull" Feb 17 13:17:28 crc kubenswrapper[4955]: E0217 13:17:28.942648 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dc80fd1-674b-4442-90f4-10da0796e4fc" containerName="registry-server" Feb 17 13:17:28 crc kubenswrapper[4955]: I0217 13:17:28.942656 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dc80fd1-674b-4442-90f4-10da0796e4fc" containerName="registry-server" Feb 17 13:17:28 crc kubenswrapper[4955]: E0217 13:17:28.942665 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="102e3d48-f56c-4e7c-956e-23484bf45776" containerName="extract" Feb 17 13:17:28 crc kubenswrapper[4955]: I0217 13:17:28.942672 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="102e3d48-f56c-4e7c-956e-23484bf45776" containerName="extract" Feb 17 13:17:28 crc kubenswrapper[4955]: E0217 13:17:28.942686 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dc80fd1-674b-4442-90f4-10da0796e4fc" containerName="extract-utilities" Feb 17 13:17:28 crc kubenswrapper[4955]: I0217 13:17:28.942694 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dc80fd1-674b-4442-90f4-10da0796e4fc" containerName="extract-utilities" Feb 17 13:17:28 crc kubenswrapper[4955]: I0217 13:17:28.944261 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="102e3d48-f56c-4e7c-956e-23484bf45776" containerName="extract" Feb 17 13:17:28 crc kubenswrapper[4955]: I0217 13:17:28.944288 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d226a93-aca2-4e11-9a77-6159b3ca087b" containerName="registry-server" Feb 17 13:17:28 crc kubenswrapper[4955]: I0217 13:17:28.944309 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="5dc80fd1-674b-4442-90f4-10da0796e4fc" containerName="registry-server" Feb 17 13:17:28 crc kubenswrapper[4955]: I0217 13:17:28.944850 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-754bdc4867-j8r95" Feb 17 13:17:28 crc kubenswrapper[4955]: I0217 13:17:28.946601 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Feb 17 13:17:28 crc kubenswrapper[4955]: I0217 13:17:28.946838 4955 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-2c5vw" Feb 17 13:17:28 crc kubenswrapper[4955]: I0217 13:17:28.946904 4955 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Feb 17 13:17:28 crc kubenswrapper[4955]: I0217 13:17:28.950021 4955 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Feb 17 13:17:28 crc kubenswrapper[4955]: I0217 13:17:28.950225 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Feb 17 13:17:28 crc kubenswrapper[4955]: I0217 13:17:28.965022 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-754bdc4867-j8r95"] Feb 17 13:17:29 crc kubenswrapper[4955]: I0217 13:17:29.099878 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6f6cf1f8-d54b-4fad-8313-c5f235a9df88-webhook-cert\") pod \"metallb-operator-controller-manager-754bdc4867-j8r95\" (UID: \"6f6cf1f8-d54b-4fad-8313-c5f235a9df88\") " pod="metallb-system/metallb-operator-controller-manager-754bdc4867-j8r95" Feb 17 13:17:29 crc kubenswrapper[4955]: I0217 13:17:29.099964 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6f6cf1f8-d54b-4fad-8313-c5f235a9df88-apiservice-cert\") pod \"metallb-operator-controller-manager-754bdc4867-j8r95\" (UID: \"6f6cf1f8-d54b-4fad-8313-c5f235a9df88\") " pod="metallb-system/metallb-operator-controller-manager-754bdc4867-j8r95" Feb 17 13:17:29 crc kubenswrapper[4955]: I0217 13:17:29.099996 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhw5x\" (UniqueName: \"kubernetes.io/projected/6f6cf1f8-d54b-4fad-8313-c5f235a9df88-kube-api-access-rhw5x\") pod \"metallb-operator-controller-manager-754bdc4867-j8r95\" (UID: \"6f6cf1f8-d54b-4fad-8313-c5f235a9df88\") " pod="metallb-system/metallb-operator-controller-manager-754bdc4867-j8r95" Feb 17 13:17:29 crc kubenswrapper[4955]: I0217 13:17:29.200861 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6f6cf1f8-d54b-4fad-8313-c5f235a9df88-webhook-cert\") pod \"metallb-operator-controller-manager-754bdc4867-j8r95\" (UID: \"6f6cf1f8-d54b-4fad-8313-c5f235a9df88\") " pod="metallb-system/metallb-operator-controller-manager-754bdc4867-j8r95" Feb 17 13:17:29 crc kubenswrapper[4955]: I0217 13:17:29.200955 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6f6cf1f8-d54b-4fad-8313-c5f235a9df88-apiservice-cert\") pod \"metallb-operator-controller-manager-754bdc4867-j8r95\" (UID: \"6f6cf1f8-d54b-4fad-8313-c5f235a9df88\") " pod="metallb-system/metallb-operator-controller-manager-754bdc4867-j8r95" Feb 17 13:17:29 crc kubenswrapper[4955]: I0217 13:17:29.200981 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhw5x\" (UniqueName: \"kubernetes.io/projected/6f6cf1f8-d54b-4fad-8313-c5f235a9df88-kube-api-access-rhw5x\") pod \"metallb-operator-controller-manager-754bdc4867-j8r95\" (UID: \"6f6cf1f8-d54b-4fad-8313-c5f235a9df88\") " pod="metallb-system/metallb-operator-controller-manager-754bdc4867-j8r95" Feb 17 13:17:29 crc kubenswrapper[4955]: I0217 13:17:29.206339 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6f6cf1f8-d54b-4fad-8313-c5f235a9df88-webhook-cert\") pod \"metallb-operator-controller-manager-754bdc4867-j8r95\" (UID: \"6f6cf1f8-d54b-4fad-8313-c5f235a9df88\") " pod="metallb-system/metallb-operator-controller-manager-754bdc4867-j8r95" Feb 17 13:17:29 crc kubenswrapper[4955]: I0217 13:17:29.215847 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhw5x\" (UniqueName: \"kubernetes.io/projected/6f6cf1f8-d54b-4fad-8313-c5f235a9df88-kube-api-access-rhw5x\") pod \"metallb-operator-controller-manager-754bdc4867-j8r95\" (UID: \"6f6cf1f8-d54b-4fad-8313-c5f235a9df88\") " pod="metallb-system/metallb-operator-controller-manager-754bdc4867-j8r95" Feb 17 13:17:29 crc kubenswrapper[4955]: I0217 13:17:29.217264 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6f6cf1f8-d54b-4fad-8313-c5f235a9df88-apiservice-cert\") pod \"metallb-operator-controller-manager-754bdc4867-j8r95\" (UID: \"6f6cf1f8-d54b-4fad-8313-c5f235a9df88\") " pod="metallb-system/metallb-operator-controller-manager-754bdc4867-j8r95" Feb 17 13:17:29 crc kubenswrapper[4955]: I0217 13:17:29.261947 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-754bdc4867-j8r95" Feb 17 13:17:29 crc kubenswrapper[4955]: I0217 13:17:29.282619 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-579d9646f7-zpz6n"] Feb 17 13:17:29 crc kubenswrapper[4955]: I0217 13:17:29.283384 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-579d9646f7-zpz6n" Feb 17 13:17:29 crc kubenswrapper[4955]: I0217 13:17:29.286692 4955 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-dxb59" Feb 17 13:17:29 crc kubenswrapper[4955]: I0217 13:17:29.286902 4955 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Feb 17 13:17:29 crc kubenswrapper[4955]: I0217 13:17:29.287023 4955 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Feb 17 13:17:29 crc kubenswrapper[4955]: I0217 13:17:29.309838 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-579d9646f7-zpz6n"] Feb 17 13:17:29 crc kubenswrapper[4955]: I0217 13:17:29.402981 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kq7dw\" (UniqueName: \"kubernetes.io/projected/4b384d59-9dc8-4d70-8b12-bd97252d1cfa-kube-api-access-kq7dw\") pod \"metallb-operator-webhook-server-579d9646f7-zpz6n\" (UID: \"4b384d59-9dc8-4d70-8b12-bd97252d1cfa\") " pod="metallb-system/metallb-operator-webhook-server-579d9646f7-zpz6n" Feb 17 13:17:29 crc kubenswrapper[4955]: I0217 13:17:29.403318 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4b384d59-9dc8-4d70-8b12-bd97252d1cfa-apiservice-cert\") pod \"metallb-operator-webhook-server-579d9646f7-zpz6n\" (UID: \"4b384d59-9dc8-4d70-8b12-bd97252d1cfa\") " pod="metallb-system/metallb-operator-webhook-server-579d9646f7-zpz6n" Feb 17 13:17:29 crc kubenswrapper[4955]: I0217 13:17:29.403430 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4b384d59-9dc8-4d70-8b12-bd97252d1cfa-webhook-cert\") pod \"metallb-operator-webhook-server-579d9646f7-zpz6n\" (UID: \"4b384d59-9dc8-4d70-8b12-bd97252d1cfa\") " pod="metallb-system/metallb-operator-webhook-server-579d9646f7-zpz6n" Feb 17 13:17:29 crc kubenswrapper[4955]: I0217 13:17:29.504510 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4b384d59-9dc8-4d70-8b12-bd97252d1cfa-webhook-cert\") pod \"metallb-operator-webhook-server-579d9646f7-zpz6n\" (UID: \"4b384d59-9dc8-4d70-8b12-bd97252d1cfa\") " pod="metallb-system/metallb-operator-webhook-server-579d9646f7-zpz6n" Feb 17 13:17:29 crc kubenswrapper[4955]: I0217 13:17:29.504559 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kq7dw\" (UniqueName: \"kubernetes.io/projected/4b384d59-9dc8-4d70-8b12-bd97252d1cfa-kube-api-access-kq7dw\") pod \"metallb-operator-webhook-server-579d9646f7-zpz6n\" (UID: \"4b384d59-9dc8-4d70-8b12-bd97252d1cfa\") " pod="metallb-system/metallb-operator-webhook-server-579d9646f7-zpz6n" Feb 17 13:17:29 crc kubenswrapper[4955]: I0217 13:17:29.504590 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4b384d59-9dc8-4d70-8b12-bd97252d1cfa-apiservice-cert\") pod \"metallb-operator-webhook-server-579d9646f7-zpz6n\" (UID: \"4b384d59-9dc8-4d70-8b12-bd97252d1cfa\") " pod="metallb-system/metallb-operator-webhook-server-579d9646f7-zpz6n" Feb 17 13:17:29 crc kubenswrapper[4955]: I0217 13:17:29.512324 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4b384d59-9dc8-4d70-8b12-bd97252d1cfa-apiservice-cert\") pod \"metallb-operator-webhook-server-579d9646f7-zpz6n\" (UID: \"4b384d59-9dc8-4d70-8b12-bd97252d1cfa\") " pod="metallb-system/metallb-operator-webhook-server-579d9646f7-zpz6n" Feb 17 13:17:29 crc kubenswrapper[4955]: I0217 13:17:29.519433 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4b384d59-9dc8-4d70-8b12-bd97252d1cfa-webhook-cert\") pod \"metallb-operator-webhook-server-579d9646f7-zpz6n\" (UID: \"4b384d59-9dc8-4d70-8b12-bd97252d1cfa\") " pod="metallb-system/metallb-operator-webhook-server-579d9646f7-zpz6n" Feb 17 13:17:29 crc kubenswrapper[4955]: I0217 13:17:29.535217 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kq7dw\" (UniqueName: \"kubernetes.io/projected/4b384d59-9dc8-4d70-8b12-bd97252d1cfa-kube-api-access-kq7dw\") pod \"metallb-operator-webhook-server-579d9646f7-zpz6n\" (UID: \"4b384d59-9dc8-4d70-8b12-bd97252d1cfa\") " pod="metallb-system/metallb-operator-webhook-server-579d9646f7-zpz6n" Feb 17 13:17:29 crc kubenswrapper[4955]: I0217 13:17:29.633944 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-579d9646f7-zpz6n" Feb 17 13:17:29 crc kubenswrapper[4955]: I0217 13:17:29.713465 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-754bdc4867-j8r95"] Feb 17 13:17:29 crc kubenswrapper[4955]: W0217 13:17:29.720475 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6f6cf1f8_d54b_4fad_8313_c5f235a9df88.slice/crio-6d3d20c664ad5196b2e59887838c5f0328fc14fb307f68313528bd56f078accc WatchSource:0}: Error finding container 6d3d20c664ad5196b2e59887838c5f0328fc14fb307f68313528bd56f078accc: Status 404 returned error can't find the container with id 6d3d20c664ad5196b2e59887838c5f0328fc14fb307f68313528bd56f078accc Feb 17 13:17:29 crc kubenswrapper[4955]: I0217 13:17:29.844722 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-754bdc4867-j8r95" event={"ID":"6f6cf1f8-d54b-4fad-8313-c5f235a9df88","Type":"ContainerStarted","Data":"6d3d20c664ad5196b2e59887838c5f0328fc14fb307f68313528bd56f078accc"} Feb 17 13:17:29 crc kubenswrapper[4955]: I0217 13:17:29.852414 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-579d9646f7-zpz6n"] Feb 17 13:17:29 crc kubenswrapper[4955]: W0217 13:17:29.857712 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b384d59_9dc8_4d70_8b12_bd97252d1cfa.slice/crio-7f3b3aa51cb2aec7effc1dca449da72abceb5f09b398b1a9461b2f8da33219e4 WatchSource:0}: Error finding container 7f3b3aa51cb2aec7effc1dca449da72abceb5f09b398b1a9461b2f8da33219e4: Status 404 returned error can't find the container with id 7f3b3aa51cb2aec7effc1dca449da72abceb5f09b398b1a9461b2f8da33219e4 Feb 17 13:17:30 crc kubenswrapper[4955]: I0217 13:17:30.513735 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-74q9j"] Feb 17 13:17:30 crc kubenswrapper[4955]: I0217 13:17:30.515473 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-74q9j" Feb 17 13:17:30 crc kubenswrapper[4955]: I0217 13:17:30.528041 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-74q9j"] Feb 17 13:17:30 crc kubenswrapper[4955]: I0217 13:17:30.619903 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e-utilities\") pod \"redhat-marketplace-74q9j\" (UID: \"d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e\") " pod="openshift-marketplace/redhat-marketplace-74q9j" Feb 17 13:17:30 crc kubenswrapper[4955]: I0217 13:17:30.619969 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e-catalog-content\") pod \"redhat-marketplace-74q9j\" (UID: \"d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e\") " pod="openshift-marketplace/redhat-marketplace-74q9j" Feb 17 13:17:30 crc kubenswrapper[4955]: I0217 13:17:30.620039 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87zj4\" (UniqueName: \"kubernetes.io/projected/d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e-kube-api-access-87zj4\") pod \"redhat-marketplace-74q9j\" (UID: \"d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e\") " pod="openshift-marketplace/redhat-marketplace-74q9j" Feb 17 13:17:30 crc kubenswrapper[4955]: I0217 13:17:30.721151 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87zj4\" (UniqueName: \"kubernetes.io/projected/d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e-kube-api-access-87zj4\") pod \"redhat-marketplace-74q9j\" (UID: \"d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e\") " pod="openshift-marketplace/redhat-marketplace-74q9j" Feb 17 13:17:30 crc kubenswrapper[4955]: I0217 13:17:30.721231 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e-utilities\") pod \"redhat-marketplace-74q9j\" (UID: \"d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e\") " pod="openshift-marketplace/redhat-marketplace-74q9j" Feb 17 13:17:30 crc kubenswrapper[4955]: I0217 13:17:30.721294 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e-catalog-content\") pod \"redhat-marketplace-74q9j\" (UID: \"d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e\") " pod="openshift-marketplace/redhat-marketplace-74q9j" Feb 17 13:17:30 crc kubenswrapper[4955]: I0217 13:17:30.721695 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e-utilities\") pod \"redhat-marketplace-74q9j\" (UID: \"d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e\") " pod="openshift-marketplace/redhat-marketplace-74q9j" Feb 17 13:17:30 crc kubenswrapper[4955]: I0217 13:17:30.721744 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e-catalog-content\") pod \"redhat-marketplace-74q9j\" (UID: \"d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e\") " pod="openshift-marketplace/redhat-marketplace-74q9j" Feb 17 13:17:30 crc kubenswrapper[4955]: I0217 13:17:30.748930 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87zj4\" (UniqueName: \"kubernetes.io/projected/d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e-kube-api-access-87zj4\") pod \"redhat-marketplace-74q9j\" (UID: \"d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e\") " pod="openshift-marketplace/redhat-marketplace-74q9j" Feb 17 13:17:30 crc kubenswrapper[4955]: I0217 13:17:30.848358 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-74q9j" Feb 17 13:17:30 crc kubenswrapper[4955]: I0217 13:17:30.852871 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-579d9646f7-zpz6n" event={"ID":"4b384d59-9dc8-4d70-8b12-bd97252d1cfa","Type":"ContainerStarted","Data":"7f3b3aa51cb2aec7effc1dca449da72abceb5f09b398b1a9461b2f8da33219e4"} Feb 17 13:17:31 crc kubenswrapper[4955]: I0217 13:17:31.064336 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-74q9j"] Feb 17 13:17:31 crc kubenswrapper[4955]: I0217 13:17:31.887619 4955 generic.go:334] "Generic (PLEG): container finished" podID="d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e" containerID="aeb0298a4da34417fad495d7625b649529c66d243e7833d8db4c9282527372c0" exitCode=0 Feb 17 13:17:31 crc kubenswrapper[4955]: I0217 13:17:31.887675 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-74q9j" event={"ID":"d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e","Type":"ContainerDied","Data":"aeb0298a4da34417fad495d7625b649529c66d243e7833d8db4c9282527372c0"} Feb 17 13:17:31 crc kubenswrapper[4955]: I0217 13:17:31.887868 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-74q9j" event={"ID":"d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e","Type":"ContainerStarted","Data":"8158dbddea6aa218b2ae968908f1c8c24445dea335e34b34f82e01e6556fc435"} Feb 17 13:17:33 crc kubenswrapper[4955]: I0217 13:17:33.900657 4955 generic.go:334] "Generic (PLEG): container finished" podID="d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e" containerID="aaec3f0d15cb3f7e089252fb3c4966fa88295a5db4ac14e47edccc19da4b564c" exitCode=0 Feb 17 13:17:33 crc kubenswrapper[4955]: I0217 13:17:33.900708 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-74q9j" event={"ID":"d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e","Type":"ContainerDied","Data":"aaec3f0d15cb3f7e089252fb3c4966fa88295a5db4ac14e47edccc19da4b564c"} Feb 17 13:17:33 crc kubenswrapper[4955]: I0217 13:17:33.905397 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-754bdc4867-j8r95" event={"ID":"6f6cf1f8-d54b-4fad-8313-c5f235a9df88","Type":"ContainerStarted","Data":"c7d415da3a6d1e03eafa950599d31d60e3ec99310d45dca50b61c756d1c9d7e2"} Feb 17 13:17:33 crc kubenswrapper[4955]: I0217 13:17:33.906019 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-754bdc4867-j8r95" Feb 17 13:17:33 crc kubenswrapper[4955]: I0217 13:17:33.949736 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-754bdc4867-j8r95" podStartSLOduration=2.46276784 podStartE2EDuration="5.949714481s" podCreationTimestamp="2026-02-17 13:17:28 +0000 UTC" firstStartedPulling="2026-02-17 13:17:29.727718607 +0000 UTC m=+788.250448160" lastFinishedPulling="2026-02-17 13:17:33.214665258 +0000 UTC m=+791.737394801" observedRunningTime="2026-02-17 13:17:33.940417367 +0000 UTC m=+792.463146910" watchObservedRunningTime="2026-02-17 13:17:33.949714481 +0000 UTC m=+792.472444024" Feb 17 13:17:34 crc kubenswrapper[4955]: I0217 13:17:34.913181 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-579d9646f7-zpz6n" event={"ID":"4b384d59-9dc8-4d70-8b12-bd97252d1cfa","Type":"ContainerStarted","Data":"cc1fed30bd72fb2d59309172c7991c540234ccad89a2d41e71e702fb0b3ff8b4"} Feb 17 13:17:34 crc kubenswrapper[4955]: I0217 13:17:34.939130 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-579d9646f7-zpz6n" podStartSLOduration=1.302578494 podStartE2EDuration="5.939115514s" podCreationTimestamp="2026-02-17 13:17:29 +0000 UTC" firstStartedPulling="2026-02-17 13:17:29.860294361 +0000 UTC m=+788.383023894" lastFinishedPulling="2026-02-17 13:17:34.496831371 +0000 UTC m=+793.019560914" observedRunningTime="2026-02-17 13:17:34.936301021 +0000 UTC m=+793.459030604" watchObservedRunningTime="2026-02-17 13:17:34.939115514 +0000 UTC m=+793.461845057" Feb 17 13:17:35 crc kubenswrapper[4955]: I0217 13:17:35.919800 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-74q9j" event={"ID":"d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e","Type":"ContainerStarted","Data":"5ae09e7d69dcfad5c6ee44aff7fa7dffd798d45805285b969e769fce2b77c624"} Feb 17 13:17:35 crc kubenswrapper[4955]: I0217 13:17:35.919939 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-579d9646f7-zpz6n" Feb 17 13:17:35 crc kubenswrapper[4955]: I0217 13:17:35.936821 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-74q9j" podStartSLOduration=2.568679898 podStartE2EDuration="5.9368021s" podCreationTimestamp="2026-02-17 13:17:30 +0000 UTC" firstStartedPulling="2026-02-17 13:17:31.898132194 +0000 UTC m=+790.420861747" lastFinishedPulling="2026-02-17 13:17:35.266254406 +0000 UTC m=+793.788983949" observedRunningTime="2026-02-17 13:17:35.934590045 +0000 UTC m=+794.457319588" watchObservedRunningTime="2026-02-17 13:17:35.9368021 +0000 UTC m=+794.459531643" Feb 17 13:17:40 crc kubenswrapper[4955]: I0217 13:17:40.848839 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-74q9j" Feb 17 13:17:40 crc kubenswrapper[4955]: I0217 13:17:40.851952 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-74q9j" Feb 17 13:17:40 crc kubenswrapper[4955]: I0217 13:17:40.926489 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-74q9j" Feb 17 13:17:41 crc kubenswrapper[4955]: I0217 13:17:41.013158 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-74q9j" Feb 17 13:17:43 crc kubenswrapper[4955]: I0217 13:17:43.901572 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-74q9j"] Feb 17 13:17:43 crc kubenswrapper[4955]: I0217 13:17:43.901841 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-74q9j" podUID="d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e" containerName="registry-server" containerID="cri-o://5ae09e7d69dcfad5c6ee44aff7fa7dffd798d45805285b969e769fce2b77c624" gracePeriod=2 Feb 17 13:17:44 crc kubenswrapper[4955]: I0217 13:17:44.278403 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-74q9j" Feb 17 13:17:44 crc kubenswrapper[4955]: I0217 13:17:44.394072 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87zj4\" (UniqueName: \"kubernetes.io/projected/d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e-kube-api-access-87zj4\") pod \"d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e\" (UID: \"d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e\") " Feb 17 13:17:44 crc kubenswrapper[4955]: I0217 13:17:44.394145 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e-catalog-content\") pod \"d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e\" (UID: \"d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e\") " Feb 17 13:17:44 crc kubenswrapper[4955]: I0217 13:17:44.394277 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e-utilities\") pod \"d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e\" (UID: \"d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e\") " Feb 17 13:17:44 crc kubenswrapper[4955]: I0217 13:17:44.395069 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e-utilities" (OuterVolumeSpecName: "utilities") pod "d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e" (UID: "d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:17:44 crc kubenswrapper[4955]: I0217 13:17:44.406474 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e-kube-api-access-87zj4" (OuterVolumeSpecName: "kube-api-access-87zj4") pod "d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e" (UID: "d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e"). InnerVolumeSpecName "kube-api-access-87zj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:17:44 crc kubenswrapper[4955]: I0217 13:17:44.416905 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e" (UID: "d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:17:44 crc kubenswrapper[4955]: I0217 13:17:44.495564 4955 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 13:17:44 crc kubenswrapper[4955]: I0217 13:17:44.495600 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-87zj4\" (UniqueName: \"kubernetes.io/projected/d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e-kube-api-access-87zj4\") on node \"crc\" DevicePath \"\"" Feb 17 13:17:44 crc kubenswrapper[4955]: I0217 13:17:44.495609 4955 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 13:17:44 crc kubenswrapper[4955]: I0217 13:17:44.984423 4955 generic.go:334] "Generic (PLEG): container finished" podID="d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e" containerID="5ae09e7d69dcfad5c6ee44aff7fa7dffd798d45805285b969e769fce2b77c624" exitCode=0 Feb 17 13:17:44 crc kubenswrapper[4955]: I0217 13:17:44.984457 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-74q9j" event={"ID":"d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e","Type":"ContainerDied","Data":"5ae09e7d69dcfad5c6ee44aff7fa7dffd798d45805285b969e769fce2b77c624"} Feb 17 13:17:44 crc kubenswrapper[4955]: I0217 13:17:44.984752 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-74q9j" event={"ID":"d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e","Type":"ContainerDied","Data":"8158dbddea6aa218b2ae968908f1c8c24445dea335e34b34f82e01e6556fc435"} Feb 17 13:17:44 crc kubenswrapper[4955]: I0217 13:17:44.984771 4955 scope.go:117] "RemoveContainer" containerID="5ae09e7d69dcfad5c6ee44aff7fa7dffd798d45805285b969e769fce2b77c624" Feb 17 13:17:44 crc kubenswrapper[4955]: I0217 13:17:44.984555 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-74q9j" Feb 17 13:17:45 crc kubenswrapper[4955]: I0217 13:17:45.016204 4955 scope.go:117] "RemoveContainer" containerID="aaec3f0d15cb3f7e089252fb3c4966fa88295a5db4ac14e47edccc19da4b564c" Feb 17 13:17:45 crc kubenswrapper[4955]: I0217 13:17:45.035059 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-74q9j"] Feb 17 13:17:45 crc kubenswrapper[4955]: I0217 13:17:45.041564 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-74q9j"] Feb 17 13:17:45 crc kubenswrapper[4955]: I0217 13:17:45.057299 4955 scope.go:117] "RemoveContainer" containerID="aeb0298a4da34417fad495d7625b649529c66d243e7833d8db4c9282527372c0" Feb 17 13:17:45 crc kubenswrapper[4955]: I0217 13:17:45.076530 4955 scope.go:117] "RemoveContainer" containerID="5ae09e7d69dcfad5c6ee44aff7fa7dffd798d45805285b969e769fce2b77c624" Feb 17 13:17:45 crc kubenswrapper[4955]: E0217 13:17:45.077133 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ae09e7d69dcfad5c6ee44aff7fa7dffd798d45805285b969e769fce2b77c624\": container with ID starting with 5ae09e7d69dcfad5c6ee44aff7fa7dffd798d45805285b969e769fce2b77c624 not found: ID does not exist" containerID="5ae09e7d69dcfad5c6ee44aff7fa7dffd798d45805285b969e769fce2b77c624" Feb 17 13:17:45 crc kubenswrapper[4955]: I0217 13:17:45.077190 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ae09e7d69dcfad5c6ee44aff7fa7dffd798d45805285b969e769fce2b77c624"} err="failed to get container status \"5ae09e7d69dcfad5c6ee44aff7fa7dffd798d45805285b969e769fce2b77c624\": rpc error: code = NotFound desc = could not find container \"5ae09e7d69dcfad5c6ee44aff7fa7dffd798d45805285b969e769fce2b77c624\": container with ID starting with 5ae09e7d69dcfad5c6ee44aff7fa7dffd798d45805285b969e769fce2b77c624 not found: ID does not exist" Feb 17 13:17:45 crc kubenswrapper[4955]: I0217 13:17:45.077223 4955 scope.go:117] "RemoveContainer" containerID="aaec3f0d15cb3f7e089252fb3c4966fa88295a5db4ac14e47edccc19da4b564c" Feb 17 13:17:45 crc kubenswrapper[4955]: E0217 13:17:45.077638 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aaec3f0d15cb3f7e089252fb3c4966fa88295a5db4ac14e47edccc19da4b564c\": container with ID starting with aaec3f0d15cb3f7e089252fb3c4966fa88295a5db4ac14e47edccc19da4b564c not found: ID does not exist" containerID="aaec3f0d15cb3f7e089252fb3c4966fa88295a5db4ac14e47edccc19da4b564c" Feb 17 13:17:45 crc kubenswrapper[4955]: I0217 13:17:45.077681 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aaec3f0d15cb3f7e089252fb3c4966fa88295a5db4ac14e47edccc19da4b564c"} err="failed to get container status \"aaec3f0d15cb3f7e089252fb3c4966fa88295a5db4ac14e47edccc19da4b564c\": rpc error: code = NotFound desc = could not find container \"aaec3f0d15cb3f7e089252fb3c4966fa88295a5db4ac14e47edccc19da4b564c\": container with ID starting with aaec3f0d15cb3f7e089252fb3c4966fa88295a5db4ac14e47edccc19da4b564c not found: ID does not exist" Feb 17 13:17:45 crc kubenswrapper[4955]: I0217 13:17:45.077712 4955 scope.go:117] "RemoveContainer" containerID="aeb0298a4da34417fad495d7625b649529c66d243e7833d8db4c9282527372c0" Feb 17 13:17:45 crc kubenswrapper[4955]: E0217 13:17:45.078448 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aeb0298a4da34417fad495d7625b649529c66d243e7833d8db4c9282527372c0\": container with ID starting with aeb0298a4da34417fad495d7625b649529c66d243e7833d8db4c9282527372c0 not found: ID does not exist" containerID="aeb0298a4da34417fad495d7625b649529c66d243e7833d8db4c9282527372c0" Feb 17 13:17:45 crc kubenswrapper[4955]: I0217 13:17:45.078478 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aeb0298a4da34417fad495d7625b649529c66d243e7833d8db4c9282527372c0"} err="failed to get container status \"aeb0298a4da34417fad495d7625b649529c66d243e7833d8db4c9282527372c0\": rpc error: code = NotFound desc = could not find container \"aeb0298a4da34417fad495d7625b649529c66d243e7833d8db4c9282527372c0\": container with ID starting with aeb0298a4da34417fad495d7625b649529c66d243e7833d8db4c9282527372c0 not found: ID does not exist" Feb 17 13:17:46 crc kubenswrapper[4955]: I0217 13:17:46.232841 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e" path="/var/lib/kubelet/pods/d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e/volumes" Feb 17 13:17:46 crc kubenswrapper[4955]: I0217 13:17:46.710096 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tp9jf"] Feb 17 13:17:46 crc kubenswrapper[4955]: E0217 13:17:46.710559 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e" containerName="extract-utilities" Feb 17 13:17:46 crc kubenswrapper[4955]: I0217 13:17:46.710569 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e" containerName="extract-utilities" Feb 17 13:17:46 crc kubenswrapper[4955]: E0217 13:17:46.710582 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e" containerName="extract-content" Feb 17 13:17:46 crc kubenswrapper[4955]: I0217 13:17:46.710587 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e" containerName="extract-content" Feb 17 13:17:46 crc kubenswrapper[4955]: E0217 13:17:46.710607 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e" containerName="registry-server" Feb 17 13:17:46 crc kubenswrapper[4955]: I0217 13:17:46.710614 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e" containerName="registry-server" Feb 17 13:17:46 crc kubenswrapper[4955]: I0217 13:17:46.710701 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8ee46fe-24fd-432c-ab31-9cf3d0a55c3e" containerName="registry-server" Feb 17 13:17:46 crc kubenswrapper[4955]: I0217 13:17:46.711557 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tp9jf" Feb 17 13:17:46 crc kubenswrapper[4955]: I0217 13:17:46.722716 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tp9jf"] Feb 17 13:17:46 crc kubenswrapper[4955]: I0217 13:17:46.832482 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxkht\" (UniqueName: \"kubernetes.io/projected/80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0-kube-api-access-lxkht\") pod \"community-operators-tp9jf\" (UID: \"80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0\") " pod="openshift-marketplace/community-operators-tp9jf" Feb 17 13:17:46 crc kubenswrapper[4955]: I0217 13:17:46.832541 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0-catalog-content\") pod \"community-operators-tp9jf\" (UID: \"80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0\") " pod="openshift-marketplace/community-operators-tp9jf" Feb 17 13:17:46 crc kubenswrapper[4955]: I0217 13:17:46.832576 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0-utilities\") pod \"community-operators-tp9jf\" (UID: \"80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0\") " pod="openshift-marketplace/community-operators-tp9jf" Feb 17 13:17:46 crc kubenswrapper[4955]: I0217 13:17:46.934378 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0-utilities\") pod \"community-operators-tp9jf\" (UID: \"80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0\") " pod="openshift-marketplace/community-operators-tp9jf" Feb 17 13:17:46 crc kubenswrapper[4955]: I0217 13:17:46.934483 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxkht\" (UniqueName: \"kubernetes.io/projected/80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0-kube-api-access-lxkht\") pod \"community-operators-tp9jf\" (UID: \"80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0\") " pod="openshift-marketplace/community-operators-tp9jf" Feb 17 13:17:46 crc kubenswrapper[4955]: I0217 13:17:46.934527 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0-catalog-content\") pod \"community-operators-tp9jf\" (UID: \"80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0\") " pod="openshift-marketplace/community-operators-tp9jf" Feb 17 13:17:46 crc kubenswrapper[4955]: I0217 13:17:46.935149 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0-utilities\") pod \"community-operators-tp9jf\" (UID: \"80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0\") " pod="openshift-marketplace/community-operators-tp9jf" Feb 17 13:17:46 crc kubenswrapper[4955]: I0217 13:17:46.935200 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0-catalog-content\") pod \"community-operators-tp9jf\" (UID: \"80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0\") " pod="openshift-marketplace/community-operators-tp9jf" Feb 17 13:17:46 crc kubenswrapper[4955]: I0217 13:17:46.956544 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxkht\" (UniqueName: \"kubernetes.io/projected/80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0-kube-api-access-lxkht\") pod \"community-operators-tp9jf\" (UID: \"80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0\") " pod="openshift-marketplace/community-operators-tp9jf" Feb 17 13:17:47 crc kubenswrapper[4955]: I0217 13:17:47.088213 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tp9jf" Feb 17 13:17:47 crc kubenswrapper[4955]: I0217 13:17:47.545436 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tp9jf"] Feb 17 13:17:48 crc kubenswrapper[4955]: I0217 13:17:48.007700 4955 generic.go:334] "Generic (PLEG): container finished" podID="80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0" containerID="8d1eeb682cf9c25a06da22dc4734e7d28b7783e1ade9b3dc0ecdaf9fc864434a" exitCode=0 Feb 17 13:17:48 crc kubenswrapper[4955]: I0217 13:17:48.007759 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tp9jf" event={"ID":"80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0","Type":"ContainerDied","Data":"8d1eeb682cf9c25a06da22dc4734e7d28b7783e1ade9b3dc0ecdaf9fc864434a"} Feb 17 13:17:48 crc kubenswrapper[4955]: I0217 13:17:48.008050 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tp9jf" event={"ID":"80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0","Type":"ContainerStarted","Data":"48b3e5dc65b9573e471d17468dc78cbfabe09f3ffb3a9134c364fc67c064930d"} Feb 17 13:17:49 crc kubenswrapper[4955]: I0217 13:17:49.027547 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tp9jf" event={"ID":"80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0","Type":"ContainerStarted","Data":"c1f83dcb569459cb865852d0653ee939b20185dee52077d6b344c2b6c5b7e7f0"} Feb 17 13:17:49 crc kubenswrapper[4955]: I0217 13:17:49.637766 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-579d9646f7-zpz6n" Feb 17 13:17:50 crc kubenswrapper[4955]: I0217 13:17:50.038120 4955 generic.go:334] "Generic (PLEG): container finished" podID="80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0" containerID="c1f83dcb569459cb865852d0653ee939b20185dee52077d6b344c2b6c5b7e7f0" exitCode=0 Feb 17 13:17:50 crc kubenswrapper[4955]: I0217 13:17:50.038209 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tp9jf" event={"ID":"80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0","Type":"ContainerDied","Data":"c1f83dcb569459cb865852d0653ee939b20185dee52077d6b344c2b6c5b7e7f0"} Feb 17 13:17:51 crc kubenswrapper[4955]: I0217 13:17:51.044765 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tp9jf" event={"ID":"80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0","Type":"ContainerStarted","Data":"590d01075fc8c25cb10b4f419aaaff98c5f31bd2ca43ab55dc3ab8e3570ed0cb"} Feb 17 13:17:57 crc kubenswrapper[4955]: I0217 13:17:57.088845 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tp9jf" Feb 17 13:17:57 crc kubenswrapper[4955]: I0217 13:17:57.089170 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tp9jf" Feb 17 13:17:57 crc kubenswrapper[4955]: I0217 13:17:57.144000 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tp9jf" Feb 17 13:17:57 crc kubenswrapper[4955]: I0217 13:17:57.168283 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tp9jf" podStartSLOduration=8.71631801 podStartE2EDuration="11.168265087s" podCreationTimestamp="2026-02-17 13:17:46 +0000 UTC" firstStartedPulling="2026-02-17 13:17:48.010234634 +0000 UTC m=+806.532964207" lastFinishedPulling="2026-02-17 13:17:50.462181741 +0000 UTC m=+808.984911284" observedRunningTime="2026-02-17 13:17:51.068240805 +0000 UTC m=+809.590970348" watchObservedRunningTime="2026-02-17 13:17:57.168265087 +0000 UTC m=+815.690994640" Feb 17 13:17:58 crc kubenswrapper[4955]: I0217 13:17:58.152954 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tp9jf" Feb 17 13:17:58 crc kubenswrapper[4955]: I0217 13:17:58.210756 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tp9jf"] Feb 17 13:18:00 crc kubenswrapper[4955]: I0217 13:18:00.100033 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tp9jf" podUID="80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0" containerName="registry-server" containerID="cri-o://590d01075fc8c25cb10b4f419aaaff98c5f31bd2ca43ab55dc3ab8e3570ed0cb" gracePeriod=2 Feb 17 13:18:01 crc kubenswrapper[4955]: I0217 13:18:01.010945 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tp9jf" Feb 17 13:18:01 crc kubenswrapper[4955]: I0217 13:18:01.108120 4955 generic.go:334] "Generic (PLEG): container finished" podID="80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0" containerID="590d01075fc8c25cb10b4f419aaaff98c5f31bd2ca43ab55dc3ab8e3570ed0cb" exitCode=0 Feb 17 13:18:01 crc kubenswrapper[4955]: I0217 13:18:01.108155 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tp9jf" event={"ID":"80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0","Type":"ContainerDied","Data":"590d01075fc8c25cb10b4f419aaaff98c5f31bd2ca43ab55dc3ab8e3570ed0cb"} Feb 17 13:18:01 crc kubenswrapper[4955]: I0217 13:18:01.108162 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tp9jf" Feb 17 13:18:01 crc kubenswrapper[4955]: I0217 13:18:01.108178 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tp9jf" event={"ID":"80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0","Type":"ContainerDied","Data":"48b3e5dc65b9573e471d17468dc78cbfabe09f3ffb3a9134c364fc67c064930d"} Feb 17 13:18:01 crc kubenswrapper[4955]: I0217 13:18:01.108193 4955 scope.go:117] "RemoveContainer" containerID="590d01075fc8c25cb10b4f419aaaff98c5f31bd2ca43ab55dc3ab8e3570ed0cb" Feb 17 13:18:01 crc kubenswrapper[4955]: I0217 13:18:01.125218 4955 scope.go:117] "RemoveContainer" containerID="c1f83dcb569459cb865852d0653ee939b20185dee52077d6b344c2b6c5b7e7f0" Feb 17 13:18:01 crc kubenswrapper[4955]: I0217 13:18:01.139257 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0-utilities\") pod \"80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0\" (UID: \"80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0\") " Feb 17 13:18:01 crc kubenswrapper[4955]: I0217 13:18:01.139490 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0-catalog-content\") pod \"80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0\" (UID: \"80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0\") " Feb 17 13:18:01 crc kubenswrapper[4955]: I0217 13:18:01.139728 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxkht\" (UniqueName: \"kubernetes.io/projected/80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0-kube-api-access-lxkht\") pod \"80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0\" (UID: \"80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0\") " Feb 17 13:18:01 crc kubenswrapper[4955]: I0217 13:18:01.142127 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0-utilities" (OuterVolumeSpecName: "utilities") pod "80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0" (UID: "80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:18:01 crc kubenswrapper[4955]: I0217 13:18:01.143016 4955 scope.go:117] "RemoveContainer" containerID="8d1eeb682cf9c25a06da22dc4734e7d28b7783e1ade9b3dc0ecdaf9fc864434a" Feb 17 13:18:01 crc kubenswrapper[4955]: I0217 13:18:01.148480 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0-kube-api-access-lxkht" (OuterVolumeSpecName: "kube-api-access-lxkht") pod "80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0" (UID: "80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0"). InnerVolumeSpecName "kube-api-access-lxkht". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:18:01 crc kubenswrapper[4955]: I0217 13:18:01.186484 4955 scope.go:117] "RemoveContainer" containerID="590d01075fc8c25cb10b4f419aaaff98c5f31bd2ca43ab55dc3ab8e3570ed0cb" Feb 17 13:18:01 crc kubenswrapper[4955]: E0217 13:18:01.187031 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"590d01075fc8c25cb10b4f419aaaff98c5f31bd2ca43ab55dc3ab8e3570ed0cb\": container with ID starting with 590d01075fc8c25cb10b4f419aaaff98c5f31bd2ca43ab55dc3ab8e3570ed0cb not found: ID does not exist" containerID="590d01075fc8c25cb10b4f419aaaff98c5f31bd2ca43ab55dc3ab8e3570ed0cb" Feb 17 13:18:01 crc kubenswrapper[4955]: I0217 13:18:01.187093 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"590d01075fc8c25cb10b4f419aaaff98c5f31bd2ca43ab55dc3ab8e3570ed0cb"} err="failed to get container status \"590d01075fc8c25cb10b4f419aaaff98c5f31bd2ca43ab55dc3ab8e3570ed0cb\": rpc error: code = NotFound desc = could not find container \"590d01075fc8c25cb10b4f419aaaff98c5f31bd2ca43ab55dc3ab8e3570ed0cb\": container with ID starting with 590d01075fc8c25cb10b4f419aaaff98c5f31bd2ca43ab55dc3ab8e3570ed0cb not found: ID does not exist" Feb 17 13:18:01 crc kubenswrapper[4955]: I0217 13:18:01.187133 4955 scope.go:117] "RemoveContainer" containerID="c1f83dcb569459cb865852d0653ee939b20185dee52077d6b344c2b6c5b7e7f0" Feb 17 13:18:01 crc kubenswrapper[4955]: E0217 13:18:01.187538 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1f83dcb569459cb865852d0653ee939b20185dee52077d6b344c2b6c5b7e7f0\": container with ID starting with c1f83dcb569459cb865852d0653ee939b20185dee52077d6b344c2b6c5b7e7f0 not found: ID does not exist" containerID="c1f83dcb569459cb865852d0653ee939b20185dee52077d6b344c2b6c5b7e7f0" Feb 17 13:18:01 crc kubenswrapper[4955]: I0217 13:18:01.187582 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1f83dcb569459cb865852d0653ee939b20185dee52077d6b344c2b6c5b7e7f0"} err="failed to get container status \"c1f83dcb569459cb865852d0653ee939b20185dee52077d6b344c2b6c5b7e7f0\": rpc error: code = NotFound desc = could not find container \"c1f83dcb569459cb865852d0653ee939b20185dee52077d6b344c2b6c5b7e7f0\": container with ID starting with c1f83dcb569459cb865852d0653ee939b20185dee52077d6b344c2b6c5b7e7f0 not found: ID does not exist" Feb 17 13:18:01 crc kubenswrapper[4955]: I0217 13:18:01.187612 4955 scope.go:117] "RemoveContainer" containerID="8d1eeb682cf9c25a06da22dc4734e7d28b7783e1ade9b3dc0ecdaf9fc864434a" Feb 17 13:18:01 crc kubenswrapper[4955]: E0217 13:18:01.188107 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d1eeb682cf9c25a06da22dc4734e7d28b7783e1ade9b3dc0ecdaf9fc864434a\": container with ID starting with 8d1eeb682cf9c25a06da22dc4734e7d28b7783e1ade9b3dc0ecdaf9fc864434a not found: ID does not exist" containerID="8d1eeb682cf9c25a06da22dc4734e7d28b7783e1ade9b3dc0ecdaf9fc864434a" Feb 17 13:18:01 crc kubenswrapper[4955]: I0217 13:18:01.188141 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d1eeb682cf9c25a06da22dc4734e7d28b7783e1ade9b3dc0ecdaf9fc864434a"} err="failed to get container status \"8d1eeb682cf9c25a06da22dc4734e7d28b7783e1ade9b3dc0ecdaf9fc864434a\": rpc error: code = NotFound desc = could not find container \"8d1eeb682cf9c25a06da22dc4734e7d28b7783e1ade9b3dc0ecdaf9fc864434a\": container with ID starting with 8d1eeb682cf9c25a06da22dc4734e7d28b7783e1ade9b3dc0ecdaf9fc864434a not found: ID does not exist" Feb 17 13:18:01 crc kubenswrapper[4955]: I0217 13:18:01.210036 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0" (UID: "80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:18:01 crc kubenswrapper[4955]: I0217 13:18:01.241841 4955 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 13:18:01 crc kubenswrapper[4955]: I0217 13:18:01.241932 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxkht\" (UniqueName: \"kubernetes.io/projected/80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0-kube-api-access-lxkht\") on node \"crc\" DevicePath \"\"" Feb 17 13:18:01 crc kubenswrapper[4955]: I0217 13:18:01.241960 4955 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 13:18:01 crc kubenswrapper[4955]: I0217 13:18:01.436829 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tp9jf"] Feb 17 13:18:01 crc kubenswrapper[4955]: I0217 13:18:01.440027 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tp9jf"] Feb 17 13:18:02 crc kubenswrapper[4955]: I0217 13:18:02.235517 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0" path="/var/lib/kubelet/pods/80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0/volumes" Feb 17 13:18:09 crc kubenswrapper[4955]: I0217 13:18:09.265705 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-754bdc4867-j8r95" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.098855 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-vzhr7"] Feb 17 13:18:10 crc kubenswrapper[4955]: E0217 13:18:10.099083 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0" containerName="extract-content" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.099098 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0" containerName="extract-content" Feb 17 13:18:10 crc kubenswrapper[4955]: E0217 13:18:10.099109 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0" containerName="registry-server" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.099117 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0" containerName="registry-server" Feb 17 13:18:10 crc kubenswrapper[4955]: E0217 13:18:10.099128 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0" containerName="extract-utilities" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.099136 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0" containerName="extract-utilities" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.099271 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="80ee7327-f65e-4b7f-8bdd-f31bf4ca91c0" containerName="registry-server" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.101471 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-vzhr7" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.109396 4955 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.109609 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.109515 4955 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-mvzlz" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.112235 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-78b44bf5bb-rcv4v"] Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.113002 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-rcv4v" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.118660 4955 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.127822 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-78b44bf5bb-rcv4v"] Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.193119 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/720ac59e-55ea-44be-b7af-9b57fec4901b-frr-conf\") pod \"frr-k8s-vzhr7\" (UID: \"720ac59e-55ea-44be-b7af-9b57fec4901b\") " pod="metallb-system/frr-k8s-vzhr7" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.193171 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/720ac59e-55ea-44be-b7af-9b57fec4901b-frr-sockets\") pod \"frr-k8s-vzhr7\" (UID: \"720ac59e-55ea-44be-b7af-9b57fec4901b\") " pod="metallb-system/frr-k8s-vzhr7" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.193197 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/92393c39-fc3d-490c-93a3-c11ddef89c02-cert\") pod \"frr-k8s-webhook-server-78b44bf5bb-rcv4v\" (UID: \"92393c39-fc3d-490c-93a3-c11ddef89c02\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-rcv4v" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.193368 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/720ac59e-55ea-44be-b7af-9b57fec4901b-frr-startup\") pod \"frr-k8s-vzhr7\" (UID: \"720ac59e-55ea-44be-b7af-9b57fec4901b\") " pod="metallb-system/frr-k8s-vzhr7" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.193435 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/720ac59e-55ea-44be-b7af-9b57fec4901b-metrics\") pod \"frr-k8s-vzhr7\" (UID: \"720ac59e-55ea-44be-b7af-9b57fec4901b\") " pod="metallb-system/frr-k8s-vzhr7" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.193458 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/720ac59e-55ea-44be-b7af-9b57fec4901b-reloader\") pod \"frr-k8s-vzhr7\" (UID: \"720ac59e-55ea-44be-b7af-9b57fec4901b\") " pod="metallb-system/frr-k8s-vzhr7" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.193471 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/720ac59e-55ea-44be-b7af-9b57fec4901b-metrics-certs\") pod \"frr-k8s-vzhr7\" (UID: \"720ac59e-55ea-44be-b7af-9b57fec4901b\") " pod="metallb-system/frr-k8s-vzhr7" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.193492 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-645mk\" (UniqueName: \"kubernetes.io/projected/720ac59e-55ea-44be-b7af-9b57fec4901b-kube-api-access-645mk\") pod \"frr-k8s-vzhr7\" (UID: \"720ac59e-55ea-44be-b7af-9b57fec4901b\") " pod="metallb-system/frr-k8s-vzhr7" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.193530 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5l4vb\" (UniqueName: \"kubernetes.io/projected/92393c39-fc3d-490c-93a3-c11ddef89c02-kube-api-access-5l4vb\") pod \"frr-k8s-webhook-server-78b44bf5bb-rcv4v\" (UID: \"92393c39-fc3d-490c-93a3-c11ddef89c02\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-rcv4v" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.255846 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-dl5vp"] Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.256628 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-dl5vp" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.258949 4955 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-dcd6w" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.259222 4955 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.261668 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.262427 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-69bbfbf88f-pvzn9"] Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.262810 4955 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.263287 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-69bbfbf88f-pvzn9" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.272123 4955 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.294384 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwdp6\" (UniqueName: \"kubernetes.io/projected/618eb288-1aa4-43a9-b26a-9e2f5378411d-kube-api-access-kwdp6\") pod \"speaker-dl5vp\" (UID: \"618eb288-1aa4-43a9-b26a-9e2f5378411d\") " pod="metallb-system/speaker-dl5vp" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.294445 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/720ac59e-55ea-44be-b7af-9b57fec4901b-frr-conf\") pod \"frr-k8s-vzhr7\" (UID: \"720ac59e-55ea-44be-b7af-9b57fec4901b\") " pod="metallb-system/frr-k8s-vzhr7" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.294464 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/618eb288-1aa4-43a9-b26a-9e2f5378411d-memberlist\") pod \"speaker-dl5vp\" (UID: \"618eb288-1aa4-43a9-b26a-9e2f5378411d\") " pod="metallb-system/speaker-dl5vp" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.294481 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/720ac59e-55ea-44be-b7af-9b57fec4901b-frr-sockets\") pod \"frr-k8s-vzhr7\" (UID: \"720ac59e-55ea-44be-b7af-9b57fec4901b\") " pod="metallb-system/frr-k8s-vzhr7" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.294496 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/92393c39-fc3d-490c-93a3-c11ddef89c02-cert\") pod \"frr-k8s-webhook-server-78b44bf5bb-rcv4v\" (UID: \"92393c39-fc3d-490c-93a3-c11ddef89c02\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-rcv4v" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.294513 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8c11830d-6699-432a-a2d3-14b135bc1665-cert\") pod \"controller-69bbfbf88f-pvzn9\" (UID: \"8c11830d-6699-432a-a2d3-14b135bc1665\") " pod="metallb-system/controller-69bbfbf88f-pvzn9" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.294545 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8c11830d-6699-432a-a2d3-14b135bc1665-metrics-certs\") pod \"controller-69bbfbf88f-pvzn9\" (UID: \"8c11830d-6699-432a-a2d3-14b135bc1665\") " pod="metallb-system/controller-69bbfbf88f-pvzn9" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.294561 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/618eb288-1aa4-43a9-b26a-9e2f5378411d-metallb-excludel2\") pod \"speaker-dl5vp\" (UID: \"618eb288-1aa4-43a9-b26a-9e2f5378411d\") " pod="metallb-system/speaker-dl5vp" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.294577 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/720ac59e-55ea-44be-b7af-9b57fec4901b-frr-startup\") pod \"frr-k8s-vzhr7\" (UID: \"720ac59e-55ea-44be-b7af-9b57fec4901b\") " pod="metallb-system/frr-k8s-vzhr7" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.294601 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/720ac59e-55ea-44be-b7af-9b57fec4901b-metrics\") pod \"frr-k8s-vzhr7\" (UID: \"720ac59e-55ea-44be-b7af-9b57fec4901b\") " pod="metallb-system/frr-k8s-vzhr7" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.294617 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/720ac59e-55ea-44be-b7af-9b57fec4901b-reloader\") pod \"frr-k8s-vzhr7\" (UID: \"720ac59e-55ea-44be-b7af-9b57fec4901b\") " pod="metallb-system/frr-k8s-vzhr7" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.294631 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/720ac59e-55ea-44be-b7af-9b57fec4901b-metrics-certs\") pod \"frr-k8s-vzhr7\" (UID: \"720ac59e-55ea-44be-b7af-9b57fec4901b\") " pod="metallb-system/frr-k8s-vzhr7" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.294647 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-645mk\" (UniqueName: \"kubernetes.io/projected/720ac59e-55ea-44be-b7af-9b57fec4901b-kube-api-access-645mk\") pod \"frr-k8s-vzhr7\" (UID: \"720ac59e-55ea-44be-b7af-9b57fec4901b\") " pod="metallb-system/frr-k8s-vzhr7" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.294668 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tmfz\" (UniqueName: \"kubernetes.io/projected/8c11830d-6699-432a-a2d3-14b135bc1665-kube-api-access-6tmfz\") pod \"controller-69bbfbf88f-pvzn9\" (UID: \"8c11830d-6699-432a-a2d3-14b135bc1665\") " pod="metallb-system/controller-69bbfbf88f-pvzn9" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.294687 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5l4vb\" (UniqueName: \"kubernetes.io/projected/92393c39-fc3d-490c-93a3-c11ddef89c02-kube-api-access-5l4vb\") pod \"frr-k8s-webhook-server-78b44bf5bb-rcv4v\" (UID: \"92393c39-fc3d-490c-93a3-c11ddef89c02\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-rcv4v" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.294705 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/618eb288-1aa4-43a9-b26a-9e2f5378411d-metrics-certs\") pod \"speaker-dl5vp\" (UID: \"618eb288-1aa4-43a9-b26a-9e2f5378411d\") " pod="metallb-system/speaker-dl5vp" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.295075 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/720ac59e-55ea-44be-b7af-9b57fec4901b-frr-conf\") pod \"frr-k8s-vzhr7\" (UID: \"720ac59e-55ea-44be-b7af-9b57fec4901b\") " pod="metallb-system/frr-k8s-vzhr7" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.295255 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/720ac59e-55ea-44be-b7af-9b57fec4901b-metrics\") pod \"frr-k8s-vzhr7\" (UID: \"720ac59e-55ea-44be-b7af-9b57fec4901b\") " pod="metallb-system/frr-k8s-vzhr7" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.295413 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/720ac59e-55ea-44be-b7af-9b57fec4901b-reloader\") pod \"frr-k8s-vzhr7\" (UID: \"720ac59e-55ea-44be-b7af-9b57fec4901b\") " pod="metallb-system/frr-k8s-vzhr7" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.295952 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/720ac59e-55ea-44be-b7af-9b57fec4901b-frr-startup\") pod \"frr-k8s-vzhr7\" (UID: \"720ac59e-55ea-44be-b7af-9b57fec4901b\") " pod="metallb-system/frr-k8s-vzhr7" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.296139 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/720ac59e-55ea-44be-b7af-9b57fec4901b-frr-sockets\") pod \"frr-k8s-vzhr7\" (UID: \"720ac59e-55ea-44be-b7af-9b57fec4901b\") " pod="metallb-system/frr-k8s-vzhr7" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.302022 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/720ac59e-55ea-44be-b7af-9b57fec4901b-metrics-certs\") pod \"frr-k8s-vzhr7\" (UID: \"720ac59e-55ea-44be-b7af-9b57fec4901b\") " pod="metallb-system/frr-k8s-vzhr7" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.302532 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/92393c39-fc3d-490c-93a3-c11ddef89c02-cert\") pod \"frr-k8s-webhook-server-78b44bf5bb-rcv4v\" (UID: \"92393c39-fc3d-490c-93a3-c11ddef89c02\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-rcv4v" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.304129 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-69bbfbf88f-pvzn9"] Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.333310 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-645mk\" (UniqueName: \"kubernetes.io/projected/720ac59e-55ea-44be-b7af-9b57fec4901b-kube-api-access-645mk\") pod \"frr-k8s-vzhr7\" (UID: \"720ac59e-55ea-44be-b7af-9b57fec4901b\") " pod="metallb-system/frr-k8s-vzhr7" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.342442 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5l4vb\" (UniqueName: \"kubernetes.io/projected/92393c39-fc3d-490c-93a3-c11ddef89c02-kube-api-access-5l4vb\") pod \"frr-k8s-webhook-server-78b44bf5bb-rcv4v\" (UID: \"92393c39-fc3d-490c-93a3-c11ddef89c02\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-rcv4v" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.395722 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/618eb288-1aa4-43a9-b26a-9e2f5378411d-memberlist\") pod \"speaker-dl5vp\" (UID: \"618eb288-1aa4-43a9-b26a-9e2f5378411d\") " pod="metallb-system/speaker-dl5vp" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.395770 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8c11830d-6699-432a-a2d3-14b135bc1665-cert\") pod \"controller-69bbfbf88f-pvzn9\" (UID: \"8c11830d-6699-432a-a2d3-14b135bc1665\") " pod="metallb-system/controller-69bbfbf88f-pvzn9" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.395823 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8c11830d-6699-432a-a2d3-14b135bc1665-metrics-certs\") pod \"controller-69bbfbf88f-pvzn9\" (UID: \"8c11830d-6699-432a-a2d3-14b135bc1665\") " pod="metallb-system/controller-69bbfbf88f-pvzn9" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.395844 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/618eb288-1aa4-43a9-b26a-9e2f5378411d-metallb-excludel2\") pod \"speaker-dl5vp\" (UID: \"618eb288-1aa4-43a9-b26a-9e2f5378411d\") " pod="metallb-system/speaker-dl5vp" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.395881 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tmfz\" (UniqueName: \"kubernetes.io/projected/8c11830d-6699-432a-a2d3-14b135bc1665-kube-api-access-6tmfz\") pod \"controller-69bbfbf88f-pvzn9\" (UID: \"8c11830d-6699-432a-a2d3-14b135bc1665\") " pod="metallb-system/controller-69bbfbf88f-pvzn9" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.395899 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/618eb288-1aa4-43a9-b26a-9e2f5378411d-metrics-certs\") pod \"speaker-dl5vp\" (UID: \"618eb288-1aa4-43a9-b26a-9e2f5378411d\") " pod="metallb-system/speaker-dl5vp" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.395928 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwdp6\" (UniqueName: \"kubernetes.io/projected/618eb288-1aa4-43a9-b26a-9e2f5378411d-kube-api-access-kwdp6\") pod \"speaker-dl5vp\" (UID: \"618eb288-1aa4-43a9-b26a-9e2f5378411d\") " pod="metallb-system/speaker-dl5vp" Feb 17 13:18:10 crc kubenswrapper[4955]: E0217 13:18:10.395926 4955 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Feb 17 13:18:10 crc kubenswrapper[4955]: E0217 13:18:10.396010 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/618eb288-1aa4-43a9-b26a-9e2f5378411d-memberlist podName:618eb288-1aa4-43a9-b26a-9e2f5378411d nodeName:}" failed. No retries permitted until 2026-02-17 13:18:10.895984709 +0000 UTC m=+829.418714352 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/618eb288-1aa4-43a9-b26a-9e2f5378411d-memberlist") pod "speaker-dl5vp" (UID: "618eb288-1aa4-43a9-b26a-9e2f5378411d") : secret "metallb-memberlist" not found Feb 17 13:18:10 crc kubenswrapper[4955]: E0217 13:18:10.396023 4955 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Feb 17 13:18:10 crc kubenswrapper[4955]: E0217 13:18:10.396115 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8c11830d-6699-432a-a2d3-14b135bc1665-metrics-certs podName:8c11830d-6699-432a-a2d3-14b135bc1665 nodeName:}" failed. No retries permitted until 2026-02-17 13:18:10.896084602 +0000 UTC m=+829.418814145 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8c11830d-6699-432a-a2d3-14b135bc1665-metrics-certs") pod "controller-69bbfbf88f-pvzn9" (UID: "8c11830d-6699-432a-a2d3-14b135bc1665") : secret "controller-certs-secret" not found Feb 17 13:18:10 crc kubenswrapper[4955]: E0217 13:18:10.396284 4955 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Feb 17 13:18:10 crc kubenswrapper[4955]: E0217 13:18:10.396328 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/618eb288-1aa4-43a9-b26a-9e2f5378411d-metrics-certs podName:618eb288-1aa4-43a9-b26a-9e2f5378411d nodeName:}" failed. No retries permitted until 2026-02-17 13:18:10.896313459 +0000 UTC m=+829.419043002 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/618eb288-1aa4-43a9-b26a-9e2f5378411d-metrics-certs") pod "speaker-dl5vp" (UID: "618eb288-1aa4-43a9-b26a-9e2f5378411d") : secret "speaker-certs-secret" not found Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.396914 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/618eb288-1aa4-43a9-b26a-9e2f5378411d-metallb-excludel2\") pod \"speaker-dl5vp\" (UID: \"618eb288-1aa4-43a9-b26a-9e2f5378411d\") " pod="metallb-system/speaker-dl5vp" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.400569 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8c11830d-6699-432a-a2d3-14b135bc1665-cert\") pod \"controller-69bbfbf88f-pvzn9\" (UID: \"8c11830d-6699-432a-a2d3-14b135bc1665\") " pod="metallb-system/controller-69bbfbf88f-pvzn9" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.412995 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwdp6\" (UniqueName: \"kubernetes.io/projected/618eb288-1aa4-43a9-b26a-9e2f5378411d-kube-api-access-kwdp6\") pod \"speaker-dl5vp\" (UID: \"618eb288-1aa4-43a9-b26a-9e2f5378411d\") " pod="metallb-system/speaker-dl5vp" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.415877 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tmfz\" (UniqueName: \"kubernetes.io/projected/8c11830d-6699-432a-a2d3-14b135bc1665-kube-api-access-6tmfz\") pod \"controller-69bbfbf88f-pvzn9\" (UID: \"8c11830d-6699-432a-a2d3-14b135bc1665\") " pod="metallb-system/controller-69bbfbf88f-pvzn9" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.421064 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-vzhr7" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.426884 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-rcv4v" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.651390 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-78b44bf5bb-rcv4v"] Feb 17 13:18:10 crc kubenswrapper[4955]: W0217 13:18:10.672346 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod92393c39_fc3d_490c_93a3_c11ddef89c02.slice/crio-222c9c34788f98f31ab6660803d66591a700ffc499b8d1030f419d9c401d6eca WatchSource:0}: Error finding container 222c9c34788f98f31ab6660803d66591a700ffc499b8d1030f419d9c401d6eca: Status 404 returned error can't find the container with id 222c9c34788f98f31ab6660803d66591a700ffc499b8d1030f419d9c401d6eca Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.904841 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/618eb288-1aa4-43a9-b26a-9e2f5378411d-metrics-certs\") pod \"speaker-dl5vp\" (UID: \"618eb288-1aa4-43a9-b26a-9e2f5378411d\") " pod="metallb-system/speaker-dl5vp" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.904979 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/618eb288-1aa4-43a9-b26a-9e2f5378411d-memberlist\") pod \"speaker-dl5vp\" (UID: \"618eb288-1aa4-43a9-b26a-9e2f5378411d\") " pod="metallb-system/speaker-dl5vp" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.905106 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8c11830d-6699-432a-a2d3-14b135bc1665-metrics-certs\") pod \"controller-69bbfbf88f-pvzn9\" (UID: \"8c11830d-6699-432a-a2d3-14b135bc1665\") " pod="metallb-system/controller-69bbfbf88f-pvzn9" Feb 17 13:18:10 crc kubenswrapper[4955]: E0217 13:18:10.905988 4955 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Feb 17 13:18:10 crc kubenswrapper[4955]: E0217 13:18:10.906093 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/618eb288-1aa4-43a9-b26a-9e2f5378411d-memberlist podName:618eb288-1aa4-43a9-b26a-9e2f5378411d nodeName:}" failed. No retries permitted until 2026-02-17 13:18:11.906068478 +0000 UTC m=+830.428798061 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/618eb288-1aa4-43a9-b26a-9e2f5378411d-memberlist") pod "speaker-dl5vp" (UID: "618eb288-1aa4-43a9-b26a-9e2f5378411d") : secret "metallb-memberlist" not found Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.914152 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8c11830d-6699-432a-a2d3-14b135bc1665-metrics-certs\") pod \"controller-69bbfbf88f-pvzn9\" (UID: \"8c11830d-6699-432a-a2d3-14b135bc1665\") " pod="metallb-system/controller-69bbfbf88f-pvzn9" Feb 17 13:18:10 crc kubenswrapper[4955]: I0217 13:18:10.914239 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/618eb288-1aa4-43a9-b26a-9e2f5378411d-metrics-certs\") pod \"speaker-dl5vp\" (UID: \"618eb288-1aa4-43a9-b26a-9e2f5378411d\") " pod="metallb-system/speaker-dl5vp" Feb 17 13:18:11 crc kubenswrapper[4955]: I0217 13:18:11.183668 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vzhr7" event={"ID":"720ac59e-55ea-44be-b7af-9b57fec4901b","Type":"ContainerStarted","Data":"68cc82a83c3352c0fbf982523e9f85772d3e235b9d193a06945e2adb36aec7cf"} Feb 17 13:18:11 crc kubenswrapper[4955]: I0217 13:18:11.184980 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-rcv4v" event={"ID":"92393c39-fc3d-490c-93a3-c11ddef89c02","Type":"ContainerStarted","Data":"222c9c34788f98f31ab6660803d66591a700ffc499b8d1030f419d9c401d6eca"} Feb 17 13:18:11 crc kubenswrapper[4955]: I0217 13:18:11.194541 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-69bbfbf88f-pvzn9" Feb 17 13:18:11 crc kubenswrapper[4955]: I0217 13:18:11.676543 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-69bbfbf88f-pvzn9"] Feb 17 13:18:11 crc kubenswrapper[4955]: W0217 13:18:11.702818 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c11830d_6699_432a_a2d3_14b135bc1665.slice/crio-39836f1ff69b08a35a87fc44acb8ab70e772559f1720f07e284e00b6b61a7313 WatchSource:0}: Error finding container 39836f1ff69b08a35a87fc44acb8ab70e772559f1720f07e284e00b6b61a7313: Status 404 returned error can't find the container with id 39836f1ff69b08a35a87fc44acb8ab70e772559f1720f07e284e00b6b61a7313 Feb 17 13:18:11 crc kubenswrapper[4955]: I0217 13:18:11.926047 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/618eb288-1aa4-43a9-b26a-9e2f5378411d-memberlist\") pod \"speaker-dl5vp\" (UID: \"618eb288-1aa4-43a9-b26a-9e2f5378411d\") " pod="metallb-system/speaker-dl5vp" Feb 17 13:18:11 crc kubenswrapper[4955]: E0217 13:18:11.926430 4955 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Feb 17 13:18:11 crc kubenswrapper[4955]: E0217 13:18:11.926670 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/618eb288-1aa4-43a9-b26a-9e2f5378411d-memberlist podName:618eb288-1aa4-43a9-b26a-9e2f5378411d nodeName:}" failed. No retries permitted until 2026-02-17 13:18:13.926650879 +0000 UTC m=+832.449380442 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/618eb288-1aa4-43a9-b26a-9e2f5378411d-memberlist") pod "speaker-dl5vp" (UID: "618eb288-1aa4-43a9-b26a-9e2f5378411d") : secret "metallb-memberlist" not found Feb 17 13:18:12 crc kubenswrapper[4955]: I0217 13:18:12.191439 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-69bbfbf88f-pvzn9" event={"ID":"8c11830d-6699-432a-a2d3-14b135bc1665","Type":"ContainerStarted","Data":"a54d4629b5b40e68a38fab8d8b50c9281146fde2b6cf64870ab9633dad5f1409"} Feb 17 13:18:12 crc kubenswrapper[4955]: I0217 13:18:12.191499 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-69bbfbf88f-pvzn9" event={"ID":"8c11830d-6699-432a-a2d3-14b135bc1665","Type":"ContainerStarted","Data":"b0ac491332603fc44a50355e8c42852562b3a56219cca4db96995be8a8989e90"} Feb 17 13:18:12 crc kubenswrapper[4955]: I0217 13:18:12.191515 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-69bbfbf88f-pvzn9" event={"ID":"8c11830d-6699-432a-a2d3-14b135bc1665","Type":"ContainerStarted","Data":"39836f1ff69b08a35a87fc44acb8ab70e772559f1720f07e284e00b6b61a7313"} Feb 17 13:18:12 crc kubenswrapper[4955]: I0217 13:18:12.191803 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-69bbfbf88f-pvzn9" Feb 17 13:18:12 crc kubenswrapper[4955]: I0217 13:18:12.206133 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-69bbfbf88f-pvzn9" podStartSLOduration=2.206113327 podStartE2EDuration="2.206113327s" podCreationTimestamp="2026-02-17 13:18:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:18:12.204912201 +0000 UTC m=+830.727641744" watchObservedRunningTime="2026-02-17 13:18:12.206113327 +0000 UTC m=+830.728842870" Feb 17 13:18:13 crc kubenswrapper[4955]: I0217 13:18:13.964733 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/618eb288-1aa4-43a9-b26a-9e2f5378411d-memberlist\") pod \"speaker-dl5vp\" (UID: \"618eb288-1aa4-43a9-b26a-9e2f5378411d\") " pod="metallb-system/speaker-dl5vp" Feb 17 13:18:13 crc kubenswrapper[4955]: I0217 13:18:13.980357 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/618eb288-1aa4-43a9-b26a-9e2f5378411d-memberlist\") pod \"speaker-dl5vp\" (UID: \"618eb288-1aa4-43a9-b26a-9e2f5378411d\") " pod="metallb-system/speaker-dl5vp" Feb 17 13:18:14 crc kubenswrapper[4955]: I0217 13:18:14.186507 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-dl5vp" Feb 17 13:18:14 crc kubenswrapper[4955]: W0217 13:18:14.235180 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod618eb288_1aa4_43a9_b26a_9e2f5378411d.slice/crio-8ebe036c934fbb82f52ec1f3ae9d7ae21edb9554dc118f8827acf3c6db699531 WatchSource:0}: Error finding container 8ebe036c934fbb82f52ec1f3ae9d7ae21edb9554dc118f8827acf3c6db699531: Status 404 returned error can't find the container with id 8ebe036c934fbb82f52ec1f3ae9d7ae21edb9554dc118f8827acf3c6db699531 Feb 17 13:18:15 crc kubenswrapper[4955]: I0217 13:18:15.219114 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-dl5vp" event={"ID":"618eb288-1aa4-43a9-b26a-9e2f5378411d","Type":"ContainerStarted","Data":"3a26a3ea8852305e650dd9951bb6755466876b6bfb657bb2e9fed67aa586e120"} Feb 17 13:18:15 crc kubenswrapper[4955]: I0217 13:18:15.219438 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-dl5vp" event={"ID":"618eb288-1aa4-43a9-b26a-9e2f5378411d","Type":"ContainerStarted","Data":"062db5cef2fabbdae1cc8ca55bc5204aa6e91e6b6fb3bf904344e722c357ba06"} Feb 17 13:18:15 crc kubenswrapper[4955]: I0217 13:18:15.219448 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-dl5vp" event={"ID":"618eb288-1aa4-43a9-b26a-9e2f5378411d","Type":"ContainerStarted","Data":"8ebe036c934fbb82f52ec1f3ae9d7ae21edb9554dc118f8827acf3c6db699531"} Feb 17 13:18:15 crc kubenswrapper[4955]: I0217 13:18:15.219608 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-dl5vp" Feb 17 13:18:15 crc kubenswrapper[4955]: I0217 13:18:15.242518 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-dl5vp" podStartSLOduration=5.242502732 podStartE2EDuration="5.242502732s" podCreationTimestamp="2026-02-17 13:18:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:18:15.239947757 +0000 UTC m=+833.762677300" watchObservedRunningTime="2026-02-17 13:18:15.242502732 +0000 UTC m=+833.765232275" Feb 17 13:18:18 crc kubenswrapper[4955]: I0217 13:18:18.242731 4955 generic.go:334] "Generic (PLEG): container finished" podID="720ac59e-55ea-44be-b7af-9b57fec4901b" containerID="de4dba7ecbb05927fc732c85aca71b2c1a126e55993587ef21c23147df43fee1" exitCode=0 Feb 17 13:18:18 crc kubenswrapper[4955]: I0217 13:18:18.242863 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vzhr7" event={"ID":"720ac59e-55ea-44be-b7af-9b57fec4901b","Type":"ContainerDied","Data":"de4dba7ecbb05927fc732c85aca71b2c1a126e55993587ef21c23147df43fee1"} Feb 17 13:18:18 crc kubenswrapper[4955]: I0217 13:18:18.246514 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-rcv4v" event={"ID":"92393c39-fc3d-490c-93a3-c11ddef89c02","Type":"ContainerStarted","Data":"ae8bb04dc9161027095e62e67751e901a277aee7bea8d9fc23019feddedd362c"} Feb 17 13:18:18 crc kubenswrapper[4955]: I0217 13:18:18.247414 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-rcv4v" Feb 17 13:18:19 crc kubenswrapper[4955]: I0217 13:18:19.255021 4955 generic.go:334] "Generic (PLEG): container finished" podID="720ac59e-55ea-44be-b7af-9b57fec4901b" containerID="ba06862daa8d50d83ff953afd34e880ae6e339649b2d0e3cf95d4c53ce336b06" exitCode=0 Feb 17 13:18:19 crc kubenswrapper[4955]: I0217 13:18:19.255112 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vzhr7" event={"ID":"720ac59e-55ea-44be-b7af-9b57fec4901b","Type":"ContainerDied","Data":"ba06862daa8d50d83ff953afd34e880ae6e339649b2d0e3cf95d4c53ce336b06"} Feb 17 13:18:19 crc kubenswrapper[4955]: I0217 13:18:19.284915 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-rcv4v" podStartSLOduration=2.475890022 podStartE2EDuration="9.284895058s" podCreationTimestamp="2026-02-17 13:18:10 +0000 UTC" firstStartedPulling="2026-02-17 13:18:10.677109907 +0000 UTC m=+829.199839450" lastFinishedPulling="2026-02-17 13:18:17.486114943 +0000 UTC m=+836.008844486" observedRunningTime="2026-02-17 13:18:18.342216161 +0000 UTC m=+836.864945704" watchObservedRunningTime="2026-02-17 13:18:19.284895058 +0000 UTC m=+837.807624611" Feb 17 13:18:20 crc kubenswrapper[4955]: I0217 13:18:20.264095 4955 generic.go:334] "Generic (PLEG): container finished" podID="720ac59e-55ea-44be-b7af-9b57fec4901b" containerID="9a4ead6bdeac7d4a737de867d89d2047025f810d72a8ce4159756dce868122f6" exitCode=0 Feb 17 13:18:20 crc kubenswrapper[4955]: I0217 13:18:20.264160 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vzhr7" event={"ID":"720ac59e-55ea-44be-b7af-9b57fec4901b","Type":"ContainerDied","Data":"9a4ead6bdeac7d4a737de867d89d2047025f810d72a8ce4159756dce868122f6"} Feb 17 13:18:21 crc kubenswrapper[4955]: I0217 13:18:21.197907 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-69bbfbf88f-pvzn9" Feb 17 13:18:21 crc kubenswrapper[4955]: I0217 13:18:21.272670 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vzhr7" event={"ID":"720ac59e-55ea-44be-b7af-9b57fec4901b","Type":"ContainerStarted","Data":"90ea81fb23cada2ab978cffd292812381bddc3171202b17195a17c0f593d4475"} Feb 17 13:18:21 crc kubenswrapper[4955]: I0217 13:18:21.272714 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vzhr7" event={"ID":"720ac59e-55ea-44be-b7af-9b57fec4901b","Type":"ContainerStarted","Data":"3ae9cd1df4506176f78433525912448212ae5e3d8a82d77fb277074117213228"} Feb 17 13:18:21 crc kubenswrapper[4955]: I0217 13:18:21.272727 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vzhr7" event={"ID":"720ac59e-55ea-44be-b7af-9b57fec4901b","Type":"ContainerStarted","Data":"4ef236aa12180013f7f7400bcd75380e13fd16befba5272b8e65debbaf04014c"} Feb 17 13:18:21 crc kubenswrapper[4955]: I0217 13:18:21.272739 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vzhr7" event={"ID":"720ac59e-55ea-44be-b7af-9b57fec4901b","Type":"ContainerStarted","Data":"4ca4854b3909eb5c01126921be0b48f6e5c2bd08643acfe1e6b9d461a6abe999"} Feb 17 13:18:22 crc kubenswrapper[4955]: I0217 13:18:22.289039 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vzhr7" event={"ID":"720ac59e-55ea-44be-b7af-9b57fec4901b","Type":"ContainerStarted","Data":"71c14611f18bf701a9da560fcd8db5384238126a46127609b42a13327e02bf62"} Feb 17 13:18:22 crc kubenswrapper[4955]: I0217 13:18:22.289434 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-vzhr7" Feb 17 13:18:22 crc kubenswrapper[4955]: I0217 13:18:22.289455 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vzhr7" event={"ID":"720ac59e-55ea-44be-b7af-9b57fec4901b","Type":"ContainerStarted","Data":"6e9bfe1d80c42b64c4db9ab97e9711cc52dfb4c4a4ae02a4e01ec85e53f46760"} Feb 17 13:18:24 crc kubenswrapper[4955]: I0217 13:18:24.192543 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-dl5vp" Feb 17 13:18:24 crc kubenswrapper[4955]: I0217 13:18:24.208203 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-vzhr7" podStartSLOduration=7.286801076 podStartE2EDuration="14.208172872s" podCreationTimestamp="2026-02-17 13:18:10 +0000 UTC" firstStartedPulling="2026-02-17 13:18:10.536395433 +0000 UTC m=+829.059124976" lastFinishedPulling="2026-02-17 13:18:17.457767209 +0000 UTC m=+835.980496772" observedRunningTime="2026-02-17 13:18:22.331190844 +0000 UTC m=+840.853920407" watchObservedRunningTime="2026-02-17 13:18:24.208172872 +0000 UTC m=+842.730902465" Feb 17 13:18:25 crc kubenswrapper[4955]: I0217 13:18:25.421765 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-vzhr7" Feb 17 13:18:25 crc kubenswrapper[4955]: I0217 13:18:25.478469 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-vzhr7" Feb 17 13:18:27 crc kubenswrapper[4955]: I0217 13:18:27.213897 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-drhvx"] Feb 17 13:18:27 crc kubenswrapper[4955]: I0217 13:18:27.215132 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-drhvx" Feb 17 13:18:27 crc kubenswrapper[4955]: I0217 13:18:27.225708 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Feb 17 13:18:27 crc kubenswrapper[4955]: I0217 13:18:27.225867 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Feb 17 13:18:27 crc kubenswrapper[4955]: I0217 13:18:27.226036 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-blrbl" Feb 17 13:18:27 crc kubenswrapper[4955]: I0217 13:18:27.233685 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-drhvx"] Feb 17 13:18:27 crc kubenswrapper[4955]: I0217 13:18:27.303425 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hw9qd\" (UniqueName: \"kubernetes.io/projected/f314c133-16d8-4e7c-9d46-eecba7fcfd0f-kube-api-access-hw9qd\") pod \"openstack-operator-index-drhvx\" (UID: \"f314c133-16d8-4e7c-9d46-eecba7fcfd0f\") " pod="openstack-operators/openstack-operator-index-drhvx" Feb 17 13:18:27 crc kubenswrapper[4955]: I0217 13:18:27.404887 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hw9qd\" (UniqueName: \"kubernetes.io/projected/f314c133-16d8-4e7c-9d46-eecba7fcfd0f-kube-api-access-hw9qd\") pod \"openstack-operator-index-drhvx\" (UID: \"f314c133-16d8-4e7c-9d46-eecba7fcfd0f\") " pod="openstack-operators/openstack-operator-index-drhvx" Feb 17 13:18:27 crc kubenswrapper[4955]: I0217 13:18:27.422490 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hw9qd\" (UniqueName: \"kubernetes.io/projected/f314c133-16d8-4e7c-9d46-eecba7fcfd0f-kube-api-access-hw9qd\") pod \"openstack-operator-index-drhvx\" (UID: \"f314c133-16d8-4e7c-9d46-eecba7fcfd0f\") " pod="openstack-operators/openstack-operator-index-drhvx" Feb 17 13:18:27 crc kubenswrapper[4955]: I0217 13:18:27.560106 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-drhvx" Feb 17 13:18:28 crc kubenswrapper[4955]: I0217 13:18:28.042427 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-drhvx"] Feb 17 13:18:28 crc kubenswrapper[4955]: I0217 13:18:28.331950 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-drhvx" event={"ID":"f314c133-16d8-4e7c-9d46-eecba7fcfd0f","Type":"ContainerStarted","Data":"88a580c8444f0012bdee74ee16fd0bf65f4f686536f4832ed2ab1fd38d0fdf8a"} Feb 17 13:18:30 crc kubenswrapper[4955]: I0217 13:18:30.349839 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-drhvx" event={"ID":"f314c133-16d8-4e7c-9d46-eecba7fcfd0f","Type":"ContainerStarted","Data":"d8e971afab22aa90d3b58196b33da84df286c1ba7b804e8cecea74bbae0005fe"} Feb 17 13:18:30 crc kubenswrapper[4955]: I0217 13:18:30.369618 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-drhvx" podStartSLOduration=1.298161691 podStartE2EDuration="3.369602428s" podCreationTimestamp="2026-02-17 13:18:27 +0000 UTC" firstStartedPulling="2026-02-17 13:18:28.059619154 +0000 UTC m=+846.582348707" lastFinishedPulling="2026-02-17 13:18:30.131059901 +0000 UTC m=+848.653789444" observedRunningTime="2026-02-17 13:18:30.369384362 +0000 UTC m=+848.892113905" watchObservedRunningTime="2026-02-17 13:18:30.369602428 +0000 UTC m=+848.892331971" Feb 17 13:18:30 crc kubenswrapper[4955]: I0217 13:18:30.424024 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-vzhr7" Feb 17 13:18:30 crc kubenswrapper[4955]: I0217 13:18:30.430321 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-rcv4v" Feb 17 13:18:30 crc kubenswrapper[4955]: I0217 13:18:30.593620 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-drhvx"] Feb 17 13:18:31 crc kubenswrapper[4955]: I0217 13:18:31.198154 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-5qzcl"] Feb 17 13:18:31 crc kubenswrapper[4955]: I0217 13:18:31.198958 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-5qzcl" Feb 17 13:18:31 crc kubenswrapper[4955]: I0217 13:18:31.216722 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-5qzcl"] Feb 17 13:18:31 crc kubenswrapper[4955]: I0217 13:18:31.257083 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwkrr\" (UniqueName: \"kubernetes.io/projected/38d53049-7d41-4938-b2d6-c7c7568ba144-kube-api-access-qwkrr\") pod \"openstack-operator-index-5qzcl\" (UID: \"38d53049-7d41-4938-b2d6-c7c7568ba144\") " pod="openstack-operators/openstack-operator-index-5qzcl" Feb 17 13:18:31 crc kubenswrapper[4955]: I0217 13:18:31.358456 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwkrr\" (UniqueName: \"kubernetes.io/projected/38d53049-7d41-4938-b2d6-c7c7568ba144-kube-api-access-qwkrr\") pod \"openstack-operator-index-5qzcl\" (UID: \"38d53049-7d41-4938-b2d6-c7c7568ba144\") " pod="openstack-operators/openstack-operator-index-5qzcl" Feb 17 13:18:31 crc kubenswrapper[4955]: I0217 13:18:31.382175 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwkrr\" (UniqueName: \"kubernetes.io/projected/38d53049-7d41-4938-b2d6-c7c7568ba144-kube-api-access-qwkrr\") pod \"openstack-operator-index-5qzcl\" (UID: \"38d53049-7d41-4938-b2d6-c7c7568ba144\") " pod="openstack-operators/openstack-operator-index-5qzcl" Feb 17 13:18:31 crc kubenswrapper[4955]: I0217 13:18:31.532168 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-5qzcl" Feb 17 13:18:31 crc kubenswrapper[4955]: I0217 13:18:31.786234 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-5qzcl"] Feb 17 13:18:31 crc kubenswrapper[4955]: W0217 13:18:31.800966 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod38d53049_7d41_4938_b2d6_c7c7568ba144.slice/crio-b69b34440de163b8cee0714fedeba045ef225e9d3e6dd2d2de9705006befeba2 WatchSource:0}: Error finding container b69b34440de163b8cee0714fedeba045ef225e9d3e6dd2d2de9705006befeba2: Status 404 returned error can't find the container with id b69b34440de163b8cee0714fedeba045ef225e9d3e6dd2d2de9705006befeba2 Feb 17 13:18:32 crc kubenswrapper[4955]: I0217 13:18:32.365142 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-5qzcl" event={"ID":"38d53049-7d41-4938-b2d6-c7c7568ba144","Type":"ContainerStarted","Data":"0d029198bf40a6a8576662ec94dfb5488d8a6855b3479e9b734677e6115f1e82"} Feb 17 13:18:32 crc kubenswrapper[4955]: I0217 13:18:32.367158 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-5qzcl" event={"ID":"38d53049-7d41-4938-b2d6-c7c7568ba144","Type":"ContainerStarted","Data":"b69b34440de163b8cee0714fedeba045ef225e9d3e6dd2d2de9705006befeba2"} Feb 17 13:18:32 crc kubenswrapper[4955]: I0217 13:18:32.365207 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-drhvx" podUID="f314c133-16d8-4e7c-9d46-eecba7fcfd0f" containerName="registry-server" containerID="cri-o://d8e971afab22aa90d3b58196b33da84df286c1ba7b804e8cecea74bbae0005fe" gracePeriod=2 Feb 17 13:18:32 crc kubenswrapper[4955]: I0217 13:18:32.396939 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-5qzcl" podStartSLOduration=1.3569913470000001 podStartE2EDuration="1.396910531s" podCreationTimestamp="2026-02-17 13:18:31 +0000 UTC" firstStartedPulling="2026-02-17 13:18:31.80268704 +0000 UTC m=+850.325416593" lastFinishedPulling="2026-02-17 13:18:31.842606224 +0000 UTC m=+850.365335777" observedRunningTime="2026-02-17 13:18:32.392247977 +0000 UTC m=+850.914977520" watchObservedRunningTime="2026-02-17 13:18:32.396910531 +0000 UTC m=+850.919640084" Feb 17 13:18:32 crc kubenswrapper[4955]: I0217 13:18:32.806324 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-drhvx" Feb 17 13:18:32 crc kubenswrapper[4955]: I0217 13:18:32.881298 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hw9qd\" (UniqueName: \"kubernetes.io/projected/f314c133-16d8-4e7c-9d46-eecba7fcfd0f-kube-api-access-hw9qd\") pod \"f314c133-16d8-4e7c-9d46-eecba7fcfd0f\" (UID: \"f314c133-16d8-4e7c-9d46-eecba7fcfd0f\") " Feb 17 13:18:32 crc kubenswrapper[4955]: I0217 13:18:32.906268 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f314c133-16d8-4e7c-9d46-eecba7fcfd0f-kube-api-access-hw9qd" (OuterVolumeSpecName: "kube-api-access-hw9qd") pod "f314c133-16d8-4e7c-9d46-eecba7fcfd0f" (UID: "f314c133-16d8-4e7c-9d46-eecba7fcfd0f"). InnerVolumeSpecName "kube-api-access-hw9qd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:18:32 crc kubenswrapper[4955]: I0217 13:18:32.983066 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hw9qd\" (UniqueName: \"kubernetes.io/projected/f314c133-16d8-4e7c-9d46-eecba7fcfd0f-kube-api-access-hw9qd\") on node \"crc\" DevicePath \"\"" Feb 17 13:18:33 crc kubenswrapper[4955]: I0217 13:18:33.379585 4955 generic.go:334] "Generic (PLEG): container finished" podID="f314c133-16d8-4e7c-9d46-eecba7fcfd0f" containerID="d8e971afab22aa90d3b58196b33da84df286c1ba7b804e8cecea74bbae0005fe" exitCode=0 Feb 17 13:18:33 crc kubenswrapper[4955]: I0217 13:18:33.379862 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-drhvx" Feb 17 13:18:33 crc kubenswrapper[4955]: I0217 13:18:33.379752 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-drhvx" event={"ID":"f314c133-16d8-4e7c-9d46-eecba7fcfd0f","Type":"ContainerDied","Data":"d8e971afab22aa90d3b58196b33da84df286c1ba7b804e8cecea74bbae0005fe"} Feb 17 13:18:33 crc kubenswrapper[4955]: I0217 13:18:33.379992 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-drhvx" event={"ID":"f314c133-16d8-4e7c-9d46-eecba7fcfd0f","Type":"ContainerDied","Data":"88a580c8444f0012bdee74ee16fd0bf65f4f686536f4832ed2ab1fd38d0fdf8a"} Feb 17 13:18:33 crc kubenswrapper[4955]: I0217 13:18:33.380076 4955 scope.go:117] "RemoveContainer" containerID="d8e971afab22aa90d3b58196b33da84df286c1ba7b804e8cecea74bbae0005fe" Feb 17 13:18:33 crc kubenswrapper[4955]: I0217 13:18:33.422053 4955 scope.go:117] "RemoveContainer" containerID="d8e971afab22aa90d3b58196b33da84df286c1ba7b804e8cecea74bbae0005fe" Feb 17 13:18:33 crc kubenswrapper[4955]: E0217 13:18:33.422736 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8e971afab22aa90d3b58196b33da84df286c1ba7b804e8cecea74bbae0005fe\": container with ID starting with d8e971afab22aa90d3b58196b33da84df286c1ba7b804e8cecea74bbae0005fe not found: ID does not exist" containerID="d8e971afab22aa90d3b58196b33da84df286c1ba7b804e8cecea74bbae0005fe" Feb 17 13:18:33 crc kubenswrapper[4955]: I0217 13:18:33.422842 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8e971afab22aa90d3b58196b33da84df286c1ba7b804e8cecea74bbae0005fe"} err="failed to get container status \"d8e971afab22aa90d3b58196b33da84df286c1ba7b804e8cecea74bbae0005fe\": rpc error: code = NotFound desc = could not find container \"d8e971afab22aa90d3b58196b33da84df286c1ba7b804e8cecea74bbae0005fe\": container with ID starting with d8e971afab22aa90d3b58196b33da84df286c1ba7b804e8cecea74bbae0005fe not found: ID does not exist" Feb 17 13:18:33 crc kubenswrapper[4955]: I0217 13:18:33.428976 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-drhvx"] Feb 17 13:18:33 crc kubenswrapper[4955]: I0217 13:18:33.433723 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-drhvx"] Feb 17 13:18:34 crc kubenswrapper[4955]: I0217 13:18:34.237923 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f314c133-16d8-4e7c-9d46-eecba7fcfd0f" path="/var/lib/kubelet/pods/f314c133-16d8-4e7c-9d46-eecba7fcfd0f/volumes" Feb 17 13:18:41 crc kubenswrapper[4955]: I0217 13:18:41.532899 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-5qzcl" Feb 17 13:18:41 crc kubenswrapper[4955]: I0217 13:18:41.533291 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-5qzcl" Feb 17 13:18:41 crc kubenswrapper[4955]: I0217 13:18:41.567727 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-5qzcl" Feb 17 13:18:42 crc kubenswrapper[4955]: I0217 13:18:42.490749 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-5qzcl" Feb 17 13:18:43 crc kubenswrapper[4955]: I0217 13:18:43.435579 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p"] Feb 17 13:18:43 crc kubenswrapper[4955]: E0217 13:18:43.435916 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f314c133-16d8-4e7c-9d46-eecba7fcfd0f" containerName="registry-server" Feb 17 13:18:43 crc kubenswrapper[4955]: I0217 13:18:43.435933 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="f314c133-16d8-4e7c-9d46-eecba7fcfd0f" containerName="registry-server" Feb 17 13:18:43 crc kubenswrapper[4955]: I0217 13:18:43.436063 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="f314c133-16d8-4e7c-9d46-eecba7fcfd0f" containerName="registry-server" Feb 17 13:18:43 crc kubenswrapper[4955]: I0217 13:18:43.437357 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p" Feb 17 13:18:43 crc kubenswrapper[4955]: I0217 13:18:43.439357 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-nl959" Feb 17 13:18:43 crc kubenswrapper[4955]: I0217 13:18:43.447014 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p"] Feb 17 13:18:43 crc kubenswrapper[4955]: I0217 13:18:43.538315 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f328d9c1-907b-4705-8580-0b5da54053c6-bundle\") pod \"d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p\" (UID: \"f328d9c1-907b-4705-8580-0b5da54053c6\") " pod="openstack-operators/d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p" Feb 17 13:18:43 crc kubenswrapper[4955]: I0217 13:18:43.539604 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f328d9c1-907b-4705-8580-0b5da54053c6-util\") pod \"d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p\" (UID: \"f328d9c1-907b-4705-8580-0b5da54053c6\") " pod="openstack-operators/d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p" Feb 17 13:18:43 crc kubenswrapper[4955]: I0217 13:18:43.539903 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxvdt\" (UniqueName: \"kubernetes.io/projected/f328d9c1-907b-4705-8580-0b5da54053c6-kube-api-access-nxvdt\") pod \"d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p\" (UID: \"f328d9c1-907b-4705-8580-0b5da54053c6\") " pod="openstack-operators/d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p" Feb 17 13:18:43 crc kubenswrapper[4955]: I0217 13:18:43.641715 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f328d9c1-907b-4705-8580-0b5da54053c6-util\") pod \"d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p\" (UID: \"f328d9c1-907b-4705-8580-0b5da54053c6\") " pod="openstack-operators/d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p" Feb 17 13:18:43 crc kubenswrapper[4955]: I0217 13:18:43.642130 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxvdt\" (UniqueName: \"kubernetes.io/projected/f328d9c1-907b-4705-8580-0b5da54053c6-kube-api-access-nxvdt\") pod \"d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p\" (UID: \"f328d9c1-907b-4705-8580-0b5da54053c6\") " pod="openstack-operators/d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p" Feb 17 13:18:43 crc kubenswrapper[4955]: I0217 13:18:43.642495 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f328d9c1-907b-4705-8580-0b5da54053c6-util\") pod \"d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p\" (UID: \"f328d9c1-907b-4705-8580-0b5da54053c6\") " pod="openstack-operators/d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p" Feb 17 13:18:43 crc kubenswrapper[4955]: I0217 13:18:43.644019 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f328d9c1-907b-4705-8580-0b5da54053c6-bundle\") pod \"d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p\" (UID: \"f328d9c1-907b-4705-8580-0b5da54053c6\") " pod="openstack-operators/d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p" Feb 17 13:18:43 crc kubenswrapper[4955]: I0217 13:18:43.643428 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f328d9c1-907b-4705-8580-0b5da54053c6-bundle\") pod \"d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p\" (UID: \"f328d9c1-907b-4705-8580-0b5da54053c6\") " pod="openstack-operators/d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p" Feb 17 13:18:43 crc kubenswrapper[4955]: I0217 13:18:43.672248 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxvdt\" (UniqueName: \"kubernetes.io/projected/f328d9c1-907b-4705-8580-0b5da54053c6-kube-api-access-nxvdt\") pod \"d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p\" (UID: \"f328d9c1-907b-4705-8580-0b5da54053c6\") " pod="openstack-operators/d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p" Feb 17 13:18:43 crc kubenswrapper[4955]: I0217 13:18:43.762019 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p" Feb 17 13:18:43 crc kubenswrapper[4955]: I0217 13:18:43.955954 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p"] Feb 17 13:18:44 crc kubenswrapper[4955]: I0217 13:18:44.482170 4955 generic.go:334] "Generic (PLEG): container finished" podID="f328d9c1-907b-4705-8580-0b5da54053c6" containerID="d8351d33d3ce233d543dadfb9e6e760fe7ee7a093d05134d1d5236caddfc6f47" exitCode=0 Feb 17 13:18:44 crc kubenswrapper[4955]: I0217 13:18:44.482541 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p" event={"ID":"f328d9c1-907b-4705-8580-0b5da54053c6","Type":"ContainerDied","Data":"d8351d33d3ce233d543dadfb9e6e760fe7ee7a093d05134d1d5236caddfc6f47"} Feb 17 13:18:44 crc kubenswrapper[4955]: I0217 13:18:44.482766 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p" event={"ID":"f328d9c1-907b-4705-8580-0b5da54053c6","Type":"ContainerStarted","Data":"caa59bc0b543f33e497f0475f8ee2498ed10a4c01a52ba62e9ca986ceb03cb40"} Feb 17 13:18:45 crc kubenswrapper[4955]: I0217 13:18:45.493629 4955 generic.go:334] "Generic (PLEG): container finished" podID="f328d9c1-907b-4705-8580-0b5da54053c6" containerID="7cdd3dc6f617cb7aa0b357bfc03d377160a80c473ecb5bedd79e039eefd6fa00" exitCode=0 Feb 17 13:18:45 crc kubenswrapper[4955]: I0217 13:18:45.493712 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p" event={"ID":"f328d9c1-907b-4705-8580-0b5da54053c6","Type":"ContainerDied","Data":"7cdd3dc6f617cb7aa0b357bfc03d377160a80c473ecb5bedd79e039eefd6fa00"} Feb 17 13:18:46 crc kubenswrapper[4955]: I0217 13:18:46.502317 4955 generic.go:334] "Generic (PLEG): container finished" podID="f328d9c1-907b-4705-8580-0b5da54053c6" containerID="be815be4a93eb20ec03f70762b44c98d558b538cdc9c6923edb5cc1f91f52538" exitCode=0 Feb 17 13:18:46 crc kubenswrapper[4955]: I0217 13:18:46.502357 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p" event={"ID":"f328d9c1-907b-4705-8580-0b5da54053c6","Type":"ContainerDied","Data":"be815be4a93eb20ec03f70762b44c98d558b538cdc9c6923edb5cc1f91f52538"} Feb 17 13:18:47 crc kubenswrapper[4955]: I0217 13:18:47.780191 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p" Feb 17 13:18:47 crc kubenswrapper[4955]: I0217 13:18:47.899229 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f328d9c1-907b-4705-8580-0b5da54053c6-util\") pod \"f328d9c1-907b-4705-8580-0b5da54053c6\" (UID: \"f328d9c1-907b-4705-8580-0b5da54053c6\") " Feb 17 13:18:47 crc kubenswrapper[4955]: I0217 13:18:47.899560 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f328d9c1-907b-4705-8580-0b5da54053c6-bundle\") pod \"f328d9c1-907b-4705-8580-0b5da54053c6\" (UID: \"f328d9c1-907b-4705-8580-0b5da54053c6\") " Feb 17 13:18:47 crc kubenswrapper[4955]: I0217 13:18:47.899594 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxvdt\" (UniqueName: \"kubernetes.io/projected/f328d9c1-907b-4705-8580-0b5da54053c6-kube-api-access-nxvdt\") pod \"f328d9c1-907b-4705-8580-0b5da54053c6\" (UID: \"f328d9c1-907b-4705-8580-0b5da54053c6\") " Feb 17 13:18:47 crc kubenswrapper[4955]: I0217 13:18:47.900264 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f328d9c1-907b-4705-8580-0b5da54053c6-bundle" (OuterVolumeSpecName: "bundle") pod "f328d9c1-907b-4705-8580-0b5da54053c6" (UID: "f328d9c1-907b-4705-8580-0b5da54053c6"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:18:47 crc kubenswrapper[4955]: I0217 13:18:47.909115 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f328d9c1-907b-4705-8580-0b5da54053c6-kube-api-access-nxvdt" (OuterVolumeSpecName: "kube-api-access-nxvdt") pod "f328d9c1-907b-4705-8580-0b5da54053c6" (UID: "f328d9c1-907b-4705-8580-0b5da54053c6"). InnerVolumeSpecName "kube-api-access-nxvdt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:18:47 crc kubenswrapper[4955]: I0217 13:18:47.928455 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f328d9c1-907b-4705-8580-0b5da54053c6-util" (OuterVolumeSpecName: "util") pod "f328d9c1-907b-4705-8580-0b5da54053c6" (UID: "f328d9c1-907b-4705-8580-0b5da54053c6"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:18:48 crc kubenswrapper[4955]: I0217 13:18:48.001475 4955 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f328d9c1-907b-4705-8580-0b5da54053c6-util\") on node \"crc\" DevicePath \"\"" Feb 17 13:18:48 crc kubenswrapper[4955]: I0217 13:18:48.001528 4955 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f328d9c1-907b-4705-8580-0b5da54053c6-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:18:48 crc kubenswrapper[4955]: I0217 13:18:48.001549 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nxvdt\" (UniqueName: \"kubernetes.io/projected/f328d9c1-907b-4705-8580-0b5da54053c6-kube-api-access-nxvdt\") on node \"crc\" DevicePath \"\"" Feb 17 13:18:48 crc kubenswrapper[4955]: I0217 13:18:48.519762 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p" event={"ID":"f328d9c1-907b-4705-8580-0b5da54053c6","Type":"ContainerDied","Data":"caa59bc0b543f33e497f0475f8ee2498ed10a4c01a52ba62e9ca986ceb03cb40"} Feb 17 13:18:48 crc kubenswrapper[4955]: I0217 13:18:48.519816 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="caa59bc0b543f33e497f0475f8ee2498ed10a4c01a52ba62e9ca986ceb03cb40" Feb 17 13:18:48 crc kubenswrapper[4955]: I0217 13:18:48.519837 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p" Feb 17 13:18:56 crc kubenswrapper[4955]: I0217 13:18:56.045026 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-57f9c85f89-8zxwf"] Feb 17 13:18:56 crc kubenswrapper[4955]: E0217 13:18:56.045814 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f328d9c1-907b-4705-8580-0b5da54053c6" containerName="util" Feb 17 13:18:56 crc kubenswrapper[4955]: I0217 13:18:56.045829 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="f328d9c1-907b-4705-8580-0b5da54053c6" containerName="util" Feb 17 13:18:56 crc kubenswrapper[4955]: E0217 13:18:56.045847 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f328d9c1-907b-4705-8580-0b5da54053c6" containerName="pull" Feb 17 13:18:56 crc kubenswrapper[4955]: I0217 13:18:56.045854 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="f328d9c1-907b-4705-8580-0b5da54053c6" containerName="pull" Feb 17 13:18:56 crc kubenswrapper[4955]: E0217 13:18:56.045875 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f328d9c1-907b-4705-8580-0b5da54053c6" containerName="extract" Feb 17 13:18:56 crc kubenswrapper[4955]: I0217 13:18:56.045882 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="f328d9c1-907b-4705-8580-0b5da54053c6" containerName="extract" Feb 17 13:18:56 crc kubenswrapper[4955]: I0217 13:18:56.046007 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="f328d9c1-907b-4705-8580-0b5da54053c6" containerName="extract" Feb 17 13:18:56 crc kubenswrapper[4955]: I0217 13:18:56.046487 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-57f9c85f89-8zxwf" Feb 17 13:18:56 crc kubenswrapper[4955]: I0217 13:18:56.048991 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-49874" Feb 17 13:18:56 crc kubenswrapper[4955]: I0217 13:18:56.096418 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-57f9c85f89-8zxwf"] Feb 17 13:18:56 crc kubenswrapper[4955]: I0217 13:18:56.119665 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqsdr\" (UniqueName: \"kubernetes.io/projected/5b01c583-8db4-4939-ac7b-e0d16b9f72bf-kube-api-access-jqsdr\") pod \"openstack-operator-controller-init-57f9c85f89-8zxwf\" (UID: \"5b01c583-8db4-4939-ac7b-e0d16b9f72bf\") " pod="openstack-operators/openstack-operator-controller-init-57f9c85f89-8zxwf" Feb 17 13:18:56 crc kubenswrapper[4955]: I0217 13:18:56.220670 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqsdr\" (UniqueName: \"kubernetes.io/projected/5b01c583-8db4-4939-ac7b-e0d16b9f72bf-kube-api-access-jqsdr\") pod \"openstack-operator-controller-init-57f9c85f89-8zxwf\" (UID: \"5b01c583-8db4-4939-ac7b-e0d16b9f72bf\") " pod="openstack-operators/openstack-operator-controller-init-57f9c85f89-8zxwf" Feb 17 13:18:56 crc kubenswrapper[4955]: I0217 13:18:56.242518 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqsdr\" (UniqueName: \"kubernetes.io/projected/5b01c583-8db4-4939-ac7b-e0d16b9f72bf-kube-api-access-jqsdr\") pod \"openstack-operator-controller-init-57f9c85f89-8zxwf\" (UID: \"5b01c583-8db4-4939-ac7b-e0d16b9f72bf\") " pod="openstack-operators/openstack-operator-controller-init-57f9c85f89-8zxwf" Feb 17 13:18:56 crc kubenswrapper[4955]: I0217 13:18:56.363686 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-57f9c85f89-8zxwf" Feb 17 13:18:56 crc kubenswrapper[4955]: I0217 13:18:56.848315 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-57f9c85f89-8zxwf"] Feb 17 13:18:57 crc kubenswrapper[4955]: I0217 13:18:57.581627 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-57f9c85f89-8zxwf" event={"ID":"5b01c583-8db4-4939-ac7b-e0d16b9f72bf","Type":"ContainerStarted","Data":"bbdb5277decab5ad12300e61cd198aaace338ea9d71cd0e3550bd67737f463b7"} Feb 17 13:19:01 crc kubenswrapper[4955]: I0217 13:19:01.607956 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-57f9c85f89-8zxwf" event={"ID":"5b01c583-8db4-4939-ac7b-e0d16b9f72bf","Type":"ContainerStarted","Data":"70b74332bc90603cdd87694d59f3cce6213b24c19d0b5eb8129270daa742c333"} Feb 17 13:19:01 crc kubenswrapper[4955]: I0217 13:19:01.608456 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-57f9c85f89-8zxwf" Feb 17 13:19:04 crc kubenswrapper[4955]: I0217 13:19:04.574998 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 13:19:04 crc kubenswrapper[4955]: I0217 13:19:04.575454 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 13:19:06 crc kubenswrapper[4955]: I0217 13:19:06.367771 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-57f9c85f89-8zxwf" Feb 17 13:19:06 crc kubenswrapper[4955]: I0217 13:19:06.415979 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-57f9c85f89-8zxwf" podStartSLOduration=6.686060249 podStartE2EDuration="10.415949257s" podCreationTimestamp="2026-02-17 13:18:56 +0000 UTC" firstStartedPulling="2026-02-17 13:18:56.860384685 +0000 UTC m=+875.383114228" lastFinishedPulling="2026-02-17 13:19:00.590273693 +0000 UTC m=+879.113003236" observedRunningTime="2026-02-17 13:19:01.647392621 +0000 UTC m=+880.170122164" watchObservedRunningTime="2026-02-17 13:19:06.415949257 +0000 UTC m=+884.938678840" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.102748 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-868647ff47-ffkxx"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.104406 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-ffkxx" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.106884 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5d946d989d-lpxnk"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.107589 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-lpxnk" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.108512 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-zznkl" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.110239 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-jnrh2" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.120739 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5d946d989d-lpxnk"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.128739 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-868647ff47-ffkxx"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.134094 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d8bf5c495-2c8ww"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.134250 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ld2kb\" (UniqueName: \"kubernetes.io/projected/6d1e254f-7d68-45c5-bda0-e3b33155ae56-kube-api-access-ld2kb\") pod \"cinder-operator-controller-manager-5d946d989d-lpxnk\" (UID: \"6d1e254f-7d68-45c5-bda0-e3b33155ae56\") " pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-lpxnk" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.134317 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdvhq\" (UniqueName: \"kubernetes.io/projected/602419fb-5033-4889-b587-e1eb97001910-kube-api-access-xdvhq\") pod \"barbican-operator-controller-manager-868647ff47-ffkxx\" (UID: \"602419fb-5033-4889-b587-e1eb97001910\") " pod="openstack-operators/barbican-operator-controller-manager-868647ff47-ffkxx" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.134997 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-2c8ww" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.139167 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-cqm7r" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.143992 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d8bf5c495-2c8ww"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.154555 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987464f4-g967h"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.155201 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987464f4-g967h" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.157208 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-zcm9c" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.171894 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-69f49c598c-75hpg"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.194683 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987464f4-g967h"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.194877 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-75hpg" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.206042 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-rnnqk" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.237053 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ld2kb\" (UniqueName: \"kubernetes.io/projected/6d1e254f-7d68-45c5-bda0-e3b33155ae56-kube-api-access-ld2kb\") pod \"cinder-operator-controller-manager-5d946d989d-lpxnk\" (UID: \"6d1e254f-7d68-45c5-bda0-e3b33155ae56\") " pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-lpxnk" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.237117 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdvhq\" (UniqueName: \"kubernetes.io/projected/602419fb-5033-4889-b587-e1eb97001910-kube-api-access-xdvhq\") pod \"barbican-operator-controller-manager-868647ff47-ffkxx\" (UID: \"602419fb-5033-4889-b587-e1eb97001910\") " pod="openstack-operators/barbican-operator-controller-manager-868647ff47-ffkxx" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.254195 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-69f49c598c-75hpg"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.254899 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b9b8895d5-k2x44"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.255588 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-k2x44" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.261824 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-qtwvm" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.266769 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ld2kb\" (UniqueName: \"kubernetes.io/projected/6d1e254f-7d68-45c5-bda0-e3b33155ae56-kube-api-access-ld2kb\") pod \"cinder-operator-controller-manager-5d946d989d-lpxnk\" (UID: \"6d1e254f-7d68-45c5-bda0-e3b33155ae56\") " pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-lpxnk" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.267674 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b9b8895d5-k2x44"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.272682 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-79d975b745-gpgxj"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.273481 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-79d975b745-gpgxj" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.278593 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.278648 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-hlbsk" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.281135 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-79d975b745-gpgxj"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.287398 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-554564d7fc-h2rm6"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.288367 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-h2rm6" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.290988 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-4b92c" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.293668 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-554564d7fc-h2rm6"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.312260 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b4d948c87-mkqrg"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.313274 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-mkqrg" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.316921 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdvhq\" (UniqueName: \"kubernetes.io/projected/602419fb-5033-4889-b587-e1eb97001910-kube-api-access-xdvhq\") pod \"barbican-operator-controller-manager-868647ff47-ffkxx\" (UID: \"602419fb-5033-4889-b587-e1eb97001910\") " pod="openstack-operators/barbican-operator-controller-manager-868647ff47-ffkxx" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.318001 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-vx5qd" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.327794 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-54f6768c69-4f8kb"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.328540 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-4f8kb" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.334430 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-9pf96" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.337839 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-54f6768c69-4f8kb"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.338439 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rm7sb\" (UniqueName: \"kubernetes.io/projected/1e4fd408-9f09-4289-92ce-7c5ae18f235a-kube-api-access-rm7sb\") pod \"designate-operator-controller-manager-6d8bf5c495-2c8ww\" (UID: \"1e4fd408-9f09-4289-92ce-7c5ae18f235a\") " pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-2c8ww" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.338486 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trbd4\" (UniqueName: \"kubernetes.io/projected/071c3b13-9304-48b6-bf22-d89dd62825a0-kube-api-access-trbd4\") pod \"heat-operator-controller-manager-69f49c598c-75hpg\" (UID: \"071c3b13-9304-48b6-bf22-d89dd62825a0\") " pod="openstack-operators/heat-operator-controller-manager-69f49c598c-75hpg" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.338526 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhrvd\" (UniqueName: \"kubernetes.io/projected/5c339089-ba43-481b-aaa8-f8bf09fd106d-kube-api-access-zhrvd\") pod \"glance-operator-controller-manager-77987464f4-g967h\" (UID: \"5c339089-ba43-481b-aaa8-f8bf09fd106d\") " pod="openstack-operators/glance-operator-controller-manager-77987464f4-g967h" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.362010 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b4d948c87-mkqrg"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.365840 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6994f66f48-nzn4k"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.366620 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-nzn4k" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.368898 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-tdrt2" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.372295 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-64ddbf8bb-xkvlp"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.373011 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-xkvlp" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.375502 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-zjmhb" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.377399 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6994f66f48-nzn4k"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.402443 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-64ddbf8bb-xkvlp"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.412459 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-567668f5cf-5fxtz"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.413302 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-5fxtz" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.416236 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-c7njf" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.420139 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-567668f5cf-5fxtz"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.426064 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-ffkxx" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.429014 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f8888797-wkdhh"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.429749 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-wkdhh" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.438947 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-qktcv" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.439579 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-lpxnk" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.439718 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf592\" (UniqueName: \"kubernetes.io/projected/cc84cdf7-92aa-4134-a784-d821c076cde3-kube-api-access-wf592\") pod \"infra-operator-controller-manager-79d975b745-gpgxj\" (UID: \"cc84cdf7-92aa-4134-a784-d821c076cde3\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-gpgxj" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.439772 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhrvd\" (UniqueName: \"kubernetes.io/projected/5c339089-ba43-481b-aaa8-f8bf09fd106d-kube-api-access-zhrvd\") pod \"glance-operator-controller-manager-77987464f4-g967h\" (UID: \"5c339089-ba43-481b-aaa8-f8bf09fd106d\") " pod="openstack-operators/glance-operator-controller-manager-77987464f4-g967h" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.439816 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5td7\" (UniqueName: \"kubernetes.io/projected/9019acce-c356-421a-bd47-a163a8857f30-kube-api-access-f5td7\") pod \"ironic-operator-controller-manager-554564d7fc-h2rm6\" (UID: \"9019acce-c356-421a-bd47-a163a8857f30\") " pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-h2rm6" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.439866 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rm7sb\" (UniqueName: \"kubernetes.io/projected/1e4fd408-9f09-4289-92ce-7c5ae18f235a-kube-api-access-rm7sb\") pod \"designate-operator-controller-manager-6d8bf5c495-2c8ww\" (UID: \"1e4fd408-9f09-4289-92ce-7c5ae18f235a\") " pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-2c8ww" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.439890 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5x9m4\" (UniqueName: \"kubernetes.io/projected/6ade752d-a3d0-4449-8028-67405763ccc8-kube-api-access-5x9m4\") pod \"horizon-operator-controller-manager-5b9b8895d5-k2x44\" (UID: \"6ade752d-a3d0-4449-8028-67405763ccc8\") " pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-k2x44" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.439907 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9dhr\" (UniqueName: \"kubernetes.io/projected/d6decb51-f3f9-4183-a20a-6aa4491dba61-kube-api-access-t9dhr\") pod \"keystone-operator-controller-manager-b4d948c87-mkqrg\" (UID: \"d6decb51-f3f9-4183-a20a-6aa4491dba61\") " pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-mkqrg" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.439929 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trbd4\" (UniqueName: \"kubernetes.io/projected/071c3b13-9304-48b6-bf22-d89dd62825a0-kube-api-access-trbd4\") pod \"heat-operator-controller-manager-69f49c598c-75hpg\" (UID: \"071c3b13-9304-48b6-bf22-d89dd62825a0\") " pod="openstack-operators/heat-operator-controller-manager-69f49c598c-75hpg" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.439946 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cc84cdf7-92aa-4134-a784-d821c076cde3-cert\") pod \"infra-operator-controller-manager-79d975b745-gpgxj\" (UID: \"cc84cdf7-92aa-4134-a784-d821c076cde3\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-gpgxj" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.439966 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvc5d\" (UniqueName: \"kubernetes.io/projected/01eb7df7-9e0a-4e78-85a3-0d456dec139f-kube-api-access-rvc5d\") pod \"manila-operator-controller-manager-54f6768c69-4f8kb\" (UID: \"01eb7df7-9e0a-4e78-85a3-0d456dec139f\") " pod="openstack-operators/manila-operator-controller-manager-54f6768c69-4f8kb" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.464162 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trbd4\" (UniqueName: \"kubernetes.io/projected/071c3b13-9304-48b6-bf22-d89dd62825a0-kube-api-access-trbd4\") pod \"heat-operator-controller-manager-69f49c598c-75hpg\" (UID: \"071c3b13-9304-48b6-bf22-d89dd62825a0\") " pod="openstack-operators/heat-operator-controller-manager-69f49c598c-75hpg" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.464254 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f8888797-wkdhh"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.477253 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cxkmq8"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.478090 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cxkmq8" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.481277 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.481508 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-5btmn" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.483451 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rm7sb\" (UniqueName: \"kubernetes.io/projected/1e4fd408-9f09-4289-92ce-7c5ae18f235a-kube-api-access-rm7sb\") pod \"designate-operator-controller-manager-6d8bf5c495-2c8ww\" (UID: \"1e4fd408-9f09-4289-92ce-7c5ae18f235a\") " pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-2c8ww" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.484356 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhrvd\" (UniqueName: \"kubernetes.io/projected/5c339089-ba43-481b-aaa8-f8bf09fd106d-kube-api-access-zhrvd\") pod \"glance-operator-controller-manager-77987464f4-g967h\" (UID: \"5c339089-ba43-481b-aaa8-f8bf09fd106d\") " pod="openstack-operators/glance-operator-controller-manager-77987464f4-g967h" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.495110 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987464f4-g967h" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.528100 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cxkmq8"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.536222 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-d44cf6b75-vq5wq"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.537088 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-vq5wq" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.538742 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-pv2sl" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.540483 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5td7\" (UniqueName: \"kubernetes.io/projected/9019acce-c356-421a-bd47-a163a8857f30-kube-api-access-f5td7\") pod \"ironic-operator-controller-manager-554564d7fc-h2rm6\" (UID: \"9019acce-c356-421a-bd47-a163a8857f30\") " pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-h2rm6" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.540530 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqjzq\" (UniqueName: \"kubernetes.io/projected/308b7713-e37c-4890-8b56-11379bad3f8d-kube-api-access-cqjzq\") pod \"mariadb-operator-controller-manager-6994f66f48-nzn4k\" (UID: \"308b7713-e37c-4890-8b56-11379bad3f8d\") " pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-nzn4k" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.540563 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjhfm\" (UniqueName: \"kubernetes.io/projected/3f752291-5ba5-4074-a126-e99f6f0242aa-kube-api-access-zjhfm\") pod \"octavia-operator-controller-manager-69f8888797-wkdhh\" (UID: \"3f752291-5ba5-4074-a126-e99f6f0242aa\") " pod="openstack-operators/octavia-operator-controller-manager-69f8888797-wkdhh" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.540585 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8j754\" (UniqueName: \"kubernetes.io/projected/744f25f9-1641-4864-8a55-ea14ead19cae-kube-api-access-8j754\") pod \"nova-operator-controller-manager-567668f5cf-5fxtz\" (UID: \"744f25f9-1641-4864-8a55-ea14ead19cae\") " pod="openstack-operators/nova-operator-controller-manager-567668f5cf-5fxtz" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.540610 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5x9m4\" (UniqueName: \"kubernetes.io/projected/6ade752d-a3d0-4449-8028-67405763ccc8-kube-api-access-5x9m4\") pod \"horizon-operator-controller-manager-5b9b8895d5-k2x44\" (UID: \"6ade752d-a3d0-4449-8028-67405763ccc8\") " pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-k2x44" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.540628 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9dhr\" (UniqueName: \"kubernetes.io/projected/d6decb51-f3f9-4183-a20a-6aa4491dba61-kube-api-access-t9dhr\") pod \"keystone-operator-controller-manager-b4d948c87-mkqrg\" (UID: \"d6decb51-f3f9-4183-a20a-6aa4491dba61\") " pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-mkqrg" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.540642 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jjgc\" (UniqueName: \"kubernetes.io/projected/037562f9-c04f-4e9d-8fa5-3e98ac7c8fc5-kube-api-access-9jjgc\") pod \"neutron-operator-controller-manager-64ddbf8bb-xkvlp\" (UID: \"037562f9-c04f-4e9d-8fa5-3e98ac7c8fc5\") " pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-xkvlp" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.540664 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cc84cdf7-92aa-4134-a784-d821c076cde3-cert\") pod \"infra-operator-controller-manager-79d975b745-gpgxj\" (UID: \"cc84cdf7-92aa-4134-a784-d821c076cde3\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-gpgxj" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.540684 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvc5d\" (UniqueName: \"kubernetes.io/projected/01eb7df7-9e0a-4e78-85a3-0d456dec139f-kube-api-access-rvc5d\") pod \"manila-operator-controller-manager-54f6768c69-4f8kb\" (UID: \"01eb7df7-9e0a-4e78-85a3-0d456dec139f\") " pod="openstack-operators/manila-operator-controller-manager-54f6768c69-4f8kb" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.540701 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wf592\" (UniqueName: \"kubernetes.io/projected/cc84cdf7-92aa-4134-a784-d821c076cde3-kube-api-access-wf592\") pod \"infra-operator-controller-manager-79d975b745-gpgxj\" (UID: \"cc84cdf7-92aa-4134-a784-d821c076cde3\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-gpgxj" Feb 17 13:19:26 crc kubenswrapper[4955]: E0217 13:19:26.540941 4955 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 17 13:19:26 crc kubenswrapper[4955]: E0217 13:19:26.540988 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cc84cdf7-92aa-4134-a784-d821c076cde3-cert podName:cc84cdf7-92aa-4134-a784-d821c076cde3 nodeName:}" failed. No retries permitted until 2026-02-17 13:19:27.040974267 +0000 UTC m=+905.563703810 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/cc84cdf7-92aa-4134-a784-d821c076cde3-cert") pod "infra-operator-controller-manager-79d975b745-gpgxj" (UID: "cc84cdf7-92aa-4134-a784-d821c076cde3") : secret "infra-operator-webhook-server-cert" not found Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.549127 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-75hpg" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.554940 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-8497b45c89-fqrz9"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.555926 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-fqrz9" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.559274 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-lzj5j" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.560147 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-68f46476f-5qz9n"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.561674 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-68f46476f-5qz9n" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.565286 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9dhr\" (UniqueName: \"kubernetes.io/projected/d6decb51-f3f9-4183-a20a-6aa4491dba61-kube-api-access-t9dhr\") pod \"keystone-operator-controller-manager-b4d948c87-mkqrg\" (UID: \"d6decb51-f3f9-4183-a20a-6aa4491dba61\") " pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-mkqrg" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.567235 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-lhb8n" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.573571 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5td7\" (UniqueName: \"kubernetes.io/projected/9019acce-c356-421a-bd47-a163a8857f30-kube-api-access-f5td7\") pod \"ironic-operator-controller-manager-554564d7fc-h2rm6\" (UID: \"9019acce-c356-421a-bd47-a163a8857f30\") " pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-h2rm6" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.573774 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf592\" (UniqueName: \"kubernetes.io/projected/cc84cdf7-92aa-4134-a784-d821c076cde3-kube-api-access-wf592\") pod \"infra-operator-controller-manager-79d975b745-gpgxj\" (UID: \"cc84cdf7-92aa-4134-a784-d821c076cde3\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-gpgxj" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.579297 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvc5d\" (UniqueName: \"kubernetes.io/projected/01eb7df7-9e0a-4e78-85a3-0d456dec139f-kube-api-access-rvc5d\") pod \"manila-operator-controller-manager-54f6768c69-4f8kb\" (UID: \"01eb7df7-9e0a-4e78-85a3-0d456dec139f\") " pod="openstack-operators/manila-operator-controller-manager-54f6768c69-4f8kb" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.583490 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5x9m4\" (UniqueName: \"kubernetes.io/projected/6ade752d-a3d0-4449-8028-67405763ccc8-kube-api-access-5x9m4\") pod \"horizon-operator-controller-manager-5b9b8895d5-k2x44\" (UID: \"6ade752d-a3d0-4449-8028-67405763ccc8\") " pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-k2x44" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.602281 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-d44cf6b75-vq5wq"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.602477 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-k2x44" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.621812 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-68f46476f-5qz9n"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.644962 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cd2ef1c9-23d1-4671-bc71-2c077afdfe08-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cxkmq8\" (UID: \"cd2ef1c9-23d1-4671-bc71-2c077afdfe08\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cxkmq8" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.645882 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7j9bm\" (UniqueName: \"kubernetes.io/projected/cd2ef1c9-23d1-4671-bc71-2c077afdfe08-kube-api-access-7j9bm\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cxkmq8\" (UID: \"cd2ef1c9-23d1-4671-bc71-2c077afdfe08\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cxkmq8" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.645990 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqjzq\" (UniqueName: \"kubernetes.io/projected/308b7713-e37c-4890-8b56-11379bad3f8d-kube-api-access-cqjzq\") pod \"mariadb-operator-controller-manager-6994f66f48-nzn4k\" (UID: \"308b7713-e37c-4890-8b56-11379bad3f8d\") " pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-nzn4k" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.646060 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjhfm\" (UniqueName: \"kubernetes.io/projected/3f752291-5ba5-4074-a126-e99f6f0242aa-kube-api-access-zjhfm\") pod \"octavia-operator-controller-manager-69f8888797-wkdhh\" (UID: \"3f752291-5ba5-4074-a126-e99f6f0242aa\") " pod="openstack-operators/octavia-operator-controller-manager-69f8888797-wkdhh" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.646135 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8j754\" (UniqueName: \"kubernetes.io/projected/744f25f9-1641-4864-8a55-ea14ead19cae-kube-api-access-8j754\") pod \"nova-operator-controller-manager-567668f5cf-5fxtz\" (UID: \"744f25f9-1641-4864-8a55-ea14ead19cae\") " pod="openstack-operators/nova-operator-controller-manager-567668f5cf-5fxtz" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.646165 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdgn6\" (UniqueName: \"kubernetes.io/projected/17cecde4-2eb0-445e-bcf4-f73f719627df-kube-api-access-wdgn6\") pod \"ovn-operator-controller-manager-d44cf6b75-vq5wq\" (UID: \"17cecde4-2eb0-445e-bcf4-f73f719627df\") " pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-vq5wq" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.646194 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jjgc\" (UniqueName: \"kubernetes.io/projected/037562f9-c04f-4e9d-8fa5-3e98ac7c8fc5-kube-api-access-9jjgc\") pod \"neutron-operator-controller-manager-64ddbf8bb-xkvlp\" (UID: \"037562f9-c04f-4e9d-8fa5-3e98ac7c8fc5\") " pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-xkvlp" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.660419 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-h2rm6" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.640957 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-8497b45c89-fqrz9"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.676332 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-mkqrg" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.679310 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8j754\" (UniqueName: \"kubernetes.io/projected/744f25f9-1641-4864-8a55-ea14ead19cae-kube-api-access-8j754\") pod \"nova-operator-controller-manager-567668f5cf-5fxtz\" (UID: \"744f25f9-1641-4864-8a55-ea14ead19cae\") " pod="openstack-operators/nova-operator-controller-manager-567668f5cf-5fxtz" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.682481 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-q68lr"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.685807 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-q68lr" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.690122 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-kq47l" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.693337 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqjzq\" (UniqueName: \"kubernetes.io/projected/308b7713-e37c-4890-8b56-11379bad3f8d-kube-api-access-cqjzq\") pod \"mariadb-operator-controller-manager-6994f66f48-nzn4k\" (UID: \"308b7713-e37c-4890-8b56-11379bad3f8d\") " pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-nzn4k" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.694185 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-q68lr"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.701756 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jjgc\" (UniqueName: \"kubernetes.io/projected/037562f9-c04f-4e9d-8fa5-3e98ac7c8fc5-kube-api-access-9jjgc\") pod \"neutron-operator-controller-manager-64ddbf8bb-xkvlp\" (UID: \"037562f9-c04f-4e9d-8fa5-3e98ac7c8fc5\") " pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-xkvlp" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.702682 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-4f8kb" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.706421 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjhfm\" (UniqueName: \"kubernetes.io/projected/3f752291-5ba5-4074-a126-e99f6f0242aa-kube-api-access-zjhfm\") pod \"octavia-operator-controller-manager-69f8888797-wkdhh\" (UID: \"3f752291-5ba5-4074-a126-e99f6f0242aa\") " pod="openstack-operators/octavia-operator-controller-manager-69f8888797-wkdhh" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.726215 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-nzn4k" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.739961 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-xkvlp" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.745500 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-7866795846-xqxjt"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.746938 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-7866795846-xqxjt" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.747842 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdgn6\" (UniqueName: \"kubernetes.io/projected/17cecde4-2eb0-445e-bcf4-f73f719627df-kube-api-access-wdgn6\") pod \"ovn-operator-controller-manager-d44cf6b75-vq5wq\" (UID: \"17cecde4-2eb0-445e-bcf4-f73f719627df\") " pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-vq5wq" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.748142 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cd2ef1c9-23d1-4671-bc71-2c077afdfe08-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cxkmq8\" (UID: \"cd2ef1c9-23d1-4671-bc71-2c077afdfe08\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cxkmq8" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.748175 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9s5w\" (UniqueName: \"kubernetes.io/projected/60b19c88-d91f-422e-9012-ad97daf661bb-kube-api-access-j9s5w\") pod \"swift-operator-controller-manager-68f46476f-5qz9n\" (UID: \"60b19c88-d91f-422e-9012-ad97daf661bb\") " pod="openstack-operators/swift-operator-controller-manager-68f46476f-5qz9n" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.748196 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7j9bm\" (UniqueName: \"kubernetes.io/projected/cd2ef1c9-23d1-4671-bc71-2c077afdfe08-kube-api-access-7j9bm\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cxkmq8\" (UID: \"cd2ef1c9-23d1-4671-bc71-2c077afdfe08\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cxkmq8" Feb 17 13:19:26 crc kubenswrapper[4955]: E0217 13:19:26.748279 4955 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.748331 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kg9g\" (UniqueName: \"kubernetes.io/projected/788f9fac-dd10-49c4-8a25-4578d69aeb2b-kube-api-access-9kg9g\") pod \"placement-operator-controller-manager-8497b45c89-fqrz9\" (UID: \"788f9fac-dd10-49c4-8a25-4578d69aeb2b\") " pod="openstack-operators/placement-operator-controller-manager-8497b45c89-fqrz9" Feb 17 13:19:26 crc kubenswrapper[4955]: E0217 13:19:26.748359 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cd2ef1c9-23d1-4671-bc71-2c077afdfe08-cert podName:cd2ef1c9-23d1-4671-bc71-2c077afdfe08 nodeName:}" failed. No retries permitted until 2026-02-17 13:19:27.248342431 +0000 UTC m=+905.771071974 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/cd2ef1c9-23d1-4671-bc71-2c077afdfe08-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9cxkmq8" (UID: "cd2ef1c9-23d1-4671-bc71-2c077afdfe08") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.749995 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-g6kj6" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.758367 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-7866795846-xqxjt"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.766910 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7j9bm\" (UniqueName: \"kubernetes.io/projected/cd2ef1c9-23d1-4671-bc71-2c077afdfe08-kube-api-access-7j9bm\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cxkmq8\" (UID: \"cd2ef1c9-23d1-4671-bc71-2c077afdfe08\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cxkmq8" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.770272 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-2c8ww" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.772979 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdgn6\" (UniqueName: \"kubernetes.io/projected/17cecde4-2eb0-445e-bcf4-f73f719627df-kube-api-access-wdgn6\") pod \"ovn-operator-controller-manager-d44cf6b75-vq5wq\" (UID: \"17cecde4-2eb0-445e-bcf4-f73f719627df\") " pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-vq5wq" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.779066 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5db88f68c-v6jkn"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.782350 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-v6jkn" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.788922 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5db88f68c-v6jkn"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.793665 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-2ggb4" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.811205 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-5fxtz" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.829210 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-wkdhh" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.852595 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9s5w\" (UniqueName: \"kubernetes.io/projected/60b19c88-d91f-422e-9012-ad97daf661bb-kube-api-access-j9s5w\") pod \"swift-operator-controller-manager-68f46476f-5qz9n\" (UID: \"60b19c88-d91f-422e-9012-ad97daf661bb\") " pod="openstack-operators/swift-operator-controller-manager-68f46476f-5qz9n" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.852670 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvmbq\" (UniqueName: \"kubernetes.io/projected/72fe0652-eee2-4e49-83b9-3c1790e8ca9e-kube-api-access-xvmbq\") pod \"telemetry-operator-controller-manager-7f45b4ff68-q68lr\" (UID: \"72fe0652-eee2-4e49-83b9-3c1790e8ca9e\") " pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-q68lr" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.852708 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kg9g\" (UniqueName: \"kubernetes.io/projected/788f9fac-dd10-49c4-8a25-4578d69aeb2b-kube-api-access-9kg9g\") pod \"placement-operator-controller-manager-8497b45c89-fqrz9\" (UID: \"788f9fac-dd10-49c4-8a25-4578d69aeb2b\") " pod="openstack-operators/placement-operator-controller-manager-8497b45c89-fqrz9" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.852758 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwnjp\" (UniqueName: \"kubernetes.io/projected/90e44f80-9802-4134-8aa3-519bc93f4bd3-kube-api-access-rwnjp\") pod \"test-operator-controller-manager-7866795846-xqxjt\" (UID: \"90e44f80-9802-4134-8aa3-519bc93f4bd3\") " pod="openstack-operators/test-operator-controller-manager-7866795846-xqxjt" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.867811 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-vq5wq" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.876979 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kg9g\" (UniqueName: \"kubernetes.io/projected/788f9fac-dd10-49c4-8a25-4578d69aeb2b-kube-api-access-9kg9g\") pod \"placement-operator-controller-manager-8497b45c89-fqrz9\" (UID: \"788f9fac-dd10-49c4-8a25-4578d69aeb2b\") " pod="openstack-operators/placement-operator-controller-manager-8497b45c89-fqrz9" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.877140 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9s5w\" (UniqueName: \"kubernetes.io/projected/60b19c88-d91f-422e-9012-ad97daf661bb-kube-api-access-j9s5w\") pod \"swift-operator-controller-manager-68f46476f-5qz9n\" (UID: \"60b19c88-d91f-422e-9012-ad97daf661bb\") " pod="openstack-operators/swift-operator-controller-manager-68f46476f-5qz9n" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.879385 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5699594bc4-nrc2h"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.880557 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5699594bc4-nrc2h" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.882334 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.882527 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.882716 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-428dv" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.898881 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5699594bc4-nrc2h"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.902359 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-fqrz9" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.924909 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-nmxwn"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.925996 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-nmxwn" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.929190 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-qlnsv" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.933305 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-nmxwn"] Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.942983 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-68f46476f-5qz9n" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.953543 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwnjp\" (UniqueName: \"kubernetes.io/projected/90e44f80-9802-4134-8aa3-519bc93f4bd3-kube-api-access-rwnjp\") pod \"test-operator-controller-manager-7866795846-xqxjt\" (UID: \"90e44f80-9802-4134-8aa3-519bc93f4bd3\") " pod="openstack-operators/test-operator-controller-manager-7866795846-xqxjt" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.954089 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvmbq\" (UniqueName: \"kubernetes.io/projected/72fe0652-eee2-4e49-83b9-3c1790e8ca9e-kube-api-access-xvmbq\") pod \"telemetry-operator-controller-manager-7f45b4ff68-q68lr\" (UID: \"72fe0652-eee2-4e49-83b9-3c1790e8ca9e\") " pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-q68lr" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.954159 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jblnv\" (UniqueName: \"kubernetes.io/projected/a1f4d718-9535-4290-94eb-93d6288d6047-kube-api-access-jblnv\") pod \"watcher-operator-controller-manager-5db88f68c-v6jkn\" (UID: \"a1f4d718-9535-4290-94eb-93d6288d6047\") " pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-v6jkn" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.979340 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvmbq\" (UniqueName: \"kubernetes.io/projected/72fe0652-eee2-4e49-83b9-3c1790e8ca9e-kube-api-access-xvmbq\") pod \"telemetry-operator-controller-manager-7f45b4ff68-q68lr\" (UID: \"72fe0652-eee2-4e49-83b9-3c1790e8ca9e\") " pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-q68lr" Feb 17 13:19:26 crc kubenswrapper[4955]: I0217 13:19:26.979622 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwnjp\" (UniqueName: \"kubernetes.io/projected/90e44f80-9802-4134-8aa3-519bc93f4bd3-kube-api-access-rwnjp\") pod \"test-operator-controller-manager-7866795846-xqxjt\" (UID: \"90e44f80-9802-4134-8aa3-519bc93f4bd3\") " pod="openstack-operators/test-operator-controller-manager-7866795846-xqxjt" Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.057701 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jblnv\" (UniqueName: \"kubernetes.io/projected/a1f4d718-9535-4290-94eb-93d6288d6047-kube-api-access-jblnv\") pod \"watcher-operator-controller-manager-5db88f68c-v6jkn\" (UID: \"a1f4d718-9535-4290-94eb-93d6288d6047\") " pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-v6jkn" Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.057768 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c93c2a50-6396-4a32-9b88-dcdc341ec6c5-webhook-certs\") pod \"openstack-operator-controller-manager-5699594bc4-nrc2h\" (UID: \"c93c2a50-6396-4a32-9b88-dcdc341ec6c5\") " pod="openstack-operators/openstack-operator-controller-manager-5699594bc4-nrc2h" Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.057802 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpmlw\" (UniqueName: \"kubernetes.io/projected/c93c2a50-6396-4a32-9b88-dcdc341ec6c5-kube-api-access-bpmlw\") pod \"openstack-operator-controller-manager-5699594bc4-nrc2h\" (UID: \"c93c2a50-6396-4a32-9b88-dcdc341ec6c5\") " pod="openstack-operators/openstack-operator-controller-manager-5699594bc4-nrc2h" Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.057830 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cc84cdf7-92aa-4134-a784-d821c076cde3-cert\") pod \"infra-operator-controller-manager-79d975b745-gpgxj\" (UID: \"cc84cdf7-92aa-4134-a784-d821c076cde3\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-gpgxj" Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.057868 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7sq8s\" (UniqueName: \"kubernetes.io/projected/afaac24a-0894-4001-9b50-d3d7c2de90bc-kube-api-access-7sq8s\") pod \"rabbitmq-cluster-operator-manager-668c99d594-nmxwn\" (UID: \"afaac24a-0894-4001-9b50-d3d7c2de90bc\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-nmxwn" Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.057888 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c93c2a50-6396-4a32-9b88-dcdc341ec6c5-metrics-certs\") pod \"openstack-operator-controller-manager-5699594bc4-nrc2h\" (UID: \"c93c2a50-6396-4a32-9b88-dcdc341ec6c5\") " pod="openstack-operators/openstack-operator-controller-manager-5699594bc4-nrc2h" Feb 17 13:19:27 crc kubenswrapper[4955]: E0217 13:19:27.058245 4955 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 17 13:19:27 crc kubenswrapper[4955]: E0217 13:19:27.058281 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cc84cdf7-92aa-4134-a784-d821c076cde3-cert podName:cc84cdf7-92aa-4134-a784-d821c076cde3 nodeName:}" failed. No retries permitted until 2026-02-17 13:19:28.058269653 +0000 UTC m=+906.580999186 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/cc84cdf7-92aa-4134-a784-d821c076cde3-cert") pod "infra-operator-controller-manager-79d975b745-gpgxj" (UID: "cc84cdf7-92aa-4134-a784-d821c076cde3") : secret "infra-operator-webhook-server-cert" not found Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.088108 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jblnv\" (UniqueName: \"kubernetes.io/projected/a1f4d718-9535-4290-94eb-93d6288d6047-kube-api-access-jblnv\") pod \"watcher-operator-controller-manager-5db88f68c-v6jkn\" (UID: \"a1f4d718-9535-4290-94eb-93d6288d6047\") " pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-v6jkn" Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.095940 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-7866795846-xqxjt" Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.096373 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-q68lr" Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.097547 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-868647ff47-ffkxx"] Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.138164 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-v6jkn" Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.158968 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c93c2a50-6396-4a32-9b88-dcdc341ec6c5-webhook-certs\") pod \"openstack-operator-controller-manager-5699594bc4-nrc2h\" (UID: \"c93c2a50-6396-4a32-9b88-dcdc341ec6c5\") " pod="openstack-operators/openstack-operator-controller-manager-5699594bc4-nrc2h" Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.159050 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpmlw\" (UniqueName: \"kubernetes.io/projected/c93c2a50-6396-4a32-9b88-dcdc341ec6c5-kube-api-access-bpmlw\") pod \"openstack-operator-controller-manager-5699594bc4-nrc2h\" (UID: \"c93c2a50-6396-4a32-9b88-dcdc341ec6c5\") " pod="openstack-operators/openstack-operator-controller-manager-5699594bc4-nrc2h" Feb 17 13:19:27 crc kubenswrapper[4955]: E0217 13:19:27.159109 4955 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 17 13:19:27 crc kubenswrapper[4955]: E0217 13:19:27.159180 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c93c2a50-6396-4a32-9b88-dcdc341ec6c5-webhook-certs podName:c93c2a50-6396-4a32-9b88-dcdc341ec6c5 nodeName:}" failed. No retries permitted until 2026-02-17 13:19:27.659161885 +0000 UTC m=+906.181891428 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/c93c2a50-6396-4a32-9b88-dcdc341ec6c5-webhook-certs") pod "openstack-operator-controller-manager-5699594bc4-nrc2h" (UID: "c93c2a50-6396-4a32-9b88-dcdc341ec6c5") : secret "webhook-server-cert" not found Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.159217 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7sq8s\" (UniqueName: \"kubernetes.io/projected/afaac24a-0894-4001-9b50-d3d7c2de90bc-kube-api-access-7sq8s\") pod \"rabbitmq-cluster-operator-manager-668c99d594-nmxwn\" (UID: \"afaac24a-0894-4001-9b50-d3d7c2de90bc\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-nmxwn" Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.159248 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c93c2a50-6396-4a32-9b88-dcdc341ec6c5-metrics-certs\") pod \"openstack-operator-controller-manager-5699594bc4-nrc2h\" (UID: \"c93c2a50-6396-4a32-9b88-dcdc341ec6c5\") " pod="openstack-operators/openstack-operator-controller-manager-5699594bc4-nrc2h" Feb 17 13:19:27 crc kubenswrapper[4955]: E0217 13:19:27.159383 4955 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 17 13:19:27 crc kubenswrapper[4955]: E0217 13:19:27.159404 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c93c2a50-6396-4a32-9b88-dcdc341ec6c5-metrics-certs podName:c93c2a50-6396-4a32-9b88-dcdc341ec6c5 nodeName:}" failed. No retries permitted until 2026-02-17 13:19:27.659398602 +0000 UTC m=+906.182128145 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c93c2a50-6396-4a32-9b88-dcdc341ec6c5-metrics-certs") pod "openstack-operator-controller-manager-5699594bc4-nrc2h" (UID: "c93c2a50-6396-4a32-9b88-dcdc341ec6c5") : secret "metrics-server-cert" not found Feb 17 13:19:27 crc kubenswrapper[4955]: W0217 13:19:27.162699 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod602419fb_5033_4889_b587_e1eb97001910.slice/crio-a7f38b60f99215be9252bf8bc5910cf741f08b74890dec17d3d0afc549ce2757 WatchSource:0}: Error finding container a7f38b60f99215be9252bf8bc5910cf741f08b74890dec17d3d0afc549ce2757: Status 404 returned error can't find the container with id a7f38b60f99215be9252bf8bc5910cf741f08b74890dec17d3d0afc549ce2757 Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.178774 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7sq8s\" (UniqueName: \"kubernetes.io/projected/afaac24a-0894-4001-9b50-d3d7c2de90bc-kube-api-access-7sq8s\") pod \"rabbitmq-cluster-operator-manager-668c99d594-nmxwn\" (UID: \"afaac24a-0894-4001-9b50-d3d7c2de90bc\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-nmxwn" Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.179235 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpmlw\" (UniqueName: \"kubernetes.io/projected/c93c2a50-6396-4a32-9b88-dcdc341ec6c5-kube-api-access-bpmlw\") pod \"openstack-operator-controller-manager-5699594bc4-nrc2h\" (UID: \"c93c2a50-6396-4a32-9b88-dcdc341ec6c5\") " pod="openstack-operators/openstack-operator-controller-manager-5699594bc4-nrc2h" Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.217864 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5d946d989d-lpxnk"] Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.260608 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cd2ef1c9-23d1-4671-bc71-2c077afdfe08-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cxkmq8\" (UID: \"cd2ef1c9-23d1-4671-bc71-2c077afdfe08\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cxkmq8" Feb 17 13:19:27 crc kubenswrapper[4955]: E0217 13:19:27.260828 4955 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 17 13:19:27 crc kubenswrapper[4955]: E0217 13:19:27.260885 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cd2ef1c9-23d1-4671-bc71-2c077afdfe08-cert podName:cd2ef1c9-23d1-4671-bc71-2c077afdfe08 nodeName:}" failed. No retries permitted until 2026-02-17 13:19:28.26086508 +0000 UTC m=+906.783594623 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/cd2ef1c9-23d1-4671-bc71-2c077afdfe08-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9cxkmq8" (UID: "cd2ef1c9-23d1-4671-bc71-2c077afdfe08") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.296104 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-nmxwn" Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.453094 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-69f49c598c-75hpg"] Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.477999 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987464f4-g967h"] Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.485931 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-54f6768c69-4f8kb"] Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.559422 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b9b8895d5-k2x44"] Feb 17 13:19:27 crc kubenswrapper[4955]: W0217 13:19:27.574342 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod071c3b13_9304_48b6_bf22_d89dd62825a0.slice/crio-07727f6b2610c5021ed587c7df345f144dede6a472fdd316c09443224e5eaf00 WatchSource:0}: Error finding container 07727f6b2610c5021ed587c7df345f144dede6a472fdd316c09443224e5eaf00: Status 404 returned error can't find the container with id 07727f6b2610c5021ed587c7df345f144dede6a472fdd316c09443224e5eaf00 Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.591075 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6994f66f48-nzn4k"] Feb 17 13:19:27 crc kubenswrapper[4955]: W0217 13:19:27.604105 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod308b7713_e37c_4890_8b56_11379bad3f8d.slice/crio-c22196d49ec40479354a59a7d1ca9b677c7432a618555513a465b3451d0ea7c7 WatchSource:0}: Error finding container c22196d49ec40479354a59a7d1ca9b677c7432a618555513a465b3451d0ea7c7: Status 404 returned error can't find the container with id c22196d49ec40479354a59a7d1ca9b677c7432a618555513a465b3451d0ea7c7 Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.628343 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b4d948c87-mkqrg"] Feb 17 13:19:27 crc kubenswrapper[4955]: W0217 13:19:27.640363 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6decb51_f3f9_4183_a20a_6aa4491dba61.slice/crio-6b968f6988bfc06a7ea09c302f01819b9062f6f90193f0783a9cc0c750e4195c WatchSource:0}: Error finding container 6b968f6988bfc06a7ea09c302f01819b9062f6f90193f0783a9cc0c750e4195c: Status 404 returned error can't find the container with id 6b968f6988bfc06a7ea09c302f01819b9062f6f90193f0783a9cc0c750e4195c Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.679024 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c93c2a50-6396-4a32-9b88-dcdc341ec6c5-metrics-certs\") pod \"openstack-operator-controller-manager-5699594bc4-nrc2h\" (UID: \"c93c2a50-6396-4a32-9b88-dcdc341ec6c5\") " pod="openstack-operators/openstack-operator-controller-manager-5699594bc4-nrc2h" Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.679130 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c93c2a50-6396-4a32-9b88-dcdc341ec6c5-webhook-certs\") pod \"openstack-operator-controller-manager-5699594bc4-nrc2h\" (UID: \"c93c2a50-6396-4a32-9b88-dcdc341ec6c5\") " pod="openstack-operators/openstack-operator-controller-manager-5699594bc4-nrc2h" Feb 17 13:19:27 crc kubenswrapper[4955]: E0217 13:19:27.679156 4955 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 17 13:19:27 crc kubenswrapper[4955]: E0217 13:19:27.679234 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c93c2a50-6396-4a32-9b88-dcdc341ec6c5-metrics-certs podName:c93c2a50-6396-4a32-9b88-dcdc341ec6c5 nodeName:}" failed. No retries permitted until 2026-02-17 13:19:28.679211949 +0000 UTC m=+907.201941572 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c93c2a50-6396-4a32-9b88-dcdc341ec6c5-metrics-certs") pod "openstack-operator-controller-manager-5699594bc4-nrc2h" (UID: "c93c2a50-6396-4a32-9b88-dcdc341ec6c5") : secret "metrics-server-cert" not found Feb 17 13:19:27 crc kubenswrapper[4955]: E0217 13:19:27.679282 4955 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 17 13:19:27 crc kubenswrapper[4955]: E0217 13:19:27.679363 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c93c2a50-6396-4a32-9b88-dcdc341ec6c5-webhook-certs podName:c93c2a50-6396-4a32-9b88-dcdc341ec6c5 nodeName:}" failed. No retries permitted until 2026-02-17 13:19:28.679345633 +0000 UTC m=+907.202075176 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/c93c2a50-6396-4a32-9b88-dcdc341ec6c5-webhook-certs") pod "openstack-operator-controller-manager-5699594bc4-nrc2h" (UID: "c93c2a50-6396-4a32-9b88-dcdc341ec6c5") : secret "webhook-server-cert" not found Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.813819 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-d44cf6b75-vq5wq"] Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.818355 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-567668f5cf-5fxtz"] Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.819366 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-4f8kb" event={"ID":"01eb7df7-9e0a-4e78-85a3-0d456dec139f","Type":"ContainerStarted","Data":"a7539c8a1b82d5776bf8c0e38f86ae78e90de49e468ce6a72f975482ff259746"} Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.822932 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-75hpg" event={"ID":"071c3b13-9304-48b6-bf22-d89dd62825a0","Type":"ContainerStarted","Data":"07727f6b2610c5021ed587c7df345f144dede6a472fdd316c09443224e5eaf00"} Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.823723 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f8888797-wkdhh"] Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.824535 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-k2x44" event={"ID":"6ade752d-a3d0-4449-8028-67405763ccc8","Type":"ContainerStarted","Data":"9cc367731afda9bb5213927510a887e40e4281a420ea0d14a45c1789db9efcdf"} Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.830527 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-64ddbf8bb-xkvlp"] Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.832362 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-mkqrg" event={"ID":"d6decb51-f3f9-4183-a20a-6aa4491dba61","Type":"ContainerStarted","Data":"6b968f6988bfc06a7ea09c302f01819b9062f6f90193f0783a9cc0c750e4195c"} Feb 17 13:19:27 crc kubenswrapper[4955]: W0217 13:19:27.832529 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17cecde4_2eb0_445e_bcf4_f73f719627df.slice/crio-ce1afb19f947665cb654ab41bed37f5b1d293a0660ee975d9938dfbffbfeec30 WatchSource:0}: Error finding container ce1afb19f947665cb654ab41bed37f5b1d293a0660ee975d9938dfbffbfeec30: Status 404 returned error can't find the container with id ce1afb19f947665cb654ab41bed37f5b1d293a0660ee975d9938dfbffbfeec30 Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.833805 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-nzn4k" event={"ID":"308b7713-e37c-4890-8b56-11379bad3f8d","Type":"ContainerStarted","Data":"c22196d49ec40479354a59a7d1ca9b677c7432a618555513a465b3451d0ea7c7"} Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.836017 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-ffkxx" event={"ID":"602419fb-5033-4889-b587-e1eb97001910","Type":"ContainerStarted","Data":"a7f38b60f99215be9252bf8bc5910cf741f08b74890dec17d3d0afc549ce2757"} Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.837749 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987464f4-g967h" event={"ID":"5c339089-ba43-481b-aaa8-f8bf09fd106d","Type":"ContainerStarted","Data":"ee2f19c6b0da5d34661cd2f3367cb0fba030591f4c6c9f77158c489b18497efa"} Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.846451 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-554564d7fc-h2rm6"] Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.849524 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-lpxnk" event={"ID":"6d1e254f-7d68-45c5-bda0-e3b33155ae56","Type":"ContainerStarted","Data":"dc402856fd2ed48cec9f38be81aee4a57fe8a6a46dedcd40ac82b01b86ccb8a8"} Feb 17 13:19:27 crc kubenswrapper[4955]: W0217 13:19:27.856750 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod037562f9_c04f_4e9d_8fa5_3e98ac7c8fc5.slice/crio-7b8f57da426193b6b90e91eedb08e7c48e7cda39626a113ab0bb3eb451e05e06 WatchSource:0}: Error finding container 7b8f57da426193b6b90e91eedb08e7c48e7cda39626a113ab0bb3eb451e05e06: Status 404 returned error can't find the container with id 7b8f57da426193b6b90e91eedb08e7c48e7cda39626a113ab0bb3eb451e05e06 Feb 17 13:19:27 crc kubenswrapper[4955]: W0217 13:19:27.867230 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9019acce_c356_421a_bd47_a163a8857f30.slice/crio-abb5510f2dae340d9724402dffa35b9bfd26a5d131cfcba6667d71ab308abc65 WatchSource:0}: Error finding container abb5510f2dae340d9724402dffa35b9bfd26a5d131cfcba6667d71ab308abc65: Status 404 returned error can't find the container with id abb5510f2dae340d9724402dffa35b9bfd26a5d131cfcba6667d71ab308abc65 Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.970481 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d8bf5c495-2c8ww"] Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.979863 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-68f46476f-5qz9n"] Feb 17 13:19:27 crc kubenswrapper[4955]: I0217 13:19:27.984156 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-8497b45c89-fqrz9"] Feb 17 13:19:27 crc kubenswrapper[4955]: E0217 13:19:27.991614 4955 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:c1e33e962043cd6e3d09ebd225cb72781451dba7af2d57522e5c6eedbdc91642,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rm7sb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-6d8bf5c495-2c8ww_openstack-operators(1e4fd408-9f09-4289-92ce-7c5ae18f235a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 17 13:19:27 crc kubenswrapper[4955]: E0217 13:19:27.992912 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-2c8ww" podUID="1e4fd408-9f09-4289-92ce-7c5ae18f235a" Feb 17 13:19:27 crc kubenswrapper[4955]: E0217 13:19:27.993856 4955 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:3d676f1281e24ef07de617570d2f7fbf625032e41866d1551a856c052248bb04,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-j9s5w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-68f46476f-5qz9n_openstack-operators(60b19c88-d91f-422e-9012-ad97daf661bb): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 17 13:19:27 crc kubenswrapper[4955]: E0217 13:19:27.994546 4955 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:a57336b9f95b703f80453db87e43a2834ca1bdc89480796d28ebbe0a9702ecfd,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9kg9g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-8497b45c89-fqrz9_openstack-operators(788f9fac-dd10-49c4-8a25-4578d69aeb2b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 17 13:19:27 crc kubenswrapper[4955]: E0217 13:19:27.994948 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-68f46476f-5qz9n" podUID="60b19c88-d91f-422e-9012-ad97daf661bb" Feb 17 13:19:27 crc kubenswrapper[4955]: E0217 13:19:27.995655 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-fqrz9" podUID="788f9fac-dd10-49c4-8a25-4578d69aeb2b" Feb 17 13:19:28 crc kubenswrapper[4955]: I0217 13:19:28.083559 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cc84cdf7-92aa-4134-a784-d821c076cde3-cert\") pod \"infra-operator-controller-manager-79d975b745-gpgxj\" (UID: \"cc84cdf7-92aa-4134-a784-d821c076cde3\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-gpgxj" Feb 17 13:19:28 crc kubenswrapper[4955]: E0217 13:19:28.083759 4955 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 17 13:19:28 crc kubenswrapper[4955]: E0217 13:19:28.083852 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cc84cdf7-92aa-4134-a784-d821c076cde3-cert podName:cc84cdf7-92aa-4134-a784-d821c076cde3 nodeName:}" failed. No retries permitted until 2026-02-17 13:19:30.083832675 +0000 UTC m=+908.606562218 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/cc84cdf7-92aa-4134-a784-d821c076cde3-cert") pod "infra-operator-controller-manager-79d975b745-gpgxj" (UID: "cc84cdf7-92aa-4134-a784-d821c076cde3") : secret "infra-operator-webhook-server-cert" not found Feb 17 13:19:28 crc kubenswrapper[4955]: I0217 13:19:28.114960 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-7866795846-xqxjt"] Feb 17 13:19:28 crc kubenswrapper[4955]: I0217 13:19:28.119627 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5db88f68c-v6jkn"] Feb 17 13:19:28 crc kubenswrapper[4955]: W0217 13:19:28.119871 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod72fe0652_eee2_4e49_83b9_3c1790e8ca9e.slice/crio-a0cac26e1fd26907f4911578623b9546da9a5da4edc443ed3aab4070edbff435 WatchSource:0}: Error finding container a0cac26e1fd26907f4911578623b9546da9a5da4edc443ed3aab4070edbff435: Status 404 returned error can't find the container with id a0cac26e1fd26907f4911578623b9546da9a5da4edc443ed3aab4070edbff435 Feb 17 13:19:28 crc kubenswrapper[4955]: I0217 13:19:28.123839 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-q68lr"] Feb 17 13:19:28 crc kubenswrapper[4955]: I0217 13:19:28.136966 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-nmxwn"] Feb 17 13:19:28 crc kubenswrapper[4955]: W0217 13:19:28.145047 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod90e44f80_9802_4134_8aa3_519bc93f4bd3.slice/crio-bc3f751f317cbeb6c485d23818515c3eacc486965a16be839d196cdf8ff87d2f WatchSource:0}: Error finding container bc3f751f317cbeb6c485d23818515c3eacc486965a16be839d196cdf8ff87d2f: Status 404 returned error can't find the container with id bc3f751f317cbeb6c485d23818515c3eacc486965a16be839d196cdf8ff87d2f Feb 17 13:19:28 crc kubenswrapper[4955]: W0217 13:19:28.145569 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda1f4d718_9535_4290_94eb_93d6288d6047.slice/crio-634570a74fcbb685aa9ee2c7091a708314b88506017a62f78932fa6884e5b601 WatchSource:0}: Error finding container 634570a74fcbb685aa9ee2c7091a708314b88506017a62f78932fa6884e5b601: Status 404 returned error can't find the container with id 634570a74fcbb685aa9ee2c7091a708314b88506017a62f78932fa6884e5b601 Feb 17 13:19:28 crc kubenswrapper[4955]: W0217 13:19:28.147923 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podafaac24a_0894_4001_9b50_d3d7c2de90bc.slice/crio-71063c2853bd783764ddb29517ed3d5d76515dde6366347686dcc8b98e68b52b WatchSource:0}: Error finding container 71063c2853bd783764ddb29517ed3d5d76515dde6366347686dcc8b98e68b52b: Status 404 returned error can't find the container with id 71063c2853bd783764ddb29517ed3d5d76515dde6366347686dcc8b98e68b52b Feb 17 13:19:28 crc kubenswrapper[4955]: E0217 13:19:28.151765 4955 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:f0fabdf79095def0f8b1c0442925548a94ca94bed4de2d3b171277129f8079e6,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rwnjp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-7866795846-xqxjt_openstack-operators(90e44f80-9802-4134-8aa3-519bc93f4bd3): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 17 13:19:28 crc kubenswrapper[4955]: E0217 13:19:28.153623 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-7866795846-xqxjt" podUID="90e44f80-9802-4134-8aa3-519bc93f4bd3" Feb 17 13:19:28 crc kubenswrapper[4955]: E0217 13:19:28.153768 4955 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:d01ae848290e880c09127d5297418dea40fc7f090fdab9bf2c578c7e7f53aec0,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jblnv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-5db88f68c-v6jkn_openstack-operators(a1f4d718-9535-4290-94eb-93d6288d6047): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 17 13:19:28 crc kubenswrapper[4955]: E0217 13:19:28.154954 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-v6jkn" podUID="a1f4d718-9535-4290-94eb-93d6288d6047" Feb 17 13:19:28 crc kubenswrapper[4955]: E0217 13:19:28.155639 4955 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7sq8s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-nmxwn_openstack-operators(afaac24a-0894-4001-9b50-d3d7c2de90bc): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 17 13:19:28 crc kubenswrapper[4955]: E0217 13:19:28.156717 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-nmxwn" podUID="afaac24a-0894-4001-9b50-d3d7c2de90bc" Feb 17 13:19:28 crc kubenswrapper[4955]: I0217 13:19:28.286322 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cd2ef1c9-23d1-4671-bc71-2c077afdfe08-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cxkmq8\" (UID: \"cd2ef1c9-23d1-4671-bc71-2c077afdfe08\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cxkmq8" Feb 17 13:19:28 crc kubenswrapper[4955]: E0217 13:19:28.287894 4955 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 17 13:19:28 crc kubenswrapper[4955]: E0217 13:19:28.287977 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cd2ef1c9-23d1-4671-bc71-2c077afdfe08-cert podName:cd2ef1c9-23d1-4671-bc71-2c077afdfe08 nodeName:}" failed. No retries permitted until 2026-02-17 13:19:30.287958955 +0000 UTC m=+908.810688498 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/cd2ef1c9-23d1-4671-bc71-2c077afdfe08-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9cxkmq8" (UID: "cd2ef1c9-23d1-4671-bc71-2c077afdfe08") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 17 13:19:28 crc kubenswrapper[4955]: I0217 13:19:28.694577 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c93c2a50-6396-4a32-9b88-dcdc341ec6c5-metrics-certs\") pod \"openstack-operator-controller-manager-5699594bc4-nrc2h\" (UID: \"c93c2a50-6396-4a32-9b88-dcdc341ec6c5\") " pod="openstack-operators/openstack-operator-controller-manager-5699594bc4-nrc2h" Feb 17 13:19:28 crc kubenswrapper[4955]: I0217 13:19:28.694672 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c93c2a50-6396-4a32-9b88-dcdc341ec6c5-webhook-certs\") pod \"openstack-operator-controller-manager-5699594bc4-nrc2h\" (UID: \"c93c2a50-6396-4a32-9b88-dcdc341ec6c5\") " pod="openstack-operators/openstack-operator-controller-manager-5699594bc4-nrc2h" Feb 17 13:19:28 crc kubenswrapper[4955]: E0217 13:19:28.694919 4955 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 17 13:19:28 crc kubenswrapper[4955]: E0217 13:19:28.695259 4955 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 17 13:19:28 crc kubenswrapper[4955]: E0217 13:19:28.695271 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c93c2a50-6396-4a32-9b88-dcdc341ec6c5-metrics-certs podName:c93c2a50-6396-4a32-9b88-dcdc341ec6c5 nodeName:}" failed. No retries permitted until 2026-02-17 13:19:30.695197467 +0000 UTC m=+909.217927010 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c93c2a50-6396-4a32-9b88-dcdc341ec6c5-metrics-certs") pod "openstack-operator-controller-manager-5699594bc4-nrc2h" (UID: "c93c2a50-6396-4a32-9b88-dcdc341ec6c5") : secret "metrics-server-cert" not found Feb 17 13:19:28 crc kubenswrapper[4955]: E0217 13:19:28.695396 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c93c2a50-6396-4a32-9b88-dcdc341ec6c5-webhook-certs podName:c93c2a50-6396-4a32-9b88-dcdc341ec6c5 nodeName:}" failed. No retries permitted until 2026-02-17 13:19:30.695381562 +0000 UTC m=+909.218111105 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/c93c2a50-6396-4a32-9b88-dcdc341ec6c5-webhook-certs") pod "openstack-operator-controller-manager-5699594bc4-nrc2h" (UID: "c93c2a50-6396-4a32-9b88-dcdc341ec6c5") : secret "webhook-server-cert" not found Feb 17 13:19:28 crc kubenswrapper[4955]: I0217 13:19:28.861763 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-fqrz9" event={"ID":"788f9fac-dd10-49c4-8a25-4578d69aeb2b","Type":"ContainerStarted","Data":"9aea15c12a7ed557cfa0d1f6507bd9b0590891e09ec0e66675ccee4798d10b86"} Feb 17 13:19:28 crc kubenswrapper[4955]: I0217 13:19:28.863596 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-vq5wq" event={"ID":"17cecde4-2eb0-445e-bcf4-f73f719627df","Type":"ContainerStarted","Data":"ce1afb19f947665cb654ab41bed37f5b1d293a0660ee975d9938dfbffbfeec30"} Feb 17 13:19:28 crc kubenswrapper[4955]: I0217 13:19:28.865729 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-xkvlp" event={"ID":"037562f9-c04f-4e9d-8fa5-3e98ac7c8fc5","Type":"ContainerStarted","Data":"7b8f57da426193b6b90e91eedb08e7c48e7cda39626a113ab0bb3eb451e05e06"} Feb 17 13:19:28 crc kubenswrapper[4955]: I0217 13:19:28.867852 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-nmxwn" event={"ID":"afaac24a-0894-4001-9b50-d3d7c2de90bc","Type":"ContainerStarted","Data":"71063c2853bd783764ddb29517ed3d5d76515dde6366347686dcc8b98e68b52b"} Feb 17 13:19:28 crc kubenswrapper[4955]: I0217 13:19:28.869116 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-wkdhh" event={"ID":"3f752291-5ba5-4074-a126-e99f6f0242aa","Type":"ContainerStarted","Data":"cdbfbf7cb39b9b4a670998dc545271a1fb2cc274a60a8ebfda9deb693c4b7abd"} Feb 17 13:19:28 crc kubenswrapper[4955]: E0217 13:19:28.870140 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-nmxwn" podUID="afaac24a-0894-4001-9b50-d3d7c2de90bc" Feb 17 13:19:28 crc kubenswrapper[4955]: E0217 13:19:28.870276 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:a57336b9f95b703f80453db87e43a2834ca1bdc89480796d28ebbe0a9702ecfd\\\"\"" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-fqrz9" podUID="788f9fac-dd10-49c4-8a25-4578d69aeb2b" Feb 17 13:19:28 crc kubenswrapper[4955]: I0217 13:19:28.872196 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-v6jkn" event={"ID":"a1f4d718-9535-4290-94eb-93d6288d6047","Type":"ContainerStarted","Data":"634570a74fcbb685aa9ee2c7091a708314b88506017a62f78932fa6884e5b601"} Feb 17 13:19:28 crc kubenswrapper[4955]: I0217 13:19:28.875244 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-h2rm6" event={"ID":"9019acce-c356-421a-bd47-a163a8857f30","Type":"ContainerStarted","Data":"abb5510f2dae340d9724402dffa35b9bfd26a5d131cfcba6667d71ab308abc65"} Feb 17 13:19:28 crc kubenswrapper[4955]: E0217 13:19:28.877578 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:d01ae848290e880c09127d5297418dea40fc7f090fdab9bf2c578c7e7f53aec0\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-v6jkn" podUID="a1f4d718-9535-4290-94eb-93d6288d6047" Feb 17 13:19:28 crc kubenswrapper[4955]: I0217 13:19:28.879464 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-7866795846-xqxjt" event={"ID":"90e44f80-9802-4134-8aa3-519bc93f4bd3","Type":"ContainerStarted","Data":"bc3f751f317cbeb6c485d23818515c3eacc486965a16be839d196cdf8ff87d2f"} Feb 17 13:19:28 crc kubenswrapper[4955]: E0217 13:19:28.881411 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:f0fabdf79095def0f8b1c0442925548a94ca94bed4de2d3b171277129f8079e6\\\"\"" pod="openstack-operators/test-operator-controller-manager-7866795846-xqxjt" podUID="90e44f80-9802-4134-8aa3-519bc93f4bd3" Feb 17 13:19:28 crc kubenswrapper[4955]: I0217 13:19:28.882373 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-2c8ww" event={"ID":"1e4fd408-9f09-4289-92ce-7c5ae18f235a","Type":"ContainerStarted","Data":"c7070c78bea8941ac2ae60fdb1227f9c491e1da1cc191f6e76a74a6d98383219"} Feb 17 13:19:28 crc kubenswrapper[4955]: E0217 13:19:28.884077 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/designate-operator@sha256:c1e33e962043cd6e3d09ebd225cb72781451dba7af2d57522e5c6eedbdc91642\\\"\"" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-2c8ww" podUID="1e4fd408-9f09-4289-92ce-7c5ae18f235a" Feb 17 13:19:28 crc kubenswrapper[4955]: I0217 13:19:28.886049 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-q68lr" event={"ID":"72fe0652-eee2-4e49-83b9-3c1790e8ca9e","Type":"ContainerStarted","Data":"a0cac26e1fd26907f4911578623b9546da9a5da4edc443ed3aab4070edbff435"} Feb 17 13:19:28 crc kubenswrapper[4955]: I0217 13:19:28.902621 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-68f46476f-5qz9n" event={"ID":"60b19c88-d91f-422e-9012-ad97daf661bb","Type":"ContainerStarted","Data":"aa58344e7ed730dc8855623e187d87b59c6f7e21dab6bde37270041577ce0711"} Feb 17 13:19:28 crc kubenswrapper[4955]: I0217 13:19:28.904343 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-5fxtz" event={"ID":"744f25f9-1641-4864-8a55-ea14ead19cae","Type":"ContainerStarted","Data":"6e6ef006f5200b83291eaeeae7e36a215c17758aaff097bb50d3213da0248c8b"} Feb 17 13:19:28 crc kubenswrapper[4955]: E0217 13:19:28.904581 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3d676f1281e24ef07de617570d2f7fbf625032e41866d1551a856c052248bb04\\\"\"" pod="openstack-operators/swift-operator-controller-manager-68f46476f-5qz9n" podUID="60b19c88-d91f-422e-9012-ad97daf661bb" Feb 17 13:19:29 crc kubenswrapper[4955]: E0217 13:19:29.944978 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:d01ae848290e880c09127d5297418dea40fc7f090fdab9bf2c578c7e7f53aec0\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-v6jkn" podUID="a1f4d718-9535-4290-94eb-93d6288d6047" Feb 17 13:19:29 crc kubenswrapper[4955]: E0217 13:19:29.945605 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-nmxwn" podUID="afaac24a-0894-4001-9b50-d3d7c2de90bc" Feb 17 13:19:29 crc kubenswrapper[4955]: E0217 13:19:29.945615 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:a57336b9f95b703f80453db87e43a2834ca1bdc89480796d28ebbe0a9702ecfd\\\"\"" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-fqrz9" podUID="788f9fac-dd10-49c4-8a25-4578d69aeb2b" Feb 17 13:19:29 crc kubenswrapper[4955]: E0217 13:19:29.945683 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3d676f1281e24ef07de617570d2f7fbf625032e41866d1551a856c052248bb04\\\"\"" pod="openstack-operators/swift-operator-controller-manager-68f46476f-5qz9n" podUID="60b19c88-d91f-422e-9012-ad97daf661bb" Feb 17 13:19:29 crc kubenswrapper[4955]: E0217 13:19:29.946879 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/designate-operator@sha256:c1e33e962043cd6e3d09ebd225cb72781451dba7af2d57522e5c6eedbdc91642\\\"\"" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-2c8ww" podUID="1e4fd408-9f09-4289-92ce-7c5ae18f235a" Feb 17 13:19:29 crc kubenswrapper[4955]: E0217 13:19:29.946938 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:f0fabdf79095def0f8b1c0442925548a94ca94bed4de2d3b171277129f8079e6\\\"\"" pod="openstack-operators/test-operator-controller-manager-7866795846-xqxjt" podUID="90e44f80-9802-4134-8aa3-519bc93f4bd3" Feb 17 13:19:30 crc kubenswrapper[4955]: I0217 13:19:30.130740 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cc84cdf7-92aa-4134-a784-d821c076cde3-cert\") pod \"infra-operator-controller-manager-79d975b745-gpgxj\" (UID: \"cc84cdf7-92aa-4134-a784-d821c076cde3\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-gpgxj" Feb 17 13:19:30 crc kubenswrapper[4955]: E0217 13:19:30.130941 4955 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 17 13:19:30 crc kubenswrapper[4955]: E0217 13:19:30.130993 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cc84cdf7-92aa-4134-a784-d821c076cde3-cert podName:cc84cdf7-92aa-4134-a784-d821c076cde3 nodeName:}" failed. No retries permitted until 2026-02-17 13:19:34.130978507 +0000 UTC m=+912.653708050 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/cc84cdf7-92aa-4134-a784-d821c076cde3-cert") pod "infra-operator-controller-manager-79d975b745-gpgxj" (UID: "cc84cdf7-92aa-4134-a784-d821c076cde3") : secret "infra-operator-webhook-server-cert" not found Feb 17 13:19:30 crc kubenswrapper[4955]: I0217 13:19:30.333495 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cd2ef1c9-23d1-4671-bc71-2c077afdfe08-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cxkmq8\" (UID: \"cd2ef1c9-23d1-4671-bc71-2c077afdfe08\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cxkmq8" Feb 17 13:19:30 crc kubenswrapper[4955]: E0217 13:19:30.333748 4955 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 17 13:19:30 crc kubenswrapper[4955]: E0217 13:19:30.333873 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cd2ef1c9-23d1-4671-bc71-2c077afdfe08-cert podName:cd2ef1c9-23d1-4671-bc71-2c077afdfe08 nodeName:}" failed. No retries permitted until 2026-02-17 13:19:34.333849091 +0000 UTC m=+912.856578694 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/cd2ef1c9-23d1-4671-bc71-2c077afdfe08-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9cxkmq8" (UID: "cd2ef1c9-23d1-4671-bc71-2c077afdfe08") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 17 13:19:30 crc kubenswrapper[4955]: I0217 13:19:30.739313 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c93c2a50-6396-4a32-9b88-dcdc341ec6c5-metrics-certs\") pod \"openstack-operator-controller-manager-5699594bc4-nrc2h\" (UID: \"c93c2a50-6396-4a32-9b88-dcdc341ec6c5\") " pod="openstack-operators/openstack-operator-controller-manager-5699594bc4-nrc2h" Feb 17 13:19:30 crc kubenswrapper[4955]: E0217 13:19:30.739490 4955 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 17 13:19:30 crc kubenswrapper[4955]: E0217 13:19:30.739563 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c93c2a50-6396-4a32-9b88-dcdc341ec6c5-metrics-certs podName:c93c2a50-6396-4a32-9b88-dcdc341ec6c5 nodeName:}" failed. No retries permitted until 2026-02-17 13:19:34.739547359 +0000 UTC m=+913.262276902 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c93c2a50-6396-4a32-9b88-dcdc341ec6c5-metrics-certs") pod "openstack-operator-controller-manager-5699594bc4-nrc2h" (UID: "c93c2a50-6396-4a32-9b88-dcdc341ec6c5") : secret "metrics-server-cert" not found Feb 17 13:19:30 crc kubenswrapper[4955]: I0217 13:19:30.739750 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c93c2a50-6396-4a32-9b88-dcdc341ec6c5-webhook-certs\") pod \"openstack-operator-controller-manager-5699594bc4-nrc2h\" (UID: \"c93c2a50-6396-4a32-9b88-dcdc341ec6c5\") " pod="openstack-operators/openstack-operator-controller-manager-5699594bc4-nrc2h" Feb 17 13:19:30 crc kubenswrapper[4955]: E0217 13:19:30.739921 4955 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 17 13:19:30 crc kubenswrapper[4955]: E0217 13:19:30.740017 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c93c2a50-6396-4a32-9b88-dcdc341ec6c5-webhook-certs podName:c93c2a50-6396-4a32-9b88-dcdc341ec6c5 nodeName:}" failed. No retries permitted until 2026-02-17 13:19:34.739994491 +0000 UTC m=+913.262724114 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/c93c2a50-6396-4a32-9b88-dcdc341ec6c5-webhook-certs") pod "openstack-operator-controller-manager-5699594bc4-nrc2h" (UID: "c93c2a50-6396-4a32-9b88-dcdc341ec6c5") : secret "webhook-server-cert" not found Feb 17 13:19:34 crc kubenswrapper[4955]: I0217 13:19:34.196940 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cc84cdf7-92aa-4134-a784-d821c076cde3-cert\") pod \"infra-operator-controller-manager-79d975b745-gpgxj\" (UID: \"cc84cdf7-92aa-4134-a784-d821c076cde3\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-gpgxj" Feb 17 13:19:34 crc kubenswrapper[4955]: E0217 13:19:34.197124 4955 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 17 13:19:34 crc kubenswrapper[4955]: E0217 13:19:34.198622 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cc84cdf7-92aa-4134-a784-d821c076cde3-cert podName:cc84cdf7-92aa-4134-a784-d821c076cde3 nodeName:}" failed. No retries permitted until 2026-02-17 13:19:42.198601555 +0000 UTC m=+920.721331098 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/cc84cdf7-92aa-4134-a784-d821c076cde3-cert") pod "infra-operator-controller-manager-79d975b745-gpgxj" (UID: "cc84cdf7-92aa-4134-a784-d821c076cde3") : secret "infra-operator-webhook-server-cert" not found Feb 17 13:19:34 crc kubenswrapper[4955]: I0217 13:19:34.401455 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cd2ef1c9-23d1-4671-bc71-2c077afdfe08-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cxkmq8\" (UID: \"cd2ef1c9-23d1-4671-bc71-2c077afdfe08\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cxkmq8" Feb 17 13:19:34 crc kubenswrapper[4955]: E0217 13:19:34.401657 4955 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 17 13:19:34 crc kubenswrapper[4955]: E0217 13:19:34.401771 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cd2ef1c9-23d1-4671-bc71-2c077afdfe08-cert podName:cd2ef1c9-23d1-4671-bc71-2c077afdfe08 nodeName:}" failed. No retries permitted until 2026-02-17 13:19:42.401748857 +0000 UTC m=+920.924478480 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/cd2ef1c9-23d1-4671-bc71-2c077afdfe08-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9cxkmq8" (UID: "cd2ef1c9-23d1-4671-bc71-2c077afdfe08") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 17 13:19:34 crc kubenswrapper[4955]: I0217 13:19:34.574852 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 13:19:34 crc kubenswrapper[4955]: I0217 13:19:34.574913 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 13:19:34 crc kubenswrapper[4955]: I0217 13:19:34.806391 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c93c2a50-6396-4a32-9b88-dcdc341ec6c5-webhook-certs\") pod \"openstack-operator-controller-manager-5699594bc4-nrc2h\" (UID: \"c93c2a50-6396-4a32-9b88-dcdc341ec6c5\") " pod="openstack-operators/openstack-operator-controller-manager-5699594bc4-nrc2h" Feb 17 13:19:34 crc kubenswrapper[4955]: I0217 13:19:34.806507 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c93c2a50-6396-4a32-9b88-dcdc341ec6c5-metrics-certs\") pod \"openstack-operator-controller-manager-5699594bc4-nrc2h\" (UID: \"c93c2a50-6396-4a32-9b88-dcdc341ec6c5\") " pod="openstack-operators/openstack-operator-controller-manager-5699594bc4-nrc2h" Feb 17 13:19:34 crc kubenswrapper[4955]: E0217 13:19:34.806584 4955 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 17 13:19:34 crc kubenswrapper[4955]: E0217 13:19:34.806680 4955 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 17 13:19:34 crc kubenswrapper[4955]: E0217 13:19:34.806684 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c93c2a50-6396-4a32-9b88-dcdc341ec6c5-webhook-certs podName:c93c2a50-6396-4a32-9b88-dcdc341ec6c5 nodeName:}" failed. No retries permitted until 2026-02-17 13:19:42.806658641 +0000 UTC m=+921.329388194 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/c93c2a50-6396-4a32-9b88-dcdc341ec6c5-webhook-certs") pod "openstack-operator-controller-manager-5699594bc4-nrc2h" (UID: "c93c2a50-6396-4a32-9b88-dcdc341ec6c5") : secret "webhook-server-cert" not found Feb 17 13:19:34 crc kubenswrapper[4955]: E0217 13:19:34.806745 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c93c2a50-6396-4a32-9b88-dcdc341ec6c5-metrics-certs podName:c93c2a50-6396-4a32-9b88-dcdc341ec6c5 nodeName:}" failed. No retries permitted until 2026-02-17 13:19:42.806730873 +0000 UTC m=+921.329460416 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c93c2a50-6396-4a32-9b88-dcdc341ec6c5-metrics-certs") pod "openstack-operator-controller-manager-5699594bc4-nrc2h" (UID: "c93c2a50-6396-4a32-9b88-dcdc341ec6c5") : secret "metrics-server-cert" not found Feb 17 13:19:41 crc kubenswrapper[4955]: E0217 13:19:41.138586 4955 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:c6ad383f55f955902b074d1ee947a2233a5fcbf40698479ae693ce056c80dcc1" Feb 17 13:19:41 crc kubenswrapper[4955]: E0217 13:19:41.139383 4955 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:c6ad383f55f955902b074d1ee947a2233a5fcbf40698479ae693ce056c80dcc1,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-t9dhr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-b4d948c87-mkqrg_openstack-operators(d6decb51-f3f9-4183-a20a-6aa4491dba61): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 17 13:19:41 crc kubenswrapper[4955]: E0217 13:19:41.140577 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-mkqrg" podUID="d6decb51-f3f9-4183-a20a-6aa4491dba61" Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.024173 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-ffkxx" event={"ID":"602419fb-5033-4889-b587-e1eb97001910","Type":"ContainerStarted","Data":"b9da6c2cc4335cffa9a51d35ff65f9e2a675f8b0ce1dcbc58d7fbb88c48fa601"} Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.024492 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-ffkxx" Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.031357 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-wkdhh" event={"ID":"3f752291-5ba5-4074-a126-e99f6f0242aa","Type":"ContainerStarted","Data":"98d64a771e336bc550ac4b51513c874fce12b17f9b42f55786785e57e0a4c292"} Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.031472 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-wkdhh" Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.038282 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-h2rm6" event={"ID":"9019acce-c356-421a-bd47-a163a8857f30","Type":"ContainerStarted","Data":"46dc1fd2cd297b7d925c3bac97018851383ac95c9d0430f434754110a33580e3"} Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.038402 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-h2rm6" Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.043640 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-nzn4k" event={"ID":"308b7713-e37c-4890-8b56-11379bad3f8d","Type":"ContainerStarted","Data":"5013b26129b091a445a42042520e30b5eff54e2d6be27ccea44888a8869cdf79"} Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.043723 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-nzn4k" Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.047626 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-4f8kb" event={"ID":"01eb7df7-9e0a-4e78-85a3-0d456dec139f","Type":"ContainerStarted","Data":"b41a9c75bdd093cf5bba7d92e3186e99cf2f35f845ef4f96932e147615a96ebd"} Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.047717 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-4f8kb" Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.050766 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-ffkxx" podStartSLOduration=2.104835547 podStartE2EDuration="16.050752656s" podCreationTimestamp="2026-02-17 13:19:26 +0000 UTC" firstStartedPulling="2026-02-17 13:19:27.197396851 +0000 UTC m=+905.720126394" lastFinishedPulling="2026-02-17 13:19:41.14331396 +0000 UTC m=+919.666043503" observedRunningTime="2026-02-17 13:19:42.042803698 +0000 UTC m=+920.565533241" watchObservedRunningTime="2026-02-17 13:19:42.050752656 +0000 UTC m=+920.573482199" Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.051156 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-vq5wq" event={"ID":"17cecde4-2eb0-445e-bcf4-f73f719627df","Type":"ContainerStarted","Data":"a940fd4dc2052440d4b16359efbe6992effdb4fab2847fe7eb5fa35b48f49b7c"} Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.051291 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-vq5wq" Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.057918 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-wkdhh" podStartSLOduration=2.757215093 podStartE2EDuration="16.057900861s" podCreationTimestamp="2026-02-17 13:19:26 +0000 UTC" firstStartedPulling="2026-02-17 13:19:27.842630982 +0000 UTC m=+906.365360535" lastFinishedPulling="2026-02-17 13:19:41.14331676 +0000 UTC m=+919.666046303" observedRunningTime="2026-02-17 13:19:42.057804568 +0000 UTC m=+920.580534111" watchObservedRunningTime="2026-02-17 13:19:42.057900861 +0000 UTC m=+920.580630404" Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.059556 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-xkvlp" event={"ID":"037562f9-c04f-4e9d-8fa5-3e98ac7c8fc5","Type":"ContainerStarted","Data":"48fa9575b5d54d81480282b24c6386f179f9624ef2730770ea1d7a6265d909e3"} Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.059655 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-xkvlp" Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.062961 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-q68lr" event={"ID":"72fe0652-eee2-4e49-83b9-3c1790e8ca9e","Type":"ContainerStarted","Data":"503080ab8b92449fee5d52c7b7933387c295e8f7b252a9b1c9113a5d5fdfad97"} Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.063075 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-q68lr" Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.065405 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-k2x44" event={"ID":"6ade752d-a3d0-4449-8028-67405763ccc8","Type":"ContainerStarted","Data":"62cef0b83896886ef1bded7ab33bce023ddac72681a5c04c26a5be50116e79da"} Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.065531 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-k2x44" Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.075024 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-5fxtz" event={"ID":"744f25f9-1641-4864-8a55-ea14ead19cae","Type":"ContainerStarted","Data":"f9433a92994025c0c5f642e59320c954c0e14c64e058bbca4bd3fb92c73a2032"} Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.075285 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-5fxtz" Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.079668 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-h2rm6" podStartSLOduration=2.839061939 podStartE2EDuration="16.079652095s" podCreationTimestamp="2026-02-17 13:19:26 +0000 UTC" firstStartedPulling="2026-02-17 13:19:27.872505158 +0000 UTC m=+906.395234701" lastFinishedPulling="2026-02-17 13:19:41.113095314 +0000 UTC m=+919.635824857" observedRunningTime="2026-02-17 13:19:42.075014192 +0000 UTC m=+920.597743735" watchObservedRunningTime="2026-02-17 13:19:42.079652095 +0000 UTC m=+920.602381638" Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.081597 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987464f4-g967h" event={"ID":"5c339089-ba43-481b-aaa8-f8bf09fd106d","Type":"ContainerStarted","Data":"51047bb28d579f749e39c2ad89b1fff7ef649bc0313222fd42e450c8baf33cb2"} Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.082330 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-77987464f4-g967h" Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.092293 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-75hpg" event={"ID":"071c3b13-9304-48b6-bf22-d89dd62825a0","Type":"ContainerStarted","Data":"22e30f06b08a76566b6c903c8de75379e8b69c6f5011c603ead5878171f02fa3"} Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.093009 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-75hpg" Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.109024 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-xkvlp" podStartSLOduration=2.826239272 podStartE2EDuration="16.109010676s" podCreationTimestamp="2026-02-17 13:19:26 +0000 UTC" firstStartedPulling="2026-02-17 13:19:27.860613078 +0000 UTC m=+906.383342621" lastFinishedPulling="2026-02-17 13:19:41.143384472 +0000 UTC m=+919.666114025" observedRunningTime="2026-02-17 13:19:42.105488245 +0000 UTC m=+920.628217788" watchObservedRunningTime="2026-02-17 13:19:42.109010676 +0000 UTC m=+920.631740209" Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.109948 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-lpxnk" event={"ID":"6d1e254f-7d68-45c5-bda0-e3b33155ae56","Type":"ContainerStarted","Data":"b006ee1ed3dcecf114fcbac8fd67ebf3d4f7b053b6b5ca724ff33579bbc5d3c4"} Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.109985 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-lpxnk" Feb 17 13:19:42 crc kubenswrapper[4955]: E0217 13:19:42.112749 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:c6ad383f55f955902b074d1ee947a2233a5fcbf40698479ae693ce056c80dcc1\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-mkqrg" podUID="d6decb51-f3f9-4183-a20a-6aa4491dba61" Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.134985 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-vq5wq" podStartSLOduration=2.85478628 podStartE2EDuration="16.13496611s" podCreationTimestamp="2026-02-17 13:19:26 +0000 UTC" firstStartedPulling="2026-02-17 13:19:27.835852618 +0000 UTC m=+906.358582171" lastFinishedPulling="2026-02-17 13:19:41.116032458 +0000 UTC m=+919.638762001" observedRunningTime="2026-02-17 13:19:42.128425632 +0000 UTC m=+920.651155185" watchObservedRunningTime="2026-02-17 13:19:42.13496611 +0000 UTC m=+920.657695653" Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.181039 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-5fxtz" podStartSLOduration=2.829527736 podStartE2EDuration="16.18102004s" podCreationTimestamp="2026-02-17 13:19:26 +0000 UTC" firstStartedPulling="2026-02-17 13:19:27.832481441 +0000 UTC m=+906.355210984" lastFinishedPulling="2026-02-17 13:19:41.183973745 +0000 UTC m=+919.706703288" observedRunningTime="2026-02-17 13:19:42.157708302 +0000 UTC m=+920.680437845" watchObservedRunningTime="2026-02-17 13:19:42.18102004 +0000 UTC m=+920.703749583" Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.185240 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-k2x44" podStartSLOduration=2.661961464 podStartE2EDuration="16.18522119s" podCreationTimestamp="2026-02-17 13:19:26 +0000 UTC" firstStartedPulling="2026-02-17 13:19:27.592572686 +0000 UTC m=+906.115302229" lastFinishedPulling="2026-02-17 13:19:41.115832372 +0000 UTC m=+919.638561955" observedRunningTime="2026-02-17 13:19:42.179699602 +0000 UTC m=+920.702429145" watchObservedRunningTime="2026-02-17 13:19:42.18522119 +0000 UTC m=+920.707950733" Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.195349 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-4f8kb" podStartSLOduration=2.602288353 podStartE2EDuration="16.19532923s" podCreationTimestamp="2026-02-17 13:19:26 +0000 UTC" firstStartedPulling="2026-02-17 13:19:27.520017016 +0000 UTC m=+906.042746559" lastFinishedPulling="2026-02-17 13:19:41.113057893 +0000 UTC m=+919.635787436" observedRunningTime="2026-02-17 13:19:42.192272362 +0000 UTC m=+920.715001905" watchObservedRunningTime="2026-02-17 13:19:42.19532923 +0000 UTC m=+920.718058773" Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.219759 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cc84cdf7-92aa-4134-a784-d821c076cde3-cert\") pod \"infra-operator-controller-manager-79d975b745-gpgxj\" (UID: \"cc84cdf7-92aa-4134-a784-d821c076cde3\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-gpgxj" Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.226437 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cc84cdf7-92aa-4134-a784-d821c076cde3-cert\") pod \"infra-operator-controller-manager-79d975b745-gpgxj\" (UID: \"cc84cdf7-92aa-4134-a784-d821c076cde3\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-gpgxj" Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.249221 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-79d975b745-gpgxj" Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.264756 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-q68lr" podStartSLOduration=3.215780306 podStartE2EDuration="16.264733369s" podCreationTimestamp="2026-02-17 13:19:26 +0000 UTC" firstStartedPulling="2026-02-17 13:19:28.12342993 +0000 UTC m=+906.646159473" lastFinishedPulling="2026-02-17 13:19:41.172382993 +0000 UTC m=+919.695112536" observedRunningTime="2026-02-17 13:19:42.227662457 +0000 UTC m=+920.750392000" watchObservedRunningTime="2026-02-17 13:19:42.264733369 +0000 UTC m=+920.787462912" Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.291494 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-nzn4k" podStartSLOduration=2.7670948859999998 podStartE2EDuration="16.291477165s" podCreationTimestamp="2026-02-17 13:19:26 +0000 UTC" firstStartedPulling="2026-02-17 13:19:27.60981584 +0000 UTC m=+906.132545383" lastFinishedPulling="2026-02-17 13:19:41.134198119 +0000 UTC m=+919.656927662" observedRunningTime="2026-02-17 13:19:42.262307919 +0000 UTC m=+920.785037462" watchObservedRunningTime="2026-02-17 13:19:42.291477165 +0000 UTC m=+920.814206708" Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.318562 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-77987464f4-g967h" podStartSLOduration=2.718425151 podStartE2EDuration="16.318541161s" podCreationTimestamp="2026-02-17 13:19:26 +0000 UTC" firstStartedPulling="2026-02-17 13:19:27.515900308 +0000 UTC m=+906.038629841" lastFinishedPulling="2026-02-17 13:19:41.116016268 +0000 UTC m=+919.638745851" observedRunningTime="2026-02-17 13:19:42.315237326 +0000 UTC m=+920.837966869" watchObservedRunningTime="2026-02-17 13:19:42.318541161 +0000 UTC m=+920.841270704" Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.332151 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-75hpg" podStartSLOduration=2.793453983 podStartE2EDuration="16.332133201s" podCreationTimestamp="2026-02-17 13:19:26 +0000 UTC" firstStartedPulling="2026-02-17 13:19:27.577395511 +0000 UTC m=+906.100125054" lastFinishedPulling="2026-02-17 13:19:41.116074689 +0000 UTC m=+919.638804272" observedRunningTime="2026-02-17 13:19:42.329594378 +0000 UTC m=+920.852323921" watchObservedRunningTime="2026-02-17 13:19:42.332133201 +0000 UTC m=+920.854862744" Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.422544 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cd2ef1c9-23d1-4671-bc71-2c077afdfe08-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cxkmq8\" (UID: \"cd2ef1c9-23d1-4671-bc71-2c077afdfe08\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cxkmq8" Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.436771 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cd2ef1c9-23d1-4671-bc71-2c077afdfe08-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9cxkmq8\" (UID: \"cd2ef1c9-23d1-4671-bc71-2c077afdfe08\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cxkmq8" Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.443107 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cxkmq8" Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.830910 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c93c2a50-6396-4a32-9b88-dcdc341ec6c5-metrics-certs\") pod \"openstack-operator-controller-manager-5699594bc4-nrc2h\" (UID: \"c93c2a50-6396-4a32-9b88-dcdc341ec6c5\") " pod="openstack-operators/openstack-operator-controller-manager-5699594bc4-nrc2h" Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.831244 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c93c2a50-6396-4a32-9b88-dcdc341ec6c5-webhook-certs\") pod \"openstack-operator-controller-manager-5699594bc4-nrc2h\" (UID: \"c93c2a50-6396-4a32-9b88-dcdc341ec6c5\") " pod="openstack-operators/openstack-operator-controller-manager-5699594bc4-nrc2h" Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.834384 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c93c2a50-6396-4a32-9b88-dcdc341ec6c5-metrics-certs\") pod \"openstack-operator-controller-manager-5699594bc4-nrc2h\" (UID: \"c93c2a50-6396-4a32-9b88-dcdc341ec6c5\") " pod="openstack-operators/openstack-operator-controller-manager-5699594bc4-nrc2h" Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.834410 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c93c2a50-6396-4a32-9b88-dcdc341ec6c5-webhook-certs\") pod \"openstack-operator-controller-manager-5699594bc4-nrc2h\" (UID: \"c93c2a50-6396-4a32-9b88-dcdc341ec6c5\") " pod="openstack-operators/openstack-operator-controller-manager-5699594bc4-nrc2h" Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.842661 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5699594bc4-nrc2h" Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.900102 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-lpxnk" podStartSLOduration=3.154493859 podStartE2EDuration="16.900087398s" podCreationTimestamp="2026-02-17 13:19:26 +0000 UTC" firstStartedPulling="2026-02-17 13:19:27.370924433 +0000 UTC m=+905.893653976" lastFinishedPulling="2026-02-17 13:19:41.116517972 +0000 UTC m=+919.639247515" observedRunningTime="2026-02-17 13:19:42.385436808 +0000 UTC m=+920.908166361" watchObservedRunningTime="2026-02-17 13:19:42.900087398 +0000 UTC m=+921.422816941" Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.904264 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cxkmq8"] Feb 17 13:19:42 crc kubenswrapper[4955]: W0217 13:19:42.920210 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd2ef1c9_23d1_4671_bc71_2c077afdfe08.slice/crio-718a6039b2f907286b882b73620f6e70c69020704965273f767d43231444a77a WatchSource:0}: Error finding container 718a6039b2f907286b882b73620f6e70c69020704965273f767d43231444a77a: Status 404 returned error can't find the container with id 718a6039b2f907286b882b73620f6e70c69020704965273f767d43231444a77a Feb 17 13:19:42 crc kubenswrapper[4955]: I0217 13:19:42.932467 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-79d975b745-gpgxj"] Feb 17 13:19:42 crc kubenswrapper[4955]: W0217 13:19:42.933889 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcc84cdf7_92aa_4134_a784_d821c076cde3.slice/crio-a1195adf9cf22daa1e95afce9897c59450f9e8d46b9f1005d87c4df2202bd06e WatchSource:0}: Error finding container a1195adf9cf22daa1e95afce9897c59450f9e8d46b9f1005d87c4df2202bd06e: Status 404 returned error can't find the container with id a1195adf9cf22daa1e95afce9897c59450f9e8d46b9f1005d87c4df2202bd06e Feb 17 13:19:43 crc kubenswrapper[4955]: I0217 13:19:43.099597 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5699594bc4-nrc2h"] Feb 17 13:19:43 crc kubenswrapper[4955]: I0217 13:19:43.118998 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-79d975b745-gpgxj" event={"ID":"cc84cdf7-92aa-4134-a784-d821c076cde3","Type":"ContainerStarted","Data":"a1195adf9cf22daa1e95afce9897c59450f9e8d46b9f1005d87c4df2202bd06e"} Feb 17 13:19:43 crc kubenswrapper[4955]: I0217 13:19:43.120296 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5699594bc4-nrc2h" event={"ID":"c93c2a50-6396-4a32-9b88-dcdc341ec6c5","Type":"ContainerStarted","Data":"7bdc7420793ec56e11f768ae5f0528b50facea95def6fe048c58db25ea63e138"} Feb 17 13:19:43 crc kubenswrapper[4955]: I0217 13:19:43.121565 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cxkmq8" event={"ID":"cd2ef1c9-23d1-4671-bc71-2c077afdfe08","Type":"ContainerStarted","Data":"718a6039b2f907286b882b73620f6e70c69020704965273f767d43231444a77a"} Feb 17 13:19:44 crc kubenswrapper[4955]: I0217 13:19:44.129902 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5699594bc4-nrc2h" event={"ID":"c93c2a50-6396-4a32-9b88-dcdc341ec6c5","Type":"ContainerStarted","Data":"395076f0eea0ef65027c7542d7dfc1911e0b3190cdff6c9378b438f4b3bd5b0f"} Feb 17 13:19:44 crc kubenswrapper[4955]: I0217 13:19:44.130773 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-5699594bc4-nrc2h" Feb 17 13:19:44 crc kubenswrapper[4955]: I0217 13:19:44.163939 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-5699594bc4-nrc2h" podStartSLOduration=18.16391761 podStartE2EDuration="18.16391761s" podCreationTimestamp="2026-02-17 13:19:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:19:44.157335131 +0000 UTC m=+922.680064684" watchObservedRunningTime="2026-02-17 13:19:44.16391761 +0000 UTC m=+922.686647153" Feb 17 13:19:46 crc kubenswrapper[4955]: I0217 13:19:46.428359 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-ffkxx" Feb 17 13:19:46 crc kubenswrapper[4955]: I0217 13:19:46.454943 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-lpxnk" Feb 17 13:19:46 crc kubenswrapper[4955]: I0217 13:19:46.503253 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-77987464f4-g967h" Feb 17 13:19:46 crc kubenswrapper[4955]: I0217 13:19:46.556848 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-75hpg" Feb 17 13:19:46 crc kubenswrapper[4955]: I0217 13:19:46.608458 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-k2x44" Feb 17 13:19:46 crc kubenswrapper[4955]: I0217 13:19:46.663076 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-h2rm6" Feb 17 13:19:46 crc kubenswrapper[4955]: I0217 13:19:46.707497 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-4f8kb" Feb 17 13:19:46 crc kubenswrapper[4955]: I0217 13:19:46.730434 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-nzn4k" Feb 17 13:19:46 crc kubenswrapper[4955]: I0217 13:19:46.752697 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-xkvlp" Feb 17 13:19:46 crc kubenswrapper[4955]: I0217 13:19:46.815925 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-5fxtz" Feb 17 13:19:46 crc kubenswrapper[4955]: I0217 13:19:46.844589 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-wkdhh" Feb 17 13:19:46 crc kubenswrapper[4955]: I0217 13:19:46.873875 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-vq5wq" Feb 17 13:19:47 crc kubenswrapper[4955]: I0217 13:19:47.098465 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-q68lr" Feb 17 13:19:48 crc kubenswrapper[4955]: I0217 13:19:48.162526 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-v6jkn" event={"ID":"a1f4d718-9535-4290-94eb-93d6288d6047","Type":"ContainerStarted","Data":"262f72fa1d10b6740e8a14cb2c268adab434af428ef63637df201e1728ed9259"} Feb 17 13:19:48 crc kubenswrapper[4955]: I0217 13:19:48.163932 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-v6jkn" Feb 17 13:19:51 crc kubenswrapper[4955]: I0217 13:19:51.193275 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-fqrz9" event={"ID":"788f9fac-dd10-49c4-8a25-4578d69aeb2b","Type":"ContainerStarted","Data":"6dad119f7ea66998f628be77ff8ead2e319fd6bde90c307e53a869536fa433da"} Feb 17 13:19:51 crc kubenswrapper[4955]: I0217 13:19:51.194617 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-fqrz9" Feb 17 13:19:51 crc kubenswrapper[4955]: I0217 13:19:51.212547 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-fqrz9" podStartSLOduration=5.920823433 podStartE2EDuration="25.212525853s" podCreationTimestamp="2026-02-17 13:19:26 +0000 UTC" firstStartedPulling="2026-02-17 13:19:27.994450163 +0000 UTC m=+906.517179706" lastFinishedPulling="2026-02-17 13:19:47.286152573 +0000 UTC m=+925.808882126" observedRunningTime="2026-02-17 13:19:51.209624369 +0000 UTC m=+929.732353912" watchObservedRunningTime="2026-02-17 13:19:51.212525853 +0000 UTC m=+929.735255396" Feb 17 13:19:51 crc kubenswrapper[4955]: I0217 13:19:51.218045 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-v6jkn" podStartSLOduration=6.086134601 podStartE2EDuration="25.21802215s" podCreationTimestamp="2026-02-17 13:19:26 +0000 UTC" firstStartedPulling="2026-02-17 13:19:28.153645376 +0000 UTC m=+906.676374919" lastFinishedPulling="2026-02-17 13:19:47.285532925 +0000 UTC m=+925.808262468" observedRunningTime="2026-02-17 13:19:48.183020068 +0000 UTC m=+926.705749611" watchObservedRunningTime="2026-02-17 13:19:51.21802215 +0000 UTC m=+929.740751693" Feb 17 13:19:52 crc kubenswrapper[4955]: I0217 13:19:52.211635 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-79d975b745-gpgxj" event={"ID":"cc84cdf7-92aa-4134-a784-d821c076cde3","Type":"ContainerStarted","Data":"cd40550a8afce8c040736db2c46b5ce3f5343c41397f86294e21b5244995cce4"} Feb 17 13:19:52 crc kubenswrapper[4955]: I0217 13:19:52.212913 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-79d975b745-gpgxj" Feb 17 13:19:52 crc kubenswrapper[4955]: I0217 13:19:52.214978 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-7866795846-xqxjt" event={"ID":"90e44f80-9802-4134-8aa3-519bc93f4bd3","Type":"ContainerStarted","Data":"4f89ba8827e1ac2eca3a96a339dfd61408bd808717a0e9690f0f9d8cc2d22bba"} Feb 17 13:19:52 crc kubenswrapper[4955]: I0217 13:19:52.215241 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-7866795846-xqxjt" Feb 17 13:19:52 crc kubenswrapper[4955]: I0217 13:19:52.217641 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-2c8ww" event={"ID":"1e4fd408-9f09-4289-92ce-7c5ae18f235a","Type":"ContainerStarted","Data":"e6f832e97b80012b9359cde43021d2df93b9094b5d321b022ed9937c6c67ffba"} Feb 17 13:19:52 crc kubenswrapper[4955]: I0217 13:19:52.217942 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-2c8ww" Feb 17 13:19:52 crc kubenswrapper[4955]: I0217 13:19:52.219333 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-nmxwn" event={"ID":"afaac24a-0894-4001-9b50-d3d7c2de90bc","Type":"ContainerStarted","Data":"91d6de8af74a4b6cba40c8a9dd9df8d9efe01ccce01d901af8b8459474326678"} Feb 17 13:19:52 crc kubenswrapper[4955]: I0217 13:19:52.221045 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-68f46476f-5qz9n" event={"ID":"60b19c88-d91f-422e-9012-ad97daf661bb","Type":"ContainerStarted","Data":"121e94429652339add0d599d266ebaf4b2f6c3e1ec8f42e9c6dc6572e1f0a6cb"} Feb 17 13:19:52 crc kubenswrapper[4955]: I0217 13:19:52.221507 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-68f46476f-5qz9n" Feb 17 13:19:52 crc kubenswrapper[4955]: I0217 13:19:52.228798 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-79d975b745-gpgxj" podStartSLOduration=21.23057866 podStartE2EDuration="26.228761178s" podCreationTimestamp="2026-02-17 13:19:26 +0000 UTC" firstStartedPulling="2026-02-17 13:19:42.936579084 +0000 UTC m=+921.459308627" lastFinishedPulling="2026-02-17 13:19:47.934761602 +0000 UTC m=+926.457491145" observedRunningTime="2026-02-17 13:19:52.224906797 +0000 UTC m=+930.747636340" watchObservedRunningTime="2026-02-17 13:19:52.228761178 +0000 UTC m=+930.751490721" Feb 17 13:19:52 crc kubenswrapper[4955]: I0217 13:19:52.233806 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cxkmq8" event={"ID":"cd2ef1c9-23d1-4671-bc71-2c077afdfe08","Type":"ContainerStarted","Data":"ba0c002441bf582c295ed3da8f9d907f7097cbef98be268ac35f9f13ca2870cf"} Feb 17 13:19:52 crc kubenswrapper[4955]: I0217 13:19:52.257125 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cxkmq8" podStartSLOduration=21.240600637 podStartE2EDuration="26.25710643s" podCreationTimestamp="2026-02-17 13:19:26 +0000 UTC" firstStartedPulling="2026-02-17 13:19:42.919424983 +0000 UTC m=+921.442154526" lastFinishedPulling="2026-02-17 13:19:47.935930786 +0000 UTC m=+926.458660319" observedRunningTime="2026-02-17 13:19:52.253065414 +0000 UTC m=+930.775794967" watchObservedRunningTime="2026-02-17 13:19:52.25710643 +0000 UTC m=+930.779835973" Feb 17 13:19:52 crc kubenswrapper[4955]: I0217 13:19:52.303976 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-7866795846-xqxjt" podStartSLOduration=3.113959309 podStartE2EDuration="26.303956723s" podCreationTimestamp="2026-02-17 13:19:26 +0000 UTC" firstStartedPulling="2026-02-17 13:19:28.151618428 +0000 UTC m=+906.674347971" lastFinishedPulling="2026-02-17 13:19:51.341615842 +0000 UTC m=+929.864345385" observedRunningTime="2026-02-17 13:19:52.301604106 +0000 UTC m=+930.824333649" watchObservedRunningTime="2026-02-17 13:19:52.303956723 +0000 UTC m=+930.826686266" Feb 17 13:19:52 crc kubenswrapper[4955]: I0217 13:19:52.305508 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-nmxwn" podStartSLOduration=3.103234401 podStartE2EDuration="26.305498967s" podCreationTimestamp="2026-02-17 13:19:26 +0000 UTC" firstStartedPulling="2026-02-17 13:19:28.15551393 +0000 UTC m=+906.678243463" lastFinishedPulling="2026-02-17 13:19:51.357778486 +0000 UTC m=+929.880508029" observedRunningTime="2026-02-17 13:19:52.287492411 +0000 UTC m=+930.810221954" watchObservedRunningTime="2026-02-17 13:19:52.305498967 +0000 UTC m=+930.828228510" Feb 17 13:19:52 crc kubenswrapper[4955]: I0217 13:19:52.315936 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-2c8ww" podStartSLOduration=3.0158442069999998 podStartE2EDuration="26.315915576s" podCreationTimestamp="2026-02-17 13:19:26 +0000 UTC" firstStartedPulling="2026-02-17 13:19:27.991320854 +0000 UTC m=+906.514050397" lastFinishedPulling="2026-02-17 13:19:51.291392233 +0000 UTC m=+929.814121766" observedRunningTime="2026-02-17 13:19:52.313589079 +0000 UTC m=+930.836318622" watchObservedRunningTime="2026-02-17 13:19:52.315915576 +0000 UTC m=+930.838645129" Feb 17 13:19:52 crc kubenswrapper[4955]: I0217 13:19:52.332361 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-68f46476f-5qz9n" podStartSLOduration=5.755382382 podStartE2EDuration="26.332344867s" podCreationTimestamp="2026-02-17 13:19:26 +0000 UTC" firstStartedPulling="2026-02-17 13:19:27.993735733 +0000 UTC m=+906.516465276" lastFinishedPulling="2026-02-17 13:19:48.570698218 +0000 UTC m=+927.093427761" observedRunningTime="2026-02-17 13:19:52.326911601 +0000 UTC m=+930.849641154" watchObservedRunningTime="2026-02-17 13:19:52.332344867 +0000 UTC m=+930.855074410" Feb 17 13:19:52 crc kubenswrapper[4955]: I0217 13:19:52.443579 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cxkmq8" Feb 17 13:19:52 crc kubenswrapper[4955]: I0217 13:19:52.851410 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-5699594bc4-nrc2h" Feb 17 13:19:56 crc kubenswrapper[4955]: I0217 13:19:56.773593 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-2c8ww" Feb 17 13:19:56 crc kubenswrapper[4955]: I0217 13:19:56.906350 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-fqrz9" Feb 17 13:19:56 crc kubenswrapper[4955]: I0217 13:19:56.946754 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-68f46476f-5qz9n" Feb 17 13:19:57 crc kubenswrapper[4955]: I0217 13:19:57.101379 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-7866795846-xqxjt" Feb 17 13:19:57 crc kubenswrapper[4955]: I0217 13:19:57.141573 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-v6jkn" Feb 17 13:20:02 crc kubenswrapper[4955]: I0217 13:20:02.256836 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-79d975b745-gpgxj" Feb 17 13:20:02 crc kubenswrapper[4955]: I0217 13:20:02.448660 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9cxkmq8" Feb 17 13:20:04 crc kubenswrapper[4955]: I0217 13:20:04.575165 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 13:20:04 crc kubenswrapper[4955]: I0217 13:20:04.575252 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 13:20:04 crc kubenswrapper[4955]: I0217 13:20:04.575317 4955 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" Feb 17 13:20:04 crc kubenswrapper[4955]: I0217 13:20:04.576213 4955 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a72972c52870687e44697a4fc3675d78bf2389f390e43593c993f04221b0f584"} pod="openshift-machine-config-operator/machine-config-daemon-29qxq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 17 13:20:04 crc kubenswrapper[4955]: I0217 13:20:04.576322 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" containerID="cri-o://a72972c52870687e44697a4fc3675d78bf2389f390e43593c993f04221b0f584" gracePeriod=600 Feb 17 13:20:04 crc kubenswrapper[4955]: E0217 13:20:04.649740 4955 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5bf36887_2a2b_4193_abdb_a15beaf46300.slice/crio-a72972c52870687e44697a4fc3675d78bf2389f390e43593c993f04221b0f584.scope\": RecentStats: unable to find data in memory cache]" Feb 17 13:20:05 crc kubenswrapper[4955]: I0217 13:20:05.343805 4955 generic.go:334] "Generic (PLEG): container finished" podID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerID="a72972c52870687e44697a4fc3675d78bf2389f390e43593c993f04221b0f584" exitCode=0 Feb 17 13:20:05 crc kubenswrapper[4955]: I0217 13:20:05.343888 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" event={"ID":"5bf36887-2a2b-4193-abdb-a15beaf46300","Type":"ContainerDied","Data":"a72972c52870687e44697a4fc3675d78bf2389f390e43593c993f04221b0f584"} Feb 17 13:20:05 crc kubenswrapper[4955]: I0217 13:20:05.344082 4955 scope.go:117] "RemoveContainer" containerID="817a479207945cb4a0cb4caff307634a33fb718eab5f99bb1c4a37862d5010ac" Feb 17 13:20:06 crc kubenswrapper[4955]: I0217 13:20:06.352810 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-mkqrg" event={"ID":"d6decb51-f3f9-4183-a20a-6aa4491dba61","Type":"ContainerStarted","Data":"d363b970fefd1e6ecb7a1de0047b31c2d9764c16bbf0a3c9cda5bbf92689a27a"} Feb 17 13:20:06 crc kubenswrapper[4955]: I0217 13:20:06.353328 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-mkqrg" Feb 17 13:20:06 crc kubenswrapper[4955]: I0217 13:20:06.355494 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" event={"ID":"5bf36887-2a2b-4193-abdb-a15beaf46300","Type":"ContainerStarted","Data":"3481e1de1949e001dc1ec5b32fb42f82ef040a09d687f51932cebedf9288edf6"} Feb 17 13:20:06 crc kubenswrapper[4955]: I0217 13:20:06.371152 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-mkqrg" podStartSLOduration=2.143623859 podStartE2EDuration="40.371135378s" podCreationTimestamp="2026-02-17 13:19:26 +0000 UTC" firstStartedPulling="2026-02-17 13:19:27.643101134 +0000 UTC m=+906.165830677" lastFinishedPulling="2026-02-17 13:20:05.870612653 +0000 UTC m=+944.393342196" observedRunningTime="2026-02-17 13:20:06.370251292 +0000 UTC m=+944.892980845" watchObservedRunningTime="2026-02-17 13:20:06.371135378 +0000 UTC m=+944.893864921" Feb 17 13:20:16 crc kubenswrapper[4955]: I0217 13:20:16.679559 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-mkqrg" Feb 17 13:20:32 crc kubenswrapper[4955]: I0217 13:20:32.925925 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-6h44n"] Feb 17 13:20:32 crc kubenswrapper[4955]: I0217 13:20:32.927387 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-6h44n" Feb 17 13:20:32 crc kubenswrapper[4955]: I0217 13:20:32.934248 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-w8d72" Feb 17 13:20:32 crc kubenswrapper[4955]: I0217 13:20:32.934248 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Feb 17 13:20:32 crc kubenswrapper[4955]: I0217 13:20:32.934753 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Feb 17 13:20:32 crc kubenswrapper[4955]: I0217 13:20:32.937302 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Feb 17 13:20:32 crc kubenswrapper[4955]: I0217 13:20:32.996576 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-6h44n"] Feb 17 13:20:33 crc kubenswrapper[4955]: I0217 13:20:33.079132 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-nq5vv"] Feb 17 13:20:33 crc kubenswrapper[4955]: I0217 13:20:33.080117 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-nq5vv" Feb 17 13:20:33 crc kubenswrapper[4955]: I0217 13:20:33.081890 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Feb 17 13:20:33 crc kubenswrapper[4955]: I0217 13:20:33.093935 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-nq5vv"] Feb 17 13:20:33 crc kubenswrapper[4955]: I0217 13:20:33.100055 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06fb1f59-b127-4db8-9911-e87a5c9112f9-config\") pod \"dnsmasq-dns-675f4bcbfc-6h44n\" (UID: \"06fb1f59-b127-4db8-9911-e87a5c9112f9\") " pod="openstack/dnsmasq-dns-675f4bcbfc-6h44n" Feb 17 13:20:33 crc kubenswrapper[4955]: I0217 13:20:33.100143 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9mk2\" (UniqueName: \"kubernetes.io/projected/06fb1f59-b127-4db8-9911-e87a5c9112f9-kube-api-access-r9mk2\") pod \"dnsmasq-dns-675f4bcbfc-6h44n\" (UID: \"06fb1f59-b127-4db8-9911-e87a5c9112f9\") " pod="openstack/dnsmasq-dns-675f4bcbfc-6h44n" Feb 17 13:20:33 crc kubenswrapper[4955]: I0217 13:20:33.201605 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9mk2\" (UniqueName: \"kubernetes.io/projected/06fb1f59-b127-4db8-9911-e87a5c9112f9-kube-api-access-r9mk2\") pod \"dnsmasq-dns-675f4bcbfc-6h44n\" (UID: \"06fb1f59-b127-4db8-9911-e87a5c9112f9\") " pod="openstack/dnsmasq-dns-675f4bcbfc-6h44n" Feb 17 13:20:33 crc kubenswrapper[4955]: I0217 13:20:33.201658 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/76ecff2f-ae2a-4cdd-b983-b17e502bc9f3-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-nq5vv\" (UID: \"76ecff2f-ae2a-4cdd-b983-b17e502bc9f3\") " pod="openstack/dnsmasq-dns-78dd6ddcc-nq5vv" Feb 17 13:20:33 crc kubenswrapper[4955]: I0217 13:20:33.201694 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52cb7\" (UniqueName: \"kubernetes.io/projected/76ecff2f-ae2a-4cdd-b983-b17e502bc9f3-kube-api-access-52cb7\") pod \"dnsmasq-dns-78dd6ddcc-nq5vv\" (UID: \"76ecff2f-ae2a-4cdd-b983-b17e502bc9f3\") " pod="openstack/dnsmasq-dns-78dd6ddcc-nq5vv" Feb 17 13:20:33 crc kubenswrapper[4955]: I0217 13:20:33.201752 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06fb1f59-b127-4db8-9911-e87a5c9112f9-config\") pod \"dnsmasq-dns-675f4bcbfc-6h44n\" (UID: \"06fb1f59-b127-4db8-9911-e87a5c9112f9\") " pod="openstack/dnsmasq-dns-675f4bcbfc-6h44n" Feb 17 13:20:33 crc kubenswrapper[4955]: I0217 13:20:33.201770 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76ecff2f-ae2a-4cdd-b983-b17e502bc9f3-config\") pod \"dnsmasq-dns-78dd6ddcc-nq5vv\" (UID: \"76ecff2f-ae2a-4cdd-b983-b17e502bc9f3\") " pod="openstack/dnsmasq-dns-78dd6ddcc-nq5vv" Feb 17 13:20:33 crc kubenswrapper[4955]: I0217 13:20:33.202512 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06fb1f59-b127-4db8-9911-e87a5c9112f9-config\") pod \"dnsmasq-dns-675f4bcbfc-6h44n\" (UID: \"06fb1f59-b127-4db8-9911-e87a5c9112f9\") " pod="openstack/dnsmasq-dns-675f4bcbfc-6h44n" Feb 17 13:20:33 crc kubenswrapper[4955]: I0217 13:20:33.220723 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9mk2\" (UniqueName: \"kubernetes.io/projected/06fb1f59-b127-4db8-9911-e87a5c9112f9-kube-api-access-r9mk2\") pod \"dnsmasq-dns-675f4bcbfc-6h44n\" (UID: \"06fb1f59-b127-4db8-9911-e87a5c9112f9\") " pod="openstack/dnsmasq-dns-675f4bcbfc-6h44n" Feb 17 13:20:33 crc kubenswrapper[4955]: I0217 13:20:33.244532 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-6h44n" Feb 17 13:20:33 crc kubenswrapper[4955]: I0217 13:20:33.303183 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52cb7\" (UniqueName: \"kubernetes.io/projected/76ecff2f-ae2a-4cdd-b983-b17e502bc9f3-kube-api-access-52cb7\") pod \"dnsmasq-dns-78dd6ddcc-nq5vv\" (UID: \"76ecff2f-ae2a-4cdd-b983-b17e502bc9f3\") " pod="openstack/dnsmasq-dns-78dd6ddcc-nq5vv" Feb 17 13:20:33 crc kubenswrapper[4955]: I0217 13:20:33.303268 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76ecff2f-ae2a-4cdd-b983-b17e502bc9f3-config\") pod \"dnsmasq-dns-78dd6ddcc-nq5vv\" (UID: \"76ecff2f-ae2a-4cdd-b983-b17e502bc9f3\") " pod="openstack/dnsmasq-dns-78dd6ddcc-nq5vv" Feb 17 13:20:33 crc kubenswrapper[4955]: I0217 13:20:33.303383 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/76ecff2f-ae2a-4cdd-b983-b17e502bc9f3-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-nq5vv\" (UID: \"76ecff2f-ae2a-4cdd-b983-b17e502bc9f3\") " pod="openstack/dnsmasq-dns-78dd6ddcc-nq5vv" Feb 17 13:20:33 crc kubenswrapper[4955]: I0217 13:20:33.304095 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/76ecff2f-ae2a-4cdd-b983-b17e502bc9f3-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-nq5vv\" (UID: \"76ecff2f-ae2a-4cdd-b983-b17e502bc9f3\") " pod="openstack/dnsmasq-dns-78dd6ddcc-nq5vv" Feb 17 13:20:33 crc kubenswrapper[4955]: I0217 13:20:33.304354 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76ecff2f-ae2a-4cdd-b983-b17e502bc9f3-config\") pod \"dnsmasq-dns-78dd6ddcc-nq5vv\" (UID: \"76ecff2f-ae2a-4cdd-b983-b17e502bc9f3\") " pod="openstack/dnsmasq-dns-78dd6ddcc-nq5vv" Feb 17 13:20:33 crc kubenswrapper[4955]: I0217 13:20:33.320338 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52cb7\" (UniqueName: \"kubernetes.io/projected/76ecff2f-ae2a-4cdd-b983-b17e502bc9f3-kube-api-access-52cb7\") pod \"dnsmasq-dns-78dd6ddcc-nq5vv\" (UID: \"76ecff2f-ae2a-4cdd-b983-b17e502bc9f3\") " pod="openstack/dnsmasq-dns-78dd6ddcc-nq5vv" Feb 17 13:20:33 crc kubenswrapper[4955]: I0217 13:20:33.393973 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-nq5vv" Feb 17 13:20:33 crc kubenswrapper[4955]: I0217 13:20:33.674217 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-6h44n"] Feb 17 13:20:33 crc kubenswrapper[4955]: I0217 13:20:33.680937 4955 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 17 13:20:33 crc kubenswrapper[4955]: I0217 13:20:33.828967 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-nq5vv"] Feb 17 13:20:33 crc kubenswrapper[4955]: W0217 13:20:33.836373 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76ecff2f_ae2a_4cdd_b983_b17e502bc9f3.slice/crio-8338feaade4f83b37dfd2ae4e259bf2cecdcc7088e09c4ddb22f97f9252e69b9 WatchSource:0}: Error finding container 8338feaade4f83b37dfd2ae4e259bf2cecdcc7088e09c4ddb22f97f9252e69b9: Status 404 returned error can't find the container with id 8338feaade4f83b37dfd2ae4e259bf2cecdcc7088e09c4ddb22f97f9252e69b9 Feb 17 13:20:34 crc kubenswrapper[4955]: I0217 13:20:34.544421 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-6h44n" event={"ID":"06fb1f59-b127-4db8-9911-e87a5c9112f9","Type":"ContainerStarted","Data":"965a0633339dad494f040e0059ba1ba39f947fdff38524a3c2285a02945ffa16"} Feb 17 13:20:34 crc kubenswrapper[4955]: I0217 13:20:34.545583 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-nq5vv" event={"ID":"76ecff2f-ae2a-4cdd-b983-b17e502bc9f3","Type":"ContainerStarted","Data":"8338feaade4f83b37dfd2ae4e259bf2cecdcc7088e09c4ddb22f97f9252e69b9"} Feb 17 13:20:35 crc kubenswrapper[4955]: I0217 13:20:35.657401 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-6h44n"] Feb 17 13:20:35 crc kubenswrapper[4955]: I0217 13:20:35.685671 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-gm68k"] Feb 17 13:20:35 crc kubenswrapper[4955]: I0217 13:20:35.688167 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-gm68k" Feb 17 13:20:35 crc kubenswrapper[4955]: I0217 13:20:35.692627 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-gm68k"] Feb 17 13:20:35 crc kubenswrapper[4955]: I0217 13:20:35.862575 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ced654f-0b1a-4485-ab1d-c98d73e7b8d5-dns-svc\") pod \"dnsmasq-dns-666b6646f7-gm68k\" (UID: \"2ced654f-0b1a-4485-ab1d-c98d73e7b8d5\") " pod="openstack/dnsmasq-dns-666b6646f7-gm68k" Feb 17 13:20:35 crc kubenswrapper[4955]: I0217 13:20:35.862626 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxrwg\" (UniqueName: \"kubernetes.io/projected/2ced654f-0b1a-4485-ab1d-c98d73e7b8d5-kube-api-access-cxrwg\") pod \"dnsmasq-dns-666b6646f7-gm68k\" (UID: \"2ced654f-0b1a-4485-ab1d-c98d73e7b8d5\") " pod="openstack/dnsmasq-dns-666b6646f7-gm68k" Feb 17 13:20:35 crc kubenswrapper[4955]: I0217 13:20:35.862682 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ced654f-0b1a-4485-ab1d-c98d73e7b8d5-config\") pod \"dnsmasq-dns-666b6646f7-gm68k\" (UID: \"2ced654f-0b1a-4485-ab1d-c98d73e7b8d5\") " pod="openstack/dnsmasq-dns-666b6646f7-gm68k" Feb 17 13:20:35 crc kubenswrapper[4955]: I0217 13:20:35.964400 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ced654f-0b1a-4485-ab1d-c98d73e7b8d5-config\") pod \"dnsmasq-dns-666b6646f7-gm68k\" (UID: \"2ced654f-0b1a-4485-ab1d-c98d73e7b8d5\") " pod="openstack/dnsmasq-dns-666b6646f7-gm68k" Feb 17 13:20:35 crc kubenswrapper[4955]: I0217 13:20:35.964500 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ced654f-0b1a-4485-ab1d-c98d73e7b8d5-dns-svc\") pod \"dnsmasq-dns-666b6646f7-gm68k\" (UID: \"2ced654f-0b1a-4485-ab1d-c98d73e7b8d5\") " pod="openstack/dnsmasq-dns-666b6646f7-gm68k" Feb 17 13:20:35 crc kubenswrapper[4955]: I0217 13:20:35.964519 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxrwg\" (UniqueName: \"kubernetes.io/projected/2ced654f-0b1a-4485-ab1d-c98d73e7b8d5-kube-api-access-cxrwg\") pod \"dnsmasq-dns-666b6646f7-gm68k\" (UID: \"2ced654f-0b1a-4485-ab1d-c98d73e7b8d5\") " pod="openstack/dnsmasq-dns-666b6646f7-gm68k" Feb 17 13:20:35 crc kubenswrapper[4955]: I0217 13:20:35.966203 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ced654f-0b1a-4485-ab1d-c98d73e7b8d5-config\") pod \"dnsmasq-dns-666b6646f7-gm68k\" (UID: \"2ced654f-0b1a-4485-ab1d-c98d73e7b8d5\") " pod="openstack/dnsmasq-dns-666b6646f7-gm68k" Feb 17 13:20:35 crc kubenswrapper[4955]: I0217 13:20:35.969589 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ced654f-0b1a-4485-ab1d-c98d73e7b8d5-dns-svc\") pod \"dnsmasq-dns-666b6646f7-gm68k\" (UID: \"2ced654f-0b1a-4485-ab1d-c98d73e7b8d5\") " pod="openstack/dnsmasq-dns-666b6646f7-gm68k" Feb 17 13:20:35 crc kubenswrapper[4955]: I0217 13:20:35.996171 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxrwg\" (UniqueName: \"kubernetes.io/projected/2ced654f-0b1a-4485-ab1d-c98d73e7b8d5-kube-api-access-cxrwg\") pod \"dnsmasq-dns-666b6646f7-gm68k\" (UID: \"2ced654f-0b1a-4485-ab1d-c98d73e7b8d5\") " pod="openstack/dnsmasq-dns-666b6646f7-gm68k" Feb 17 13:20:36 crc kubenswrapper[4955]: I0217 13:20:36.011654 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-gm68k" Feb 17 13:20:36 crc kubenswrapper[4955]: I0217 13:20:36.055978 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-nq5vv"] Feb 17 13:20:36 crc kubenswrapper[4955]: I0217 13:20:36.086573 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-f5xt5"] Feb 17 13:20:36 crc kubenswrapper[4955]: I0217 13:20:36.087970 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-f5xt5" Feb 17 13:20:36 crc kubenswrapper[4955]: I0217 13:20:36.100465 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-f5xt5"] Feb 17 13:20:36 crc kubenswrapper[4955]: I0217 13:20:36.268483 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f7eef812-b3ad-4b76-885d-c8d474cd43e0-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-f5xt5\" (UID: \"f7eef812-b3ad-4b76-885d-c8d474cd43e0\") " pod="openstack/dnsmasq-dns-57d769cc4f-f5xt5" Feb 17 13:20:36 crc kubenswrapper[4955]: I0217 13:20:36.268528 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7eef812-b3ad-4b76-885d-c8d474cd43e0-config\") pod \"dnsmasq-dns-57d769cc4f-f5xt5\" (UID: \"f7eef812-b3ad-4b76-885d-c8d474cd43e0\") " pod="openstack/dnsmasq-dns-57d769cc4f-f5xt5" Feb 17 13:20:36 crc kubenswrapper[4955]: I0217 13:20:36.268701 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjddn\" (UniqueName: \"kubernetes.io/projected/f7eef812-b3ad-4b76-885d-c8d474cd43e0-kube-api-access-pjddn\") pod \"dnsmasq-dns-57d769cc4f-f5xt5\" (UID: \"f7eef812-b3ad-4b76-885d-c8d474cd43e0\") " pod="openstack/dnsmasq-dns-57d769cc4f-f5xt5" Feb 17 13:20:36 crc kubenswrapper[4955]: I0217 13:20:36.369884 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjddn\" (UniqueName: \"kubernetes.io/projected/f7eef812-b3ad-4b76-885d-c8d474cd43e0-kube-api-access-pjddn\") pod \"dnsmasq-dns-57d769cc4f-f5xt5\" (UID: \"f7eef812-b3ad-4b76-885d-c8d474cd43e0\") " pod="openstack/dnsmasq-dns-57d769cc4f-f5xt5" Feb 17 13:20:36 crc kubenswrapper[4955]: I0217 13:20:36.369937 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f7eef812-b3ad-4b76-885d-c8d474cd43e0-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-f5xt5\" (UID: \"f7eef812-b3ad-4b76-885d-c8d474cd43e0\") " pod="openstack/dnsmasq-dns-57d769cc4f-f5xt5" Feb 17 13:20:36 crc kubenswrapper[4955]: I0217 13:20:36.369966 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7eef812-b3ad-4b76-885d-c8d474cd43e0-config\") pod \"dnsmasq-dns-57d769cc4f-f5xt5\" (UID: \"f7eef812-b3ad-4b76-885d-c8d474cd43e0\") " pod="openstack/dnsmasq-dns-57d769cc4f-f5xt5" Feb 17 13:20:36 crc kubenswrapper[4955]: I0217 13:20:36.371555 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7eef812-b3ad-4b76-885d-c8d474cd43e0-config\") pod \"dnsmasq-dns-57d769cc4f-f5xt5\" (UID: \"f7eef812-b3ad-4b76-885d-c8d474cd43e0\") " pod="openstack/dnsmasq-dns-57d769cc4f-f5xt5" Feb 17 13:20:36 crc kubenswrapper[4955]: I0217 13:20:36.371598 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f7eef812-b3ad-4b76-885d-c8d474cd43e0-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-f5xt5\" (UID: \"f7eef812-b3ad-4b76-885d-c8d474cd43e0\") " pod="openstack/dnsmasq-dns-57d769cc4f-f5xt5" Feb 17 13:20:36 crc kubenswrapper[4955]: I0217 13:20:36.407715 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjddn\" (UniqueName: \"kubernetes.io/projected/f7eef812-b3ad-4b76-885d-c8d474cd43e0-kube-api-access-pjddn\") pod \"dnsmasq-dns-57d769cc4f-f5xt5\" (UID: \"f7eef812-b3ad-4b76-885d-c8d474cd43e0\") " pod="openstack/dnsmasq-dns-57d769cc4f-f5xt5" Feb 17 13:20:36 crc kubenswrapper[4955]: I0217 13:20:36.447090 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-f5xt5" Feb 17 13:20:36 crc kubenswrapper[4955]: I0217 13:20:36.559800 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-gm68k"] Feb 17 13:20:36 crc kubenswrapper[4955]: W0217 13:20:36.570320 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ced654f_0b1a_4485_ab1d_c98d73e7b8d5.slice/crio-f9bbc2b6db45f390b0ab2c836561c5073bc73ea9c9db27367b0a0c873ec9c300 WatchSource:0}: Error finding container f9bbc2b6db45f390b0ab2c836561c5073bc73ea9c9db27367b0a0c873ec9c300: Status 404 returned error can't find the container with id f9bbc2b6db45f390b0ab2c836561c5073bc73ea9c9db27367b0a0c873ec9c300 Feb 17 13:20:36 crc kubenswrapper[4955]: I0217 13:20:36.850921 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Feb 17 13:20:36 crc kubenswrapper[4955]: I0217 13:20:36.851976 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 17 13:20:36 crc kubenswrapper[4955]: I0217 13:20:36.857422 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Feb 17 13:20:36 crc kubenswrapper[4955]: I0217 13:20:36.857567 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Feb 17 13:20:36 crc kubenswrapper[4955]: I0217 13:20:36.857953 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-nt4nx" Feb 17 13:20:36 crc kubenswrapper[4955]: I0217 13:20:36.858328 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Feb 17 13:20:36 crc kubenswrapper[4955]: I0217 13:20:36.858467 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Feb 17 13:20:36 crc kubenswrapper[4955]: I0217 13:20:36.858535 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Feb 17 13:20:36 crc kubenswrapper[4955]: I0217 13:20:36.858602 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Feb 17 13:20:36 crc kubenswrapper[4955]: I0217 13:20:36.869312 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 17 13:20:36 crc kubenswrapper[4955]: I0217 13:20:36.895103 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-f5xt5"] Feb 17 13:20:36 crc kubenswrapper[4955]: I0217 13:20:36.979748 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/34552e31-68c3-4156-addf-37ea6fdf0b6c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " pod="openstack/rabbitmq-server-0" Feb 17 13:20:36 crc kubenswrapper[4955]: I0217 13:20:36.979826 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/34552e31-68c3-4156-addf-37ea6fdf0b6c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " pod="openstack/rabbitmq-server-0" Feb 17 13:20:36 crc kubenswrapper[4955]: I0217 13:20:36.979856 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/34552e31-68c3-4156-addf-37ea6fdf0b6c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " pod="openstack/rabbitmq-server-0" Feb 17 13:20:36 crc kubenswrapper[4955]: I0217 13:20:36.979894 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " pod="openstack/rabbitmq-server-0" Feb 17 13:20:36 crc kubenswrapper[4955]: I0217 13:20:36.980092 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/34552e31-68c3-4156-addf-37ea6fdf0b6c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " pod="openstack/rabbitmq-server-0" Feb 17 13:20:36 crc kubenswrapper[4955]: I0217 13:20:36.980139 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbxhr\" (UniqueName: \"kubernetes.io/projected/34552e31-68c3-4156-addf-37ea6fdf0b6c-kube-api-access-nbxhr\") pod \"rabbitmq-server-0\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " pod="openstack/rabbitmq-server-0" Feb 17 13:20:36 crc kubenswrapper[4955]: I0217 13:20:36.980202 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/34552e31-68c3-4156-addf-37ea6fdf0b6c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " pod="openstack/rabbitmq-server-0" Feb 17 13:20:36 crc kubenswrapper[4955]: I0217 13:20:36.980278 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/34552e31-68c3-4156-addf-37ea6fdf0b6c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " pod="openstack/rabbitmq-server-0" Feb 17 13:20:36 crc kubenswrapper[4955]: I0217 13:20:36.980419 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/34552e31-68c3-4156-addf-37ea6fdf0b6c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " pod="openstack/rabbitmq-server-0" Feb 17 13:20:36 crc kubenswrapper[4955]: I0217 13:20:36.980473 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/34552e31-68c3-4156-addf-37ea6fdf0b6c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " pod="openstack/rabbitmq-server-0" Feb 17 13:20:36 crc kubenswrapper[4955]: I0217 13:20:36.980508 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/34552e31-68c3-4156-addf-37ea6fdf0b6c-config-data\") pod \"rabbitmq-server-0\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " pod="openstack/rabbitmq-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.087671 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/34552e31-68c3-4156-addf-37ea6fdf0b6c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " pod="openstack/rabbitmq-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.087753 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbxhr\" (UniqueName: \"kubernetes.io/projected/34552e31-68c3-4156-addf-37ea6fdf0b6c-kube-api-access-nbxhr\") pod \"rabbitmq-server-0\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " pod="openstack/rabbitmq-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.087797 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/34552e31-68c3-4156-addf-37ea6fdf0b6c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " pod="openstack/rabbitmq-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.087816 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/34552e31-68c3-4156-addf-37ea6fdf0b6c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " pod="openstack/rabbitmq-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.087869 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/34552e31-68c3-4156-addf-37ea6fdf0b6c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " pod="openstack/rabbitmq-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.087904 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/34552e31-68c3-4156-addf-37ea6fdf0b6c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " pod="openstack/rabbitmq-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.087924 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/34552e31-68c3-4156-addf-37ea6fdf0b6c-config-data\") pod \"rabbitmq-server-0\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " pod="openstack/rabbitmq-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.087942 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/34552e31-68c3-4156-addf-37ea6fdf0b6c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " pod="openstack/rabbitmq-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.087970 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/34552e31-68c3-4156-addf-37ea6fdf0b6c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " pod="openstack/rabbitmq-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.087993 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/34552e31-68c3-4156-addf-37ea6fdf0b6c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " pod="openstack/rabbitmq-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.088032 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " pod="openstack/rabbitmq-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.088256 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/34552e31-68c3-4156-addf-37ea6fdf0b6c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " pod="openstack/rabbitmq-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.088442 4955 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.089006 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/34552e31-68c3-4156-addf-37ea6fdf0b6c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " pod="openstack/rabbitmq-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.091611 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/34552e31-68c3-4156-addf-37ea6fdf0b6c-config-data\") pod \"rabbitmq-server-0\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " pod="openstack/rabbitmq-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.092139 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/34552e31-68c3-4156-addf-37ea6fdf0b6c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " pod="openstack/rabbitmq-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.093024 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/34552e31-68c3-4156-addf-37ea6fdf0b6c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " pod="openstack/rabbitmq-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.093473 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/34552e31-68c3-4156-addf-37ea6fdf0b6c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " pod="openstack/rabbitmq-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.094468 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/34552e31-68c3-4156-addf-37ea6fdf0b6c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " pod="openstack/rabbitmq-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.098942 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/34552e31-68c3-4156-addf-37ea6fdf0b6c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " pod="openstack/rabbitmq-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.105242 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbxhr\" (UniqueName: \"kubernetes.io/projected/34552e31-68c3-4156-addf-37ea6fdf0b6c-kube-api-access-nbxhr\") pod \"rabbitmq-server-0\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " pod="openstack/rabbitmq-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.106477 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/34552e31-68c3-4156-addf-37ea6fdf0b6c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " pod="openstack/rabbitmq-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.112877 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " pod="openstack/rabbitmq-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.166392 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.167795 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.170526 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.170632 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.170701 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.170806 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.170812 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.170911 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.170743 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-86vw7" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.175748 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.188423 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.290845 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sdlw\" (UniqueName: \"kubernetes.io/projected/c241677c-9f24-4cc0-8d15-e166174e4cd0-kube-api-access-5sdlw\") pod \"rabbitmq-cell1-server-0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.290907 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c241677c-9f24-4cc0-8d15-e166174e4cd0-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.290932 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c241677c-9f24-4cc0-8d15-e166174e4cd0-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.290951 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c241677c-9f24-4cc0-8d15-e166174e4cd0-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.290972 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.290991 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c241677c-9f24-4cc0-8d15-e166174e4cd0-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.291092 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c241677c-9f24-4cc0-8d15-e166174e4cd0-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.291112 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c241677c-9f24-4cc0-8d15-e166174e4cd0-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.291132 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c241677c-9f24-4cc0-8d15-e166174e4cd0-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.291175 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c241677c-9f24-4cc0-8d15-e166174e4cd0-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.291192 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c241677c-9f24-4cc0-8d15-e166174e4cd0-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.391947 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c241677c-9f24-4cc0-8d15-e166174e4cd0-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.392012 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c241677c-9f24-4cc0-8d15-e166174e4cd0-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.392037 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c241677c-9f24-4cc0-8d15-e166174e4cd0-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.392070 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.392091 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c241677c-9f24-4cc0-8d15-e166174e4cd0-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.392141 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c241677c-9f24-4cc0-8d15-e166174e4cd0-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.392157 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c241677c-9f24-4cc0-8d15-e166174e4cd0-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.392172 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c241677c-9f24-4cc0-8d15-e166174e4cd0-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.392247 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c241677c-9f24-4cc0-8d15-e166174e4cd0-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.392266 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c241677c-9f24-4cc0-8d15-e166174e4cd0-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.392321 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sdlw\" (UniqueName: \"kubernetes.io/projected/c241677c-9f24-4cc0-8d15-e166174e4cd0-kube-api-access-5sdlw\") pod \"rabbitmq-cell1-server-0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.393038 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c241677c-9f24-4cc0-8d15-e166174e4cd0-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.393406 4955 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.393590 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c241677c-9f24-4cc0-8d15-e166174e4cd0-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.393718 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c241677c-9f24-4cc0-8d15-e166174e4cd0-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.394735 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c241677c-9f24-4cc0-8d15-e166174e4cd0-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.395527 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c241677c-9f24-4cc0-8d15-e166174e4cd0-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.396103 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c241677c-9f24-4cc0-8d15-e166174e4cd0-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.398293 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c241677c-9f24-4cc0-8d15-e166174e4cd0-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.407852 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c241677c-9f24-4cc0-8d15-e166174e4cd0-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.410931 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c241677c-9f24-4cc0-8d15-e166174e4cd0-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.415084 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sdlw\" (UniqueName: \"kubernetes.io/projected/c241677c-9f24-4cc0-8d15-e166174e4cd0-kube-api-access-5sdlw\") pod \"rabbitmq-cell1-server-0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.417943 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.498678 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:20:37 crc kubenswrapper[4955]: I0217 13:20:37.583757 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-gm68k" event={"ID":"2ced654f-0b1a-4485-ab1d-c98d73e7b8d5","Type":"ContainerStarted","Data":"f9bbc2b6db45f390b0ab2c836561c5073bc73ea9c9db27367b0a0c873ec9c300"} Feb 17 13:20:38 crc kubenswrapper[4955]: I0217 13:20:38.335157 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Feb 17 13:20:38 crc kubenswrapper[4955]: I0217 13:20:38.336265 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Feb 17 13:20:38 crc kubenswrapper[4955]: I0217 13:20:38.339310 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-rn6bq" Feb 17 13:20:38 crc kubenswrapper[4955]: I0217 13:20:38.339595 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Feb 17 13:20:38 crc kubenswrapper[4955]: I0217 13:20:38.339975 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Feb 17 13:20:38 crc kubenswrapper[4955]: I0217 13:20:38.341389 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Feb 17 13:20:38 crc kubenswrapper[4955]: I0217 13:20:38.358405 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Feb 17 13:20:38 crc kubenswrapper[4955]: I0217 13:20:38.371741 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Feb 17 13:20:38 crc kubenswrapper[4955]: I0217 13:20:38.506443 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/def1308a-11ee-4465-89f0-e78ffc4b8b51-config-data-default\") pod \"openstack-galera-0\" (UID: \"def1308a-11ee-4465-89f0-e78ffc4b8b51\") " pod="openstack/openstack-galera-0" Feb 17 13:20:38 crc kubenswrapper[4955]: I0217 13:20:38.506847 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/def1308a-11ee-4465-89f0-e78ffc4b8b51-config-data-generated\") pod \"openstack-galera-0\" (UID: \"def1308a-11ee-4465-89f0-e78ffc4b8b51\") " pod="openstack/openstack-galera-0" Feb 17 13:20:38 crc kubenswrapper[4955]: I0217 13:20:38.506907 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/def1308a-11ee-4465-89f0-e78ffc4b8b51-operator-scripts\") pod \"openstack-galera-0\" (UID: \"def1308a-11ee-4465-89f0-e78ffc4b8b51\") " pod="openstack/openstack-galera-0" Feb 17 13:20:38 crc kubenswrapper[4955]: I0217 13:20:38.507051 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/def1308a-11ee-4465-89f0-e78ffc4b8b51-kolla-config\") pod \"openstack-galera-0\" (UID: \"def1308a-11ee-4465-89f0-e78ffc4b8b51\") " pod="openstack/openstack-galera-0" Feb 17 13:20:38 crc kubenswrapper[4955]: I0217 13:20:38.507200 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/def1308a-11ee-4465-89f0-e78ffc4b8b51-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"def1308a-11ee-4465-89f0-e78ffc4b8b51\") " pod="openstack/openstack-galera-0" Feb 17 13:20:38 crc kubenswrapper[4955]: I0217 13:20:38.507232 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkv47\" (UniqueName: \"kubernetes.io/projected/def1308a-11ee-4465-89f0-e78ffc4b8b51-kube-api-access-rkv47\") pod \"openstack-galera-0\" (UID: \"def1308a-11ee-4465-89f0-e78ffc4b8b51\") " pod="openstack/openstack-galera-0" Feb 17 13:20:38 crc kubenswrapper[4955]: I0217 13:20:38.507306 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/def1308a-11ee-4465-89f0-e78ffc4b8b51-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"def1308a-11ee-4465-89f0-e78ffc4b8b51\") " pod="openstack/openstack-galera-0" Feb 17 13:20:38 crc kubenswrapper[4955]: I0217 13:20:38.507339 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"def1308a-11ee-4465-89f0-e78ffc4b8b51\") " pod="openstack/openstack-galera-0" Feb 17 13:20:38 crc kubenswrapper[4955]: I0217 13:20:38.609145 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/def1308a-11ee-4465-89f0-e78ffc4b8b51-operator-scripts\") pod \"openstack-galera-0\" (UID: \"def1308a-11ee-4465-89f0-e78ffc4b8b51\") " pod="openstack/openstack-galera-0" Feb 17 13:20:38 crc kubenswrapper[4955]: I0217 13:20:38.609210 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/def1308a-11ee-4465-89f0-e78ffc4b8b51-kolla-config\") pod \"openstack-galera-0\" (UID: \"def1308a-11ee-4465-89f0-e78ffc4b8b51\") " pod="openstack/openstack-galera-0" Feb 17 13:20:38 crc kubenswrapper[4955]: I0217 13:20:38.609260 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/def1308a-11ee-4465-89f0-e78ffc4b8b51-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"def1308a-11ee-4465-89f0-e78ffc4b8b51\") " pod="openstack/openstack-galera-0" Feb 17 13:20:38 crc kubenswrapper[4955]: I0217 13:20:38.609275 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkv47\" (UniqueName: \"kubernetes.io/projected/def1308a-11ee-4465-89f0-e78ffc4b8b51-kube-api-access-rkv47\") pod \"openstack-galera-0\" (UID: \"def1308a-11ee-4465-89f0-e78ffc4b8b51\") " pod="openstack/openstack-galera-0" Feb 17 13:20:38 crc kubenswrapper[4955]: I0217 13:20:38.609304 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/def1308a-11ee-4465-89f0-e78ffc4b8b51-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"def1308a-11ee-4465-89f0-e78ffc4b8b51\") " pod="openstack/openstack-galera-0" Feb 17 13:20:38 crc kubenswrapper[4955]: I0217 13:20:38.609321 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"def1308a-11ee-4465-89f0-e78ffc4b8b51\") " pod="openstack/openstack-galera-0" Feb 17 13:20:38 crc kubenswrapper[4955]: I0217 13:20:38.609350 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/def1308a-11ee-4465-89f0-e78ffc4b8b51-config-data-default\") pod \"openstack-galera-0\" (UID: \"def1308a-11ee-4465-89f0-e78ffc4b8b51\") " pod="openstack/openstack-galera-0" Feb 17 13:20:38 crc kubenswrapper[4955]: I0217 13:20:38.609372 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/def1308a-11ee-4465-89f0-e78ffc4b8b51-config-data-generated\") pod \"openstack-galera-0\" (UID: \"def1308a-11ee-4465-89f0-e78ffc4b8b51\") " pod="openstack/openstack-galera-0" Feb 17 13:20:38 crc kubenswrapper[4955]: I0217 13:20:38.609813 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/def1308a-11ee-4465-89f0-e78ffc4b8b51-config-data-generated\") pod \"openstack-galera-0\" (UID: \"def1308a-11ee-4465-89f0-e78ffc4b8b51\") " pod="openstack/openstack-galera-0" Feb 17 13:20:38 crc kubenswrapper[4955]: I0217 13:20:38.610213 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/def1308a-11ee-4465-89f0-e78ffc4b8b51-kolla-config\") pod \"openstack-galera-0\" (UID: \"def1308a-11ee-4465-89f0-e78ffc4b8b51\") " pod="openstack/openstack-galera-0" Feb 17 13:20:38 crc kubenswrapper[4955]: I0217 13:20:38.610377 4955 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"def1308a-11ee-4465-89f0-e78ffc4b8b51\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/openstack-galera-0" Feb 17 13:20:38 crc kubenswrapper[4955]: I0217 13:20:38.610450 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/def1308a-11ee-4465-89f0-e78ffc4b8b51-config-data-default\") pod \"openstack-galera-0\" (UID: \"def1308a-11ee-4465-89f0-e78ffc4b8b51\") " pod="openstack/openstack-galera-0" Feb 17 13:20:38 crc kubenswrapper[4955]: I0217 13:20:38.611162 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/def1308a-11ee-4465-89f0-e78ffc4b8b51-operator-scripts\") pod \"openstack-galera-0\" (UID: \"def1308a-11ee-4465-89f0-e78ffc4b8b51\") " pod="openstack/openstack-galera-0" Feb 17 13:20:38 crc kubenswrapper[4955]: I0217 13:20:38.618699 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/def1308a-11ee-4465-89f0-e78ffc4b8b51-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"def1308a-11ee-4465-89f0-e78ffc4b8b51\") " pod="openstack/openstack-galera-0" Feb 17 13:20:38 crc kubenswrapper[4955]: I0217 13:20:38.618991 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/def1308a-11ee-4465-89f0-e78ffc4b8b51-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"def1308a-11ee-4465-89f0-e78ffc4b8b51\") " pod="openstack/openstack-galera-0" Feb 17 13:20:38 crc kubenswrapper[4955]: I0217 13:20:38.630637 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkv47\" (UniqueName: \"kubernetes.io/projected/def1308a-11ee-4465-89f0-e78ffc4b8b51-kube-api-access-rkv47\") pod \"openstack-galera-0\" (UID: \"def1308a-11ee-4465-89f0-e78ffc4b8b51\") " pod="openstack/openstack-galera-0" Feb 17 13:20:38 crc kubenswrapper[4955]: I0217 13:20:38.666029 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"def1308a-11ee-4465-89f0-e78ffc4b8b51\") " pod="openstack/openstack-galera-0" Feb 17 13:20:38 crc kubenswrapper[4955]: I0217 13:20:38.962293 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Feb 17 13:20:39 crc kubenswrapper[4955]: I0217 13:20:39.598809 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-f5xt5" event={"ID":"f7eef812-b3ad-4b76-885d-c8d474cd43e0","Type":"ContainerStarted","Data":"b51bff11b1d3b806cf7981806717ab805b59ed5bcf2ccf515625950b46687882"} Feb 17 13:20:39 crc kubenswrapper[4955]: I0217 13:20:39.711387 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 17 13:20:39 crc kubenswrapper[4955]: I0217 13:20:39.712648 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Feb 17 13:20:39 crc kubenswrapper[4955]: I0217 13:20:39.714766 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-6wjwz" Feb 17 13:20:39 crc kubenswrapper[4955]: I0217 13:20:39.717488 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Feb 17 13:20:39 crc kubenswrapper[4955]: I0217 13:20:39.717488 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Feb 17 13:20:39 crc kubenswrapper[4955]: I0217 13:20:39.722957 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Feb 17 13:20:39 crc kubenswrapper[4955]: I0217 13:20:39.723650 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 17 13:20:39 crc kubenswrapper[4955]: I0217 13:20:39.831022 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c93012c-0538-49a7-b414-6b01bb6c6dce-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"1c93012c-0538-49a7-b414-6b01bb6c6dce\") " pod="openstack/openstack-cell1-galera-0" Feb 17 13:20:39 crc kubenswrapper[4955]: I0217 13:20:39.831412 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1c93012c-0538-49a7-b414-6b01bb6c6dce-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"1c93012c-0538-49a7-b414-6b01bb6c6dce\") " pod="openstack/openstack-cell1-galera-0" Feb 17 13:20:39 crc kubenswrapper[4955]: I0217 13:20:39.831433 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1c93012c-0538-49a7-b414-6b01bb6c6dce-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"1c93012c-0538-49a7-b414-6b01bb6c6dce\") " pod="openstack/openstack-cell1-galera-0" Feb 17 13:20:39 crc kubenswrapper[4955]: I0217 13:20:39.831455 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1c93012c-0538-49a7-b414-6b01bb6c6dce\") " pod="openstack/openstack-cell1-galera-0" Feb 17 13:20:39 crc kubenswrapper[4955]: I0217 13:20:39.831492 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c93012c-0538-49a7-b414-6b01bb6c6dce-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"1c93012c-0538-49a7-b414-6b01bb6c6dce\") " pod="openstack/openstack-cell1-galera-0" Feb 17 13:20:39 crc kubenswrapper[4955]: I0217 13:20:39.831508 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1c93012c-0538-49a7-b414-6b01bb6c6dce-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"1c93012c-0538-49a7-b414-6b01bb6c6dce\") " pod="openstack/openstack-cell1-galera-0" Feb 17 13:20:39 crc kubenswrapper[4955]: I0217 13:20:39.832042 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1c93012c-0538-49a7-b414-6b01bb6c6dce-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"1c93012c-0538-49a7-b414-6b01bb6c6dce\") " pod="openstack/openstack-cell1-galera-0" Feb 17 13:20:39 crc kubenswrapper[4955]: I0217 13:20:39.832114 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9j65\" (UniqueName: \"kubernetes.io/projected/1c93012c-0538-49a7-b414-6b01bb6c6dce-kube-api-access-s9j65\") pod \"openstack-cell1-galera-0\" (UID: \"1c93012c-0538-49a7-b414-6b01bb6c6dce\") " pod="openstack/openstack-cell1-galera-0" Feb 17 13:20:39 crc kubenswrapper[4955]: I0217 13:20:39.934548 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1c93012c-0538-49a7-b414-6b01bb6c6dce-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"1c93012c-0538-49a7-b414-6b01bb6c6dce\") " pod="openstack/openstack-cell1-galera-0" Feb 17 13:20:39 crc kubenswrapper[4955]: I0217 13:20:39.934595 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c93012c-0538-49a7-b414-6b01bb6c6dce-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"1c93012c-0538-49a7-b414-6b01bb6c6dce\") " pod="openstack/openstack-cell1-galera-0" Feb 17 13:20:39 crc kubenswrapper[4955]: I0217 13:20:39.934697 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1c93012c-0538-49a7-b414-6b01bb6c6dce-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"1c93012c-0538-49a7-b414-6b01bb6c6dce\") " pod="openstack/openstack-cell1-galera-0" Feb 17 13:20:39 crc kubenswrapper[4955]: I0217 13:20:39.934716 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9j65\" (UniqueName: \"kubernetes.io/projected/1c93012c-0538-49a7-b414-6b01bb6c6dce-kube-api-access-s9j65\") pod \"openstack-cell1-galera-0\" (UID: \"1c93012c-0538-49a7-b414-6b01bb6c6dce\") " pod="openstack/openstack-cell1-galera-0" Feb 17 13:20:39 crc kubenswrapper[4955]: I0217 13:20:39.934874 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c93012c-0538-49a7-b414-6b01bb6c6dce-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"1c93012c-0538-49a7-b414-6b01bb6c6dce\") " pod="openstack/openstack-cell1-galera-0" Feb 17 13:20:39 crc kubenswrapper[4955]: I0217 13:20:39.934914 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1c93012c-0538-49a7-b414-6b01bb6c6dce-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"1c93012c-0538-49a7-b414-6b01bb6c6dce\") " pod="openstack/openstack-cell1-galera-0" Feb 17 13:20:39 crc kubenswrapper[4955]: I0217 13:20:39.934957 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1c93012c-0538-49a7-b414-6b01bb6c6dce-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"1c93012c-0538-49a7-b414-6b01bb6c6dce\") " pod="openstack/openstack-cell1-galera-0" Feb 17 13:20:39 crc kubenswrapper[4955]: I0217 13:20:39.934977 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1c93012c-0538-49a7-b414-6b01bb6c6dce\") " pod="openstack/openstack-cell1-galera-0" Feb 17 13:20:39 crc kubenswrapper[4955]: I0217 13:20:39.935932 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1c93012c-0538-49a7-b414-6b01bb6c6dce-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"1c93012c-0538-49a7-b414-6b01bb6c6dce\") " pod="openstack/openstack-cell1-galera-0" Feb 17 13:20:39 crc kubenswrapper[4955]: I0217 13:20:39.936452 4955 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1c93012c-0538-49a7-b414-6b01bb6c6dce\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-cell1-galera-0" Feb 17 13:20:39 crc kubenswrapper[4955]: I0217 13:20:39.937013 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c93012c-0538-49a7-b414-6b01bb6c6dce-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"1c93012c-0538-49a7-b414-6b01bb6c6dce\") " pod="openstack/openstack-cell1-galera-0" Feb 17 13:20:39 crc kubenswrapper[4955]: I0217 13:20:39.937447 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1c93012c-0538-49a7-b414-6b01bb6c6dce-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"1c93012c-0538-49a7-b414-6b01bb6c6dce\") " pod="openstack/openstack-cell1-galera-0" Feb 17 13:20:39 crc kubenswrapper[4955]: I0217 13:20:39.938025 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1c93012c-0538-49a7-b414-6b01bb6c6dce-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"1c93012c-0538-49a7-b414-6b01bb6c6dce\") " pod="openstack/openstack-cell1-galera-0" Feb 17 13:20:39 crc kubenswrapper[4955]: I0217 13:20:39.939252 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c93012c-0538-49a7-b414-6b01bb6c6dce-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"1c93012c-0538-49a7-b414-6b01bb6c6dce\") " pod="openstack/openstack-cell1-galera-0" Feb 17 13:20:39 crc kubenswrapper[4955]: I0217 13:20:39.940367 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1c93012c-0538-49a7-b414-6b01bb6c6dce-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"1c93012c-0538-49a7-b414-6b01bb6c6dce\") " pod="openstack/openstack-cell1-galera-0" Feb 17 13:20:39 crc kubenswrapper[4955]: I0217 13:20:39.963117 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1c93012c-0538-49a7-b414-6b01bb6c6dce\") " pod="openstack/openstack-cell1-galera-0" Feb 17 13:20:39 crc kubenswrapper[4955]: I0217 13:20:39.968601 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9j65\" (UniqueName: \"kubernetes.io/projected/1c93012c-0538-49a7-b414-6b01bb6c6dce-kube-api-access-s9j65\") pod \"openstack-cell1-galera-0\" (UID: \"1c93012c-0538-49a7-b414-6b01bb6c6dce\") " pod="openstack/openstack-cell1-galera-0" Feb 17 13:20:40 crc kubenswrapper[4955]: I0217 13:20:40.048823 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Feb 17 13:20:40 crc kubenswrapper[4955]: I0217 13:20:40.079624 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Feb 17 13:20:40 crc kubenswrapper[4955]: I0217 13:20:40.080934 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Feb 17 13:20:40 crc kubenswrapper[4955]: I0217 13:20:40.083554 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Feb 17 13:20:40 crc kubenswrapper[4955]: I0217 13:20:40.083958 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Feb 17 13:20:40 crc kubenswrapper[4955]: I0217 13:20:40.084834 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-hz9c5" Feb 17 13:20:40 crc kubenswrapper[4955]: I0217 13:20:40.094519 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Feb 17 13:20:40 crc kubenswrapper[4955]: I0217 13:20:40.241066 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccj5m\" (UniqueName: \"kubernetes.io/projected/7dab7b98-7d1a-4940-a4c3-8bbbd57fe121-kube-api-access-ccj5m\") pod \"memcached-0\" (UID: \"7dab7b98-7d1a-4940-a4c3-8bbbd57fe121\") " pod="openstack/memcached-0" Feb 17 13:20:40 crc kubenswrapper[4955]: I0217 13:20:40.241152 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7dab7b98-7d1a-4940-a4c3-8bbbd57fe121-config-data\") pod \"memcached-0\" (UID: \"7dab7b98-7d1a-4940-a4c3-8bbbd57fe121\") " pod="openstack/memcached-0" Feb 17 13:20:40 crc kubenswrapper[4955]: I0217 13:20:40.241202 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/7dab7b98-7d1a-4940-a4c3-8bbbd57fe121-memcached-tls-certs\") pod \"memcached-0\" (UID: \"7dab7b98-7d1a-4940-a4c3-8bbbd57fe121\") " pod="openstack/memcached-0" Feb 17 13:20:40 crc kubenswrapper[4955]: I0217 13:20:40.241253 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dab7b98-7d1a-4940-a4c3-8bbbd57fe121-combined-ca-bundle\") pod \"memcached-0\" (UID: \"7dab7b98-7d1a-4940-a4c3-8bbbd57fe121\") " pod="openstack/memcached-0" Feb 17 13:20:40 crc kubenswrapper[4955]: I0217 13:20:40.241314 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7dab7b98-7d1a-4940-a4c3-8bbbd57fe121-kolla-config\") pod \"memcached-0\" (UID: \"7dab7b98-7d1a-4940-a4c3-8bbbd57fe121\") " pod="openstack/memcached-0" Feb 17 13:20:40 crc kubenswrapper[4955]: I0217 13:20:40.342886 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccj5m\" (UniqueName: \"kubernetes.io/projected/7dab7b98-7d1a-4940-a4c3-8bbbd57fe121-kube-api-access-ccj5m\") pod \"memcached-0\" (UID: \"7dab7b98-7d1a-4940-a4c3-8bbbd57fe121\") " pod="openstack/memcached-0" Feb 17 13:20:40 crc kubenswrapper[4955]: I0217 13:20:40.342946 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7dab7b98-7d1a-4940-a4c3-8bbbd57fe121-config-data\") pod \"memcached-0\" (UID: \"7dab7b98-7d1a-4940-a4c3-8bbbd57fe121\") " pod="openstack/memcached-0" Feb 17 13:20:40 crc kubenswrapper[4955]: I0217 13:20:40.342999 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/7dab7b98-7d1a-4940-a4c3-8bbbd57fe121-memcached-tls-certs\") pod \"memcached-0\" (UID: \"7dab7b98-7d1a-4940-a4c3-8bbbd57fe121\") " pod="openstack/memcached-0" Feb 17 13:20:40 crc kubenswrapper[4955]: I0217 13:20:40.343032 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dab7b98-7d1a-4940-a4c3-8bbbd57fe121-combined-ca-bundle\") pod \"memcached-0\" (UID: \"7dab7b98-7d1a-4940-a4c3-8bbbd57fe121\") " pod="openstack/memcached-0" Feb 17 13:20:40 crc kubenswrapper[4955]: I0217 13:20:40.343084 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7dab7b98-7d1a-4940-a4c3-8bbbd57fe121-kolla-config\") pod \"memcached-0\" (UID: \"7dab7b98-7d1a-4940-a4c3-8bbbd57fe121\") " pod="openstack/memcached-0" Feb 17 13:20:40 crc kubenswrapper[4955]: I0217 13:20:40.344080 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7dab7b98-7d1a-4940-a4c3-8bbbd57fe121-config-data\") pod \"memcached-0\" (UID: \"7dab7b98-7d1a-4940-a4c3-8bbbd57fe121\") " pod="openstack/memcached-0" Feb 17 13:20:40 crc kubenswrapper[4955]: I0217 13:20:40.344186 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7dab7b98-7d1a-4940-a4c3-8bbbd57fe121-kolla-config\") pod \"memcached-0\" (UID: \"7dab7b98-7d1a-4940-a4c3-8bbbd57fe121\") " pod="openstack/memcached-0" Feb 17 13:20:40 crc kubenswrapper[4955]: I0217 13:20:40.346922 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/7dab7b98-7d1a-4940-a4c3-8bbbd57fe121-memcached-tls-certs\") pod \"memcached-0\" (UID: \"7dab7b98-7d1a-4940-a4c3-8bbbd57fe121\") " pod="openstack/memcached-0" Feb 17 13:20:40 crc kubenswrapper[4955]: I0217 13:20:40.352563 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dab7b98-7d1a-4940-a4c3-8bbbd57fe121-combined-ca-bundle\") pod \"memcached-0\" (UID: \"7dab7b98-7d1a-4940-a4c3-8bbbd57fe121\") " pod="openstack/memcached-0" Feb 17 13:20:40 crc kubenswrapper[4955]: I0217 13:20:40.366608 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccj5m\" (UniqueName: \"kubernetes.io/projected/7dab7b98-7d1a-4940-a4c3-8bbbd57fe121-kube-api-access-ccj5m\") pod \"memcached-0\" (UID: \"7dab7b98-7d1a-4940-a4c3-8bbbd57fe121\") " pod="openstack/memcached-0" Feb 17 13:20:40 crc kubenswrapper[4955]: I0217 13:20:40.400328 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Feb 17 13:20:42 crc kubenswrapper[4955]: I0217 13:20:42.404957 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Feb 17 13:20:42 crc kubenswrapper[4955]: I0217 13:20:42.410490 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 17 13:20:42 crc kubenswrapper[4955]: I0217 13:20:42.414256 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 17 13:20:42 crc kubenswrapper[4955]: I0217 13:20:42.416915 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-6nd78" Feb 17 13:20:42 crc kubenswrapper[4955]: I0217 13:20:42.472685 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lnjg\" (UniqueName: \"kubernetes.io/projected/c631bd4b-26ac-44e3-87aa-8878c42e6ce2-kube-api-access-9lnjg\") pod \"kube-state-metrics-0\" (UID: \"c631bd4b-26ac-44e3-87aa-8878c42e6ce2\") " pod="openstack/kube-state-metrics-0" Feb 17 13:20:42 crc kubenswrapper[4955]: I0217 13:20:42.574692 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lnjg\" (UniqueName: \"kubernetes.io/projected/c631bd4b-26ac-44e3-87aa-8878c42e6ce2-kube-api-access-9lnjg\") pod \"kube-state-metrics-0\" (UID: \"c631bd4b-26ac-44e3-87aa-8878c42e6ce2\") " pod="openstack/kube-state-metrics-0" Feb 17 13:20:42 crc kubenswrapper[4955]: I0217 13:20:42.597899 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lnjg\" (UniqueName: \"kubernetes.io/projected/c631bd4b-26ac-44e3-87aa-8878c42e6ce2-kube-api-access-9lnjg\") pod \"kube-state-metrics-0\" (UID: \"c631bd4b-26ac-44e3-87aa-8878c42e6ce2\") " pod="openstack/kube-state-metrics-0" Feb 17 13:20:42 crc kubenswrapper[4955]: I0217 13:20:42.741216 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.164003 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-f8cjz"] Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.166357 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-f8cjz" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.169026 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.169356 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.169487 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-lwst8" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.180616 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-f8cjz"] Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.224985 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-m27b4"] Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.230950 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-m27b4" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.237607 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-m27b4"] Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.330208 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/452bb98a-815b-4154-84c5-6b31338fca29-ovn-controller-tls-certs\") pod \"ovn-controller-f8cjz\" (UID: \"452bb98a-815b-4154-84c5-6b31338fca29\") " pod="openstack/ovn-controller-f8cjz" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.330320 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/452bb98a-815b-4154-84c5-6b31338fca29-combined-ca-bundle\") pod \"ovn-controller-f8cjz\" (UID: \"452bb98a-815b-4154-84c5-6b31338fca29\") " pod="openstack/ovn-controller-f8cjz" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.330394 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9534cd01-b88e-48a4-a2ca-dd968e045dbc-scripts\") pod \"ovn-controller-ovs-m27b4\" (UID: \"9534cd01-b88e-48a4-a2ca-dd968e045dbc\") " pod="openstack/ovn-controller-ovs-m27b4" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.330502 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/452bb98a-815b-4154-84c5-6b31338fca29-var-log-ovn\") pod \"ovn-controller-f8cjz\" (UID: \"452bb98a-815b-4154-84c5-6b31338fca29\") " pod="openstack/ovn-controller-f8cjz" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.330563 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/9534cd01-b88e-48a4-a2ca-dd968e045dbc-var-lib\") pod \"ovn-controller-ovs-m27b4\" (UID: \"9534cd01-b88e-48a4-a2ca-dd968e045dbc\") " pod="openstack/ovn-controller-ovs-m27b4" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.330626 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/452bb98a-815b-4154-84c5-6b31338fca29-var-run\") pod \"ovn-controller-f8cjz\" (UID: \"452bb98a-815b-4154-84c5-6b31338fca29\") " pod="openstack/ovn-controller-f8cjz" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.330668 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/452bb98a-815b-4154-84c5-6b31338fca29-scripts\") pod \"ovn-controller-f8cjz\" (UID: \"452bb98a-815b-4154-84c5-6b31338fca29\") " pod="openstack/ovn-controller-f8cjz" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.330799 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9534cd01-b88e-48a4-a2ca-dd968e045dbc-var-run\") pod \"ovn-controller-ovs-m27b4\" (UID: \"9534cd01-b88e-48a4-a2ca-dd968e045dbc\") " pod="openstack/ovn-controller-ovs-m27b4" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.330893 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/9534cd01-b88e-48a4-a2ca-dd968e045dbc-etc-ovs\") pod \"ovn-controller-ovs-m27b4\" (UID: \"9534cd01-b88e-48a4-a2ca-dd968e045dbc\") " pod="openstack/ovn-controller-ovs-m27b4" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.330981 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/9534cd01-b88e-48a4-a2ca-dd968e045dbc-var-log\") pod \"ovn-controller-ovs-m27b4\" (UID: \"9534cd01-b88e-48a4-a2ca-dd968e045dbc\") " pod="openstack/ovn-controller-ovs-m27b4" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.331117 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzg99\" (UniqueName: \"kubernetes.io/projected/452bb98a-815b-4154-84c5-6b31338fca29-kube-api-access-tzg99\") pod \"ovn-controller-f8cjz\" (UID: \"452bb98a-815b-4154-84c5-6b31338fca29\") " pod="openstack/ovn-controller-f8cjz" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.331154 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gktxx\" (UniqueName: \"kubernetes.io/projected/9534cd01-b88e-48a4-a2ca-dd968e045dbc-kube-api-access-gktxx\") pod \"ovn-controller-ovs-m27b4\" (UID: \"9534cd01-b88e-48a4-a2ca-dd968e045dbc\") " pod="openstack/ovn-controller-ovs-m27b4" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.332016 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/452bb98a-815b-4154-84c5-6b31338fca29-var-run-ovn\") pod \"ovn-controller-f8cjz\" (UID: \"452bb98a-815b-4154-84c5-6b31338fca29\") " pod="openstack/ovn-controller-f8cjz" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.433141 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzg99\" (UniqueName: \"kubernetes.io/projected/452bb98a-815b-4154-84c5-6b31338fca29-kube-api-access-tzg99\") pod \"ovn-controller-f8cjz\" (UID: \"452bb98a-815b-4154-84c5-6b31338fca29\") " pod="openstack/ovn-controller-f8cjz" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.433188 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/452bb98a-815b-4154-84c5-6b31338fca29-var-run-ovn\") pod \"ovn-controller-f8cjz\" (UID: \"452bb98a-815b-4154-84c5-6b31338fca29\") " pod="openstack/ovn-controller-f8cjz" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.433207 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gktxx\" (UniqueName: \"kubernetes.io/projected/9534cd01-b88e-48a4-a2ca-dd968e045dbc-kube-api-access-gktxx\") pod \"ovn-controller-ovs-m27b4\" (UID: \"9534cd01-b88e-48a4-a2ca-dd968e045dbc\") " pod="openstack/ovn-controller-ovs-m27b4" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.433245 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/452bb98a-815b-4154-84c5-6b31338fca29-ovn-controller-tls-certs\") pod \"ovn-controller-f8cjz\" (UID: \"452bb98a-815b-4154-84c5-6b31338fca29\") " pod="openstack/ovn-controller-f8cjz" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.433267 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/452bb98a-815b-4154-84c5-6b31338fca29-combined-ca-bundle\") pod \"ovn-controller-f8cjz\" (UID: \"452bb98a-815b-4154-84c5-6b31338fca29\") " pod="openstack/ovn-controller-f8cjz" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.433288 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9534cd01-b88e-48a4-a2ca-dd968e045dbc-scripts\") pod \"ovn-controller-ovs-m27b4\" (UID: \"9534cd01-b88e-48a4-a2ca-dd968e045dbc\") " pod="openstack/ovn-controller-ovs-m27b4" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.433325 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/452bb98a-815b-4154-84c5-6b31338fca29-var-log-ovn\") pod \"ovn-controller-f8cjz\" (UID: \"452bb98a-815b-4154-84c5-6b31338fca29\") " pod="openstack/ovn-controller-f8cjz" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.433341 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/9534cd01-b88e-48a4-a2ca-dd968e045dbc-var-lib\") pod \"ovn-controller-ovs-m27b4\" (UID: \"9534cd01-b88e-48a4-a2ca-dd968e045dbc\") " pod="openstack/ovn-controller-ovs-m27b4" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.433361 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/452bb98a-815b-4154-84c5-6b31338fca29-var-run\") pod \"ovn-controller-f8cjz\" (UID: \"452bb98a-815b-4154-84c5-6b31338fca29\") " pod="openstack/ovn-controller-f8cjz" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.433382 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/452bb98a-815b-4154-84c5-6b31338fca29-scripts\") pod \"ovn-controller-f8cjz\" (UID: \"452bb98a-815b-4154-84c5-6b31338fca29\") " pod="openstack/ovn-controller-f8cjz" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.433409 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9534cd01-b88e-48a4-a2ca-dd968e045dbc-var-run\") pod \"ovn-controller-ovs-m27b4\" (UID: \"9534cd01-b88e-48a4-a2ca-dd968e045dbc\") " pod="openstack/ovn-controller-ovs-m27b4" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.433433 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/9534cd01-b88e-48a4-a2ca-dd968e045dbc-etc-ovs\") pod \"ovn-controller-ovs-m27b4\" (UID: \"9534cd01-b88e-48a4-a2ca-dd968e045dbc\") " pod="openstack/ovn-controller-ovs-m27b4" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.433457 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/9534cd01-b88e-48a4-a2ca-dd968e045dbc-var-log\") pod \"ovn-controller-ovs-m27b4\" (UID: \"9534cd01-b88e-48a4-a2ca-dd968e045dbc\") " pod="openstack/ovn-controller-ovs-m27b4" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.434114 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/452bb98a-815b-4154-84c5-6b31338fca29-var-log-ovn\") pod \"ovn-controller-f8cjz\" (UID: \"452bb98a-815b-4154-84c5-6b31338fca29\") " pod="openstack/ovn-controller-f8cjz" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.434123 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/452bb98a-815b-4154-84c5-6b31338fca29-var-run\") pod \"ovn-controller-f8cjz\" (UID: \"452bb98a-815b-4154-84c5-6b31338fca29\") " pod="openstack/ovn-controller-f8cjz" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.434135 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9534cd01-b88e-48a4-a2ca-dd968e045dbc-var-run\") pod \"ovn-controller-ovs-m27b4\" (UID: \"9534cd01-b88e-48a4-a2ca-dd968e045dbc\") " pod="openstack/ovn-controller-ovs-m27b4" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.434281 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/9534cd01-b88e-48a4-a2ca-dd968e045dbc-var-lib\") pod \"ovn-controller-ovs-m27b4\" (UID: \"9534cd01-b88e-48a4-a2ca-dd968e045dbc\") " pod="openstack/ovn-controller-ovs-m27b4" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.434355 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/9534cd01-b88e-48a4-a2ca-dd968e045dbc-etc-ovs\") pod \"ovn-controller-ovs-m27b4\" (UID: \"9534cd01-b88e-48a4-a2ca-dd968e045dbc\") " pod="openstack/ovn-controller-ovs-m27b4" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.434617 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/452bb98a-815b-4154-84c5-6b31338fca29-var-run-ovn\") pod \"ovn-controller-f8cjz\" (UID: \"452bb98a-815b-4154-84c5-6b31338fca29\") " pod="openstack/ovn-controller-f8cjz" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.436730 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/452bb98a-815b-4154-84c5-6b31338fca29-scripts\") pod \"ovn-controller-f8cjz\" (UID: \"452bb98a-815b-4154-84c5-6b31338fca29\") " pod="openstack/ovn-controller-f8cjz" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.436842 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/9534cd01-b88e-48a4-a2ca-dd968e045dbc-var-log\") pod \"ovn-controller-ovs-m27b4\" (UID: \"9534cd01-b88e-48a4-a2ca-dd968e045dbc\") " pod="openstack/ovn-controller-ovs-m27b4" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.437489 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9534cd01-b88e-48a4-a2ca-dd968e045dbc-scripts\") pod \"ovn-controller-ovs-m27b4\" (UID: \"9534cd01-b88e-48a4-a2ca-dd968e045dbc\") " pod="openstack/ovn-controller-ovs-m27b4" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.451303 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/452bb98a-815b-4154-84c5-6b31338fca29-ovn-controller-tls-certs\") pod \"ovn-controller-f8cjz\" (UID: \"452bb98a-815b-4154-84c5-6b31338fca29\") " pod="openstack/ovn-controller-f8cjz" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.451504 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/452bb98a-815b-4154-84c5-6b31338fca29-combined-ca-bundle\") pod \"ovn-controller-f8cjz\" (UID: \"452bb98a-815b-4154-84c5-6b31338fca29\") " pod="openstack/ovn-controller-f8cjz" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.453252 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gktxx\" (UniqueName: \"kubernetes.io/projected/9534cd01-b88e-48a4-a2ca-dd968e045dbc-kube-api-access-gktxx\") pod \"ovn-controller-ovs-m27b4\" (UID: \"9534cd01-b88e-48a4-a2ca-dd968e045dbc\") " pod="openstack/ovn-controller-ovs-m27b4" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.453896 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzg99\" (UniqueName: \"kubernetes.io/projected/452bb98a-815b-4154-84c5-6b31338fca29-kube-api-access-tzg99\") pod \"ovn-controller-f8cjz\" (UID: \"452bb98a-815b-4154-84c5-6b31338fca29\") " pod="openstack/ovn-controller-f8cjz" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.485769 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-f8cjz" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.559407 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-m27b4" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.718520 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.720413 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.723700 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.724018 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.724172 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.724410 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.724548 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-59667" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.742087 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.845275 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3384848-891d-46ff-a869-61d657304b74-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b3384848-891d-46ff-a869-61d657304b74\") " pod="openstack/ovsdbserver-nb-0" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.845494 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"b3384848-891d-46ff-a869-61d657304b74\") " pod="openstack/ovsdbserver-nb-0" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.845536 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3384848-891d-46ff-a869-61d657304b74-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b3384848-891d-46ff-a869-61d657304b74\") " pod="openstack/ovsdbserver-nb-0" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.845559 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3384848-891d-46ff-a869-61d657304b74-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"b3384848-891d-46ff-a869-61d657304b74\") " pod="openstack/ovsdbserver-nb-0" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.845577 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3384848-891d-46ff-a869-61d657304b74-config\") pod \"ovsdbserver-nb-0\" (UID: \"b3384848-891d-46ff-a869-61d657304b74\") " pod="openstack/ovsdbserver-nb-0" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.845607 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b3384848-891d-46ff-a869-61d657304b74-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"b3384848-891d-46ff-a869-61d657304b74\") " pod="openstack/ovsdbserver-nb-0" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.845773 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9k4xz\" (UniqueName: \"kubernetes.io/projected/b3384848-891d-46ff-a869-61d657304b74-kube-api-access-9k4xz\") pod \"ovsdbserver-nb-0\" (UID: \"b3384848-891d-46ff-a869-61d657304b74\") " pod="openstack/ovsdbserver-nb-0" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.845972 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b3384848-891d-46ff-a869-61d657304b74-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"b3384848-891d-46ff-a869-61d657304b74\") " pod="openstack/ovsdbserver-nb-0" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.947550 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9k4xz\" (UniqueName: \"kubernetes.io/projected/b3384848-891d-46ff-a869-61d657304b74-kube-api-access-9k4xz\") pod \"ovsdbserver-nb-0\" (UID: \"b3384848-891d-46ff-a869-61d657304b74\") " pod="openstack/ovsdbserver-nb-0" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.947648 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b3384848-891d-46ff-a869-61d657304b74-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"b3384848-891d-46ff-a869-61d657304b74\") " pod="openstack/ovsdbserver-nb-0" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.947698 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3384848-891d-46ff-a869-61d657304b74-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b3384848-891d-46ff-a869-61d657304b74\") " pod="openstack/ovsdbserver-nb-0" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.947750 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"b3384848-891d-46ff-a869-61d657304b74\") " pod="openstack/ovsdbserver-nb-0" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.947803 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3384848-891d-46ff-a869-61d657304b74-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b3384848-891d-46ff-a869-61d657304b74\") " pod="openstack/ovsdbserver-nb-0" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.947827 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3384848-891d-46ff-a869-61d657304b74-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"b3384848-891d-46ff-a869-61d657304b74\") " pod="openstack/ovsdbserver-nb-0" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.947849 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3384848-891d-46ff-a869-61d657304b74-config\") pod \"ovsdbserver-nb-0\" (UID: \"b3384848-891d-46ff-a869-61d657304b74\") " pod="openstack/ovsdbserver-nb-0" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.947887 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b3384848-891d-46ff-a869-61d657304b74-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"b3384848-891d-46ff-a869-61d657304b74\") " pod="openstack/ovsdbserver-nb-0" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.948323 4955 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"b3384848-891d-46ff-a869-61d657304b74\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/ovsdbserver-nb-0" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.948926 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b3384848-891d-46ff-a869-61d657304b74-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"b3384848-891d-46ff-a869-61d657304b74\") " pod="openstack/ovsdbserver-nb-0" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.949300 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b3384848-891d-46ff-a869-61d657304b74-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"b3384848-891d-46ff-a869-61d657304b74\") " pod="openstack/ovsdbserver-nb-0" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.949645 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3384848-891d-46ff-a869-61d657304b74-config\") pod \"ovsdbserver-nb-0\" (UID: \"b3384848-891d-46ff-a869-61d657304b74\") " pod="openstack/ovsdbserver-nb-0" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.952881 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3384848-891d-46ff-a869-61d657304b74-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"b3384848-891d-46ff-a869-61d657304b74\") " pod="openstack/ovsdbserver-nb-0" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.953291 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3384848-891d-46ff-a869-61d657304b74-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b3384848-891d-46ff-a869-61d657304b74\") " pod="openstack/ovsdbserver-nb-0" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.966796 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3384848-891d-46ff-a869-61d657304b74-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b3384848-891d-46ff-a869-61d657304b74\") " pod="openstack/ovsdbserver-nb-0" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.972537 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9k4xz\" (UniqueName: \"kubernetes.io/projected/b3384848-891d-46ff-a869-61d657304b74-kube-api-access-9k4xz\") pod \"ovsdbserver-nb-0\" (UID: \"b3384848-891d-46ff-a869-61d657304b74\") " pod="openstack/ovsdbserver-nb-0" Feb 17 13:20:45 crc kubenswrapper[4955]: I0217 13:20:45.974317 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"b3384848-891d-46ff-a869-61d657304b74\") " pod="openstack/ovsdbserver-nb-0" Feb 17 13:20:46 crc kubenswrapper[4955]: I0217 13:20:46.047108 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Feb 17 13:20:49 crc kubenswrapper[4955]: I0217 13:20:49.267496 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 17 13:20:49 crc kubenswrapper[4955]: I0217 13:20:49.269044 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Feb 17 13:20:49 crc kubenswrapper[4955]: I0217 13:20:49.271479 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Feb 17 13:20:49 crc kubenswrapper[4955]: I0217 13:20:49.272042 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Feb 17 13:20:49 crc kubenswrapper[4955]: I0217 13:20:49.272372 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Feb 17 13:20:49 crc kubenswrapper[4955]: I0217 13:20:49.276880 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-4jsc2" Feb 17 13:20:49 crc kubenswrapper[4955]: I0217 13:20:49.287353 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 17 13:20:49 crc kubenswrapper[4955]: I0217 13:20:49.407594 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa892f09-0663-4fb6-b3ad-a15c300b5b65-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"aa892f09-0663-4fb6-b3ad-a15c300b5b65\") " pod="openstack/ovsdbserver-sb-0" Feb 17 13:20:49 crc kubenswrapper[4955]: I0217 13:20:49.407635 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/aa892f09-0663-4fb6-b3ad-a15c300b5b65-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"aa892f09-0663-4fb6-b3ad-a15c300b5b65\") " pod="openstack/ovsdbserver-sb-0" Feb 17 13:20:49 crc kubenswrapper[4955]: I0217 13:20:49.407664 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa892f09-0663-4fb6-b3ad-a15c300b5b65-config\") pod \"ovsdbserver-sb-0\" (UID: \"aa892f09-0663-4fb6-b3ad-a15c300b5b65\") " pod="openstack/ovsdbserver-sb-0" Feb 17 13:20:49 crc kubenswrapper[4955]: I0217 13:20:49.407688 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa892f09-0663-4fb6-b3ad-a15c300b5b65-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"aa892f09-0663-4fb6-b3ad-a15c300b5b65\") " pod="openstack/ovsdbserver-sb-0" Feb 17 13:20:49 crc kubenswrapper[4955]: I0217 13:20:49.407712 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aa892f09-0663-4fb6-b3ad-a15c300b5b65-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"aa892f09-0663-4fb6-b3ad-a15c300b5b65\") " pod="openstack/ovsdbserver-sb-0" Feb 17 13:20:49 crc kubenswrapper[4955]: I0217 13:20:49.407733 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"aa892f09-0663-4fb6-b3ad-a15c300b5b65\") " pod="openstack/ovsdbserver-sb-0" Feb 17 13:20:49 crc kubenswrapper[4955]: I0217 13:20:49.407757 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa892f09-0663-4fb6-b3ad-a15c300b5b65-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"aa892f09-0663-4fb6-b3ad-a15c300b5b65\") " pod="openstack/ovsdbserver-sb-0" Feb 17 13:20:49 crc kubenswrapper[4955]: I0217 13:20:49.407985 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwxw7\" (UniqueName: \"kubernetes.io/projected/aa892f09-0663-4fb6-b3ad-a15c300b5b65-kube-api-access-jwxw7\") pod \"ovsdbserver-sb-0\" (UID: \"aa892f09-0663-4fb6-b3ad-a15c300b5b65\") " pod="openstack/ovsdbserver-sb-0" Feb 17 13:20:49 crc kubenswrapper[4955]: I0217 13:20:49.510076 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa892f09-0663-4fb6-b3ad-a15c300b5b65-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"aa892f09-0663-4fb6-b3ad-a15c300b5b65\") " pod="openstack/ovsdbserver-sb-0" Feb 17 13:20:49 crc kubenswrapper[4955]: I0217 13:20:49.510125 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/aa892f09-0663-4fb6-b3ad-a15c300b5b65-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"aa892f09-0663-4fb6-b3ad-a15c300b5b65\") " pod="openstack/ovsdbserver-sb-0" Feb 17 13:20:49 crc kubenswrapper[4955]: I0217 13:20:49.510152 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa892f09-0663-4fb6-b3ad-a15c300b5b65-config\") pod \"ovsdbserver-sb-0\" (UID: \"aa892f09-0663-4fb6-b3ad-a15c300b5b65\") " pod="openstack/ovsdbserver-sb-0" Feb 17 13:20:49 crc kubenswrapper[4955]: I0217 13:20:49.510181 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa892f09-0663-4fb6-b3ad-a15c300b5b65-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"aa892f09-0663-4fb6-b3ad-a15c300b5b65\") " pod="openstack/ovsdbserver-sb-0" Feb 17 13:20:49 crc kubenswrapper[4955]: I0217 13:20:49.510215 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aa892f09-0663-4fb6-b3ad-a15c300b5b65-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"aa892f09-0663-4fb6-b3ad-a15c300b5b65\") " pod="openstack/ovsdbserver-sb-0" Feb 17 13:20:49 crc kubenswrapper[4955]: I0217 13:20:49.510245 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"aa892f09-0663-4fb6-b3ad-a15c300b5b65\") " pod="openstack/ovsdbserver-sb-0" Feb 17 13:20:49 crc kubenswrapper[4955]: I0217 13:20:49.510277 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa892f09-0663-4fb6-b3ad-a15c300b5b65-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"aa892f09-0663-4fb6-b3ad-a15c300b5b65\") " pod="openstack/ovsdbserver-sb-0" Feb 17 13:20:49 crc kubenswrapper[4955]: I0217 13:20:49.510343 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwxw7\" (UniqueName: \"kubernetes.io/projected/aa892f09-0663-4fb6-b3ad-a15c300b5b65-kube-api-access-jwxw7\") pod \"ovsdbserver-sb-0\" (UID: \"aa892f09-0663-4fb6-b3ad-a15c300b5b65\") " pod="openstack/ovsdbserver-sb-0" Feb 17 13:20:49 crc kubenswrapper[4955]: I0217 13:20:49.511335 4955 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"aa892f09-0663-4fb6-b3ad-a15c300b5b65\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/ovsdbserver-sb-0" Feb 17 13:20:49 crc kubenswrapper[4955]: I0217 13:20:49.512176 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa892f09-0663-4fb6-b3ad-a15c300b5b65-config\") pod \"ovsdbserver-sb-0\" (UID: \"aa892f09-0663-4fb6-b3ad-a15c300b5b65\") " pod="openstack/ovsdbserver-sb-0" Feb 17 13:20:49 crc kubenswrapper[4955]: I0217 13:20:49.512354 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aa892f09-0663-4fb6-b3ad-a15c300b5b65-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"aa892f09-0663-4fb6-b3ad-a15c300b5b65\") " pod="openstack/ovsdbserver-sb-0" Feb 17 13:20:49 crc kubenswrapper[4955]: I0217 13:20:49.513970 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/aa892f09-0663-4fb6-b3ad-a15c300b5b65-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"aa892f09-0663-4fb6-b3ad-a15c300b5b65\") " pod="openstack/ovsdbserver-sb-0" Feb 17 13:20:49 crc kubenswrapper[4955]: I0217 13:20:49.519983 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa892f09-0663-4fb6-b3ad-a15c300b5b65-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"aa892f09-0663-4fb6-b3ad-a15c300b5b65\") " pod="openstack/ovsdbserver-sb-0" Feb 17 13:20:49 crc kubenswrapper[4955]: I0217 13:20:49.532163 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwxw7\" (UniqueName: \"kubernetes.io/projected/aa892f09-0663-4fb6-b3ad-a15c300b5b65-kube-api-access-jwxw7\") pod \"ovsdbserver-sb-0\" (UID: \"aa892f09-0663-4fb6-b3ad-a15c300b5b65\") " pod="openstack/ovsdbserver-sb-0" Feb 17 13:20:49 crc kubenswrapper[4955]: I0217 13:20:49.540648 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa892f09-0663-4fb6-b3ad-a15c300b5b65-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"aa892f09-0663-4fb6-b3ad-a15c300b5b65\") " pod="openstack/ovsdbserver-sb-0" Feb 17 13:20:49 crc kubenswrapper[4955]: I0217 13:20:49.548872 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa892f09-0663-4fb6-b3ad-a15c300b5b65-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"aa892f09-0663-4fb6-b3ad-a15c300b5b65\") " pod="openstack/ovsdbserver-sb-0" Feb 17 13:20:49 crc kubenswrapper[4955]: I0217 13:20:49.557122 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"aa892f09-0663-4fb6-b3ad-a15c300b5b65\") " pod="openstack/ovsdbserver-sb-0" Feb 17 13:20:49 crc kubenswrapper[4955]: I0217 13:20:49.591801 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Feb 17 13:20:49 crc kubenswrapper[4955]: I0217 13:20:49.857930 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Feb 17 13:20:50 crc kubenswrapper[4955]: W0217 13:20:50.258370 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddef1308a_11ee_4465_89f0_e78ffc4b8b51.slice/crio-e59de605aeafbe034231760ffa4a54027c6858cfb12fbe5596cc95e782738ba1 WatchSource:0}: Error finding container e59de605aeafbe034231760ffa4a54027c6858cfb12fbe5596cc95e782738ba1: Status 404 returned error can't find the container with id e59de605aeafbe034231760ffa4a54027c6858cfb12fbe5596cc95e782738ba1 Feb 17 13:20:50 crc kubenswrapper[4955]: E0217 13:20:50.269195 4955 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Feb 17 13:20:50 crc kubenswrapper[4955]: E0217 13:20:50.269340 4955 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r9mk2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-6h44n_openstack(06fb1f59-b127-4db8-9911-e87a5c9112f9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 17 13:20:50 crc kubenswrapper[4955]: E0217 13:20:50.270512 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-6h44n" podUID="06fb1f59-b127-4db8-9911-e87a5c9112f9" Feb 17 13:20:50 crc kubenswrapper[4955]: E0217 13:20:50.279859 4955 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Feb 17 13:20:50 crc kubenswrapper[4955]: E0217 13:20:50.280009 4955 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-52cb7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-nq5vv_openstack(76ecff2f-ae2a-4cdd-b983-b17e502bc9f3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 17 13:20:50 crc kubenswrapper[4955]: E0217 13:20:50.281246 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-nq5vv" podUID="76ecff2f-ae2a-4cdd-b983-b17e502bc9f3" Feb 17 13:20:50 crc kubenswrapper[4955]: I0217 13:20:50.694865 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-f5xt5" event={"ID":"f7eef812-b3ad-4b76-885d-c8d474cd43e0","Type":"ContainerStarted","Data":"ade39e7ed0dc4c3294ee757df6a6baf431f8c1c351a564d153d5f82cf0896b64"} Feb 17 13:20:50 crc kubenswrapper[4955]: I0217 13:20:50.696291 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"def1308a-11ee-4465-89f0-e78ffc4b8b51","Type":"ContainerStarted","Data":"e59de605aeafbe034231760ffa4a54027c6858cfb12fbe5596cc95e782738ba1"} Feb 17 13:20:50 crc kubenswrapper[4955]: I0217 13:20:50.697351 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-gm68k" event={"ID":"2ced654f-0b1a-4485-ab1d-c98d73e7b8d5","Type":"ContainerStarted","Data":"d0cc4b7ef54033dce204df622bce2e57e0d7fa2b05f6484130ffc4a9da0f6376"} Feb 17 13:20:50 crc kubenswrapper[4955]: I0217 13:20:50.869532 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 17 13:20:50 crc kubenswrapper[4955]: I0217 13:20:50.993501 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.161796 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.187385 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 17 13:20:51 crc kubenswrapper[4955]: W0217 13:20:51.219324 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc241677c_9f24_4cc0_8d15_e166174e4cd0.slice/crio-4513e349aa0229a921459de11450edded87f95558f22921e8e613e999f0e2cf4 WatchSource:0}: Error finding container 4513e349aa0229a921459de11450edded87f95558f22921e8e613e999f0e2cf4: Status 404 returned error can't find the container with id 4513e349aa0229a921459de11450edded87f95558f22921e8e613e999f0e2cf4 Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.317737 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-6h44n" Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.330826 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-nq5vv" Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.408570 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 17 13:20:51 crc kubenswrapper[4955]: W0217 13:20:51.420417 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa892f09_0663_4fb6_b3ad_a15c300b5b65.slice/crio-5edec08c6cae229a2fb24424d5ff10c4388ec7aec0edaf32780dce64bf6f236d WatchSource:0}: Error finding container 5edec08c6cae229a2fb24424d5ff10c4388ec7aec0edaf32780dce64bf6f236d: Status 404 returned error can't find the container with id 5edec08c6cae229a2fb24424d5ff10c4388ec7aec0edaf32780dce64bf6f236d Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.426369 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.431766 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-f8cjz"] Feb 17 13:20:51 crc kubenswrapper[4955]: W0217 13:20:51.447899 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc631bd4b_26ac_44e3_87aa_8878c42e6ce2.slice/crio-a3bbf1e6807283776874cd38f2f3ab5172faa85f27dd4f7335fb13c561776de0 WatchSource:0}: Error finding container a3bbf1e6807283776874cd38f2f3ab5172faa85f27dd4f7335fb13c561776de0: Status 404 returned error can't find the container with id a3bbf1e6807283776874cd38f2f3ab5172faa85f27dd4f7335fb13c561776de0 Feb 17 13:20:51 crc kubenswrapper[4955]: W0217 13:20:51.449449 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod452bb98a_815b_4154_84c5_6b31338fca29.slice/crio-2b292d5d2c7eb01a6f56baae0ec30bce06a38b169d3388529fbc7ad885386a3a WatchSource:0}: Error finding container 2b292d5d2c7eb01a6f56baae0ec30bce06a38b169d3388529fbc7ad885386a3a: Status 404 returned error can't find the container with id 2b292d5d2c7eb01a6f56baae0ec30bce06a38b169d3388529fbc7ad885386a3a Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.456236 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9mk2\" (UniqueName: \"kubernetes.io/projected/06fb1f59-b127-4db8-9911-e87a5c9112f9-kube-api-access-r9mk2\") pod \"06fb1f59-b127-4db8-9911-e87a5c9112f9\" (UID: \"06fb1f59-b127-4db8-9911-e87a5c9112f9\") " Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.456486 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52cb7\" (UniqueName: \"kubernetes.io/projected/76ecff2f-ae2a-4cdd-b983-b17e502bc9f3-kube-api-access-52cb7\") pod \"76ecff2f-ae2a-4cdd-b983-b17e502bc9f3\" (UID: \"76ecff2f-ae2a-4cdd-b983-b17e502bc9f3\") " Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.456521 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/76ecff2f-ae2a-4cdd-b983-b17e502bc9f3-dns-svc\") pod \"76ecff2f-ae2a-4cdd-b983-b17e502bc9f3\" (UID: \"76ecff2f-ae2a-4cdd-b983-b17e502bc9f3\") " Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.456552 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76ecff2f-ae2a-4cdd-b983-b17e502bc9f3-config\") pod \"76ecff2f-ae2a-4cdd-b983-b17e502bc9f3\" (UID: \"76ecff2f-ae2a-4cdd-b983-b17e502bc9f3\") " Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.456594 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06fb1f59-b127-4db8-9911-e87a5c9112f9-config\") pod \"06fb1f59-b127-4db8-9911-e87a5c9112f9\" (UID: \"06fb1f59-b127-4db8-9911-e87a5c9112f9\") " Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.457576 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06fb1f59-b127-4db8-9911-e87a5c9112f9-config" (OuterVolumeSpecName: "config") pod "06fb1f59-b127-4db8-9911-e87a5c9112f9" (UID: "06fb1f59-b127-4db8-9911-e87a5c9112f9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.459207 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76ecff2f-ae2a-4cdd-b983-b17e502bc9f3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "76ecff2f-ae2a-4cdd-b983-b17e502bc9f3" (UID: "76ecff2f-ae2a-4cdd-b983-b17e502bc9f3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.459515 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76ecff2f-ae2a-4cdd-b983-b17e502bc9f3-config" (OuterVolumeSpecName: "config") pod "76ecff2f-ae2a-4cdd-b983-b17e502bc9f3" (UID: "76ecff2f-ae2a-4cdd-b983-b17e502bc9f3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.478239 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06fb1f59-b127-4db8-9911-e87a5c9112f9-kube-api-access-r9mk2" (OuterVolumeSpecName: "kube-api-access-r9mk2") pod "06fb1f59-b127-4db8-9911-e87a5c9112f9" (UID: "06fb1f59-b127-4db8-9911-e87a5c9112f9"). InnerVolumeSpecName "kube-api-access-r9mk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.478525 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76ecff2f-ae2a-4cdd-b983-b17e502bc9f3-kube-api-access-52cb7" (OuterVolumeSpecName: "kube-api-access-52cb7") pod "76ecff2f-ae2a-4cdd-b983-b17e502bc9f3" (UID: "76ecff2f-ae2a-4cdd-b983-b17e502bc9f3"). InnerVolumeSpecName "kube-api-access-52cb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.558659 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52cb7\" (UniqueName: \"kubernetes.io/projected/76ecff2f-ae2a-4cdd-b983-b17e502bc9f3-kube-api-access-52cb7\") on node \"crc\" DevicePath \"\"" Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.559095 4955 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/76ecff2f-ae2a-4cdd-b983-b17e502bc9f3-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.559111 4955 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76ecff2f-ae2a-4cdd-b983-b17e502bc9f3-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.559127 4955 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06fb1f59-b127-4db8-9911-e87a5c9112f9-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.559140 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9mk2\" (UniqueName: \"kubernetes.io/projected/06fb1f59-b127-4db8-9911-e87a5c9112f9-kube-api-access-r9mk2\") on node \"crc\" DevicePath \"\"" Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.577267 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 17 13:20:51 crc kubenswrapper[4955]: W0217 13:20:51.582922 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3384848_891d_46ff_a869_61d657304b74.slice/crio-5159916a21c27c95fe3897dff4a3a414f01f2452af43e7021e3ba03f9a34e6e8 WatchSource:0}: Error finding container 5159916a21c27c95fe3897dff4a3a414f01f2452af43e7021e3ba03f9a34e6e8: Status 404 returned error can't find the container with id 5159916a21c27c95fe3897dff4a3a414f01f2452af43e7021e3ba03f9a34e6e8 Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.679552 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-m27b4"] Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.707045 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"34552e31-68c3-4156-addf-37ea6fdf0b6c","Type":"ContainerStarted","Data":"d03b368ffd4e5b84e52261415ca43843774000447301c378956fb0f752fff984"} Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.707795 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"b3384848-891d-46ff-a869-61d657304b74","Type":"ContainerStarted","Data":"5159916a21c27c95fe3897dff4a3a414f01f2452af43e7021e3ba03f9a34e6e8"} Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.708835 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-nq5vv" event={"ID":"76ecff2f-ae2a-4cdd-b983-b17e502bc9f3","Type":"ContainerDied","Data":"8338feaade4f83b37dfd2ae4e259bf2cecdcc7088e09c4ddb22f97f9252e69b9"} Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.708901 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-nq5vv" Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.711229 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c631bd4b-26ac-44e3-87aa-8878c42e6ce2","Type":"ContainerStarted","Data":"a3bbf1e6807283776874cd38f2f3ab5172faa85f27dd4f7335fb13c561776de0"} Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.712670 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"aa892f09-0663-4fb6-b3ad-a15c300b5b65","Type":"ContainerStarted","Data":"5edec08c6cae229a2fb24424d5ff10c4388ec7aec0edaf32780dce64bf6f236d"} Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.713982 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-m27b4" event={"ID":"9534cd01-b88e-48a4-a2ca-dd968e045dbc","Type":"ContainerStarted","Data":"fbf6f71b6c8dcdf14e4b2e6b0b48032878869f5fd4bbb8e63f1059546f170202"} Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.715681 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c241677c-9f24-4cc0-8d15-e166174e4cd0","Type":"ContainerStarted","Data":"4513e349aa0229a921459de11450edded87f95558f22921e8e613e999f0e2cf4"} Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.717097 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-f8cjz" event={"ID":"452bb98a-815b-4154-84c5-6b31338fca29","Type":"ContainerStarted","Data":"2b292d5d2c7eb01a6f56baae0ec30bce06a38b169d3388529fbc7ad885386a3a"} Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.718228 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"7dab7b98-7d1a-4940-a4c3-8bbbd57fe121","Type":"ContainerStarted","Data":"6ba51654abe2b681be63cccd12d6d5cc10bb0c8e813ea4972f79ead8aca1b29c"} Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.719614 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-6h44n" event={"ID":"06fb1f59-b127-4db8-9911-e87a5c9112f9","Type":"ContainerDied","Data":"965a0633339dad494f040e0059ba1ba39f947fdff38524a3c2285a02945ffa16"} Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.719633 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-6h44n" Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.722344 4955 generic.go:334] "Generic (PLEG): container finished" podID="2ced654f-0b1a-4485-ab1d-c98d73e7b8d5" containerID="d0cc4b7ef54033dce204df622bce2e57e0d7fa2b05f6484130ffc4a9da0f6376" exitCode=0 Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.722459 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-gm68k" event={"ID":"2ced654f-0b1a-4485-ab1d-c98d73e7b8d5","Type":"ContainerDied","Data":"d0cc4b7ef54033dce204df622bce2e57e0d7fa2b05f6484130ffc4a9da0f6376"} Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.725295 4955 generic.go:334] "Generic (PLEG): container finished" podID="f7eef812-b3ad-4b76-885d-c8d474cd43e0" containerID="ade39e7ed0dc4c3294ee757df6a6baf431f8c1c351a564d153d5f82cf0896b64" exitCode=0 Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.725427 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-f5xt5" event={"ID":"f7eef812-b3ad-4b76-885d-c8d474cd43e0","Type":"ContainerDied","Data":"ade39e7ed0dc4c3294ee757df6a6baf431f8c1c351a564d153d5f82cf0896b64"} Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.732512 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1c93012c-0538-49a7-b414-6b01bb6c6dce","Type":"ContainerStarted","Data":"224a8bebdb7992c4ee06ba84aec3a225d2e3983d31d02cdf55817a927301f0c2"} Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.828164 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-6h44n"] Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.842940 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-6h44n"] Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.852877 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-nq5vv"] Feb 17 13:20:51 crc kubenswrapper[4955]: I0217 13:20:51.857609 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-nq5vv"] Feb 17 13:20:52 crc kubenswrapper[4955]: I0217 13:20:52.233846 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06fb1f59-b127-4db8-9911-e87a5c9112f9" path="/var/lib/kubelet/pods/06fb1f59-b127-4db8-9911-e87a5c9112f9/volumes" Feb 17 13:20:52 crc kubenswrapper[4955]: I0217 13:20:52.234666 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76ecff2f-ae2a-4cdd-b983-b17e502bc9f3" path="/var/lib/kubelet/pods/76ecff2f-ae2a-4cdd-b983-b17e502bc9f3/volumes" Feb 17 13:20:52 crc kubenswrapper[4955]: I0217 13:20:52.741848 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-gm68k" event={"ID":"2ced654f-0b1a-4485-ab1d-c98d73e7b8d5","Type":"ContainerStarted","Data":"5dd819e05d1a527aa01c6b9b8e67b683ad3cc4155f24d81c540081a888accde1"} Feb 17 13:20:52 crc kubenswrapper[4955]: I0217 13:20:52.742012 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-gm68k" Feb 17 13:20:52 crc kubenswrapper[4955]: I0217 13:20:52.748128 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-f5xt5" event={"ID":"f7eef812-b3ad-4b76-885d-c8d474cd43e0","Type":"ContainerStarted","Data":"977218e0be442828b357f3cabb51b2efffd7ac393678959c05c0b2f99da8b6aa"} Feb 17 13:20:52 crc kubenswrapper[4955]: I0217 13:20:52.748346 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-f5xt5" Feb 17 13:20:52 crc kubenswrapper[4955]: I0217 13:20:52.765685 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-gm68k" podStartSLOduration=3.93211295 podStartE2EDuration="17.765665754s" podCreationTimestamp="2026-02-17 13:20:35 +0000 UTC" firstStartedPulling="2026-02-17 13:20:36.574569352 +0000 UTC m=+975.097298895" lastFinishedPulling="2026-02-17 13:20:50.408122156 +0000 UTC m=+988.930851699" observedRunningTime="2026-02-17 13:20:52.763578645 +0000 UTC m=+991.286308208" watchObservedRunningTime="2026-02-17 13:20:52.765665754 +0000 UTC m=+991.288395297" Feb 17 13:20:52 crc kubenswrapper[4955]: I0217 13:20:52.785647 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-f5xt5" podStartSLOduration=5.861588092 podStartE2EDuration="16.785625844s" podCreationTimestamp="2026-02-17 13:20:36 +0000 UTC" firstStartedPulling="2026-02-17 13:20:39.472956677 +0000 UTC m=+977.995686220" lastFinishedPulling="2026-02-17 13:20:50.396994429 +0000 UTC m=+988.919723972" observedRunningTime="2026-02-17 13:20:52.77989603 +0000 UTC m=+991.302625583" watchObservedRunningTime="2026-02-17 13:20:52.785625844 +0000 UTC m=+991.308355397" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.352864 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-2s2tl"] Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.355559 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-2s2tl" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.360566 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.365737 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-2s2tl"] Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.512665 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-f5xt5"] Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.529472 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a9d7a6-de50-4a56-b18a-ed69f82ea656-combined-ca-bundle\") pod \"ovn-controller-metrics-2s2tl\" (UID: \"c2a9d7a6-de50-4a56-b18a-ed69f82ea656\") " pod="openstack/ovn-controller-metrics-2s2tl" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.529551 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/c2a9d7a6-de50-4a56-b18a-ed69f82ea656-ovn-rundir\") pod \"ovn-controller-metrics-2s2tl\" (UID: \"c2a9d7a6-de50-4a56-b18a-ed69f82ea656\") " pod="openstack/ovn-controller-metrics-2s2tl" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.529594 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2a9d7a6-de50-4a56-b18a-ed69f82ea656-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-2s2tl\" (UID: \"c2a9d7a6-de50-4a56-b18a-ed69f82ea656\") " pod="openstack/ovn-controller-metrics-2s2tl" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.532467 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/c2a9d7a6-de50-4a56-b18a-ed69f82ea656-ovs-rundir\") pod \"ovn-controller-metrics-2s2tl\" (UID: \"c2a9d7a6-de50-4a56-b18a-ed69f82ea656\") " pod="openstack/ovn-controller-metrics-2s2tl" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.532569 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2a9d7a6-de50-4a56-b18a-ed69f82ea656-config\") pod \"ovn-controller-metrics-2s2tl\" (UID: \"c2a9d7a6-de50-4a56-b18a-ed69f82ea656\") " pod="openstack/ovn-controller-metrics-2s2tl" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.532627 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4z9m\" (UniqueName: \"kubernetes.io/projected/c2a9d7a6-de50-4a56-b18a-ed69f82ea656-kube-api-access-r4z9m\") pod \"ovn-controller-metrics-2s2tl\" (UID: \"c2a9d7a6-de50-4a56-b18a-ed69f82ea656\") " pod="openstack/ovn-controller-metrics-2s2tl" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.559683 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-snn92"] Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.564894 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-snn92" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.569062 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.581600 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-snn92"] Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.633849 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2a9d7a6-de50-4a56-b18a-ed69f82ea656-config\") pod \"ovn-controller-metrics-2s2tl\" (UID: \"c2a9d7a6-de50-4a56-b18a-ed69f82ea656\") " pod="openstack/ovn-controller-metrics-2s2tl" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.633903 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4z9m\" (UniqueName: \"kubernetes.io/projected/c2a9d7a6-de50-4a56-b18a-ed69f82ea656-kube-api-access-r4z9m\") pod \"ovn-controller-metrics-2s2tl\" (UID: \"c2a9d7a6-de50-4a56-b18a-ed69f82ea656\") " pod="openstack/ovn-controller-metrics-2s2tl" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.633983 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a9d7a6-de50-4a56-b18a-ed69f82ea656-combined-ca-bundle\") pod \"ovn-controller-metrics-2s2tl\" (UID: \"c2a9d7a6-de50-4a56-b18a-ed69f82ea656\") " pod="openstack/ovn-controller-metrics-2s2tl" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.634012 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/c2a9d7a6-de50-4a56-b18a-ed69f82ea656-ovn-rundir\") pod \"ovn-controller-metrics-2s2tl\" (UID: \"c2a9d7a6-de50-4a56-b18a-ed69f82ea656\") " pod="openstack/ovn-controller-metrics-2s2tl" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.634035 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2a9d7a6-de50-4a56-b18a-ed69f82ea656-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-2s2tl\" (UID: \"c2a9d7a6-de50-4a56-b18a-ed69f82ea656\") " pod="openstack/ovn-controller-metrics-2s2tl" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.634056 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/c2a9d7a6-de50-4a56-b18a-ed69f82ea656-ovs-rundir\") pod \"ovn-controller-metrics-2s2tl\" (UID: \"c2a9d7a6-de50-4a56-b18a-ed69f82ea656\") " pod="openstack/ovn-controller-metrics-2s2tl" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.634351 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/c2a9d7a6-de50-4a56-b18a-ed69f82ea656-ovs-rundir\") pod \"ovn-controller-metrics-2s2tl\" (UID: \"c2a9d7a6-de50-4a56-b18a-ed69f82ea656\") " pod="openstack/ovn-controller-metrics-2s2tl" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.634532 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/c2a9d7a6-de50-4a56-b18a-ed69f82ea656-ovn-rundir\") pod \"ovn-controller-metrics-2s2tl\" (UID: \"c2a9d7a6-de50-4a56-b18a-ed69f82ea656\") " pod="openstack/ovn-controller-metrics-2s2tl" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.634583 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2a9d7a6-de50-4a56-b18a-ed69f82ea656-config\") pod \"ovn-controller-metrics-2s2tl\" (UID: \"c2a9d7a6-de50-4a56-b18a-ed69f82ea656\") " pod="openstack/ovn-controller-metrics-2s2tl" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.645821 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a9d7a6-de50-4a56-b18a-ed69f82ea656-combined-ca-bundle\") pod \"ovn-controller-metrics-2s2tl\" (UID: \"c2a9d7a6-de50-4a56-b18a-ed69f82ea656\") " pod="openstack/ovn-controller-metrics-2s2tl" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.650226 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2a9d7a6-de50-4a56-b18a-ed69f82ea656-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-2s2tl\" (UID: \"c2a9d7a6-de50-4a56-b18a-ed69f82ea656\") " pod="openstack/ovn-controller-metrics-2s2tl" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.651315 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4z9m\" (UniqueName: \"kubernetes.io/projected/c2a9d7a6-de50-4a56-b18a-ed69f82ea656-kube-api-access-r4z9m\") pod \"ovn-controller-metrics-2s2tl\" (UID: \"c2a9d7a6-de50-4a56-b18a-ed69f82ea656\") " pod="openstack/ovn-controller-metrics-2s2tl" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.692873 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-2s2tl" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.720739 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-gm68k"] Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.737641 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/701369d5-716b-4e3d-b3ad-c896ffba2f49-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-snn92\" (UID: \"701369d5-716b-4e3d-b3ad-c896ffba2f49\") " pod="openstack/dnsmasq-dns-5bf47b49b7-snn92" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.737746 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/701369d5-716b-4e3d-b3ad-c896ffba2f49-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-snn92\" (UID: \"701369d5-716b-4e3d-b3ad-c896ffba2f49\") " pod="openstack/dnsmasq-dns-5bf47b49b7-snn92" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.737804 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/701369d5-716b-4e3d-b3ad-c896ffba2f49-config\") pod \"dnsmasq-dns-5bf47b49b7-snn92\" (UID: \"701369d5-716b-4e3d-b3ad-c896ffba2f49\") " pod="openstack/dnsmasq-dns-5bf47b49b7-snn92" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.737822 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6r4x\" (UniqueName: \"kubernetes.io/projected/701369d5-716b-4e3d-b3ad-c896ffba2f49-kube-api-access-v6r4x\") pod \"dnsmasq-dns-5bf47b49b7-snn92\" (UID: \"701369d5-716b-4e3d-b3ad-c896ffba2f49\") " pod="openstack/dnsmasq-dns-5bf47b49b7-snn92" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.744039 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8554648995-jpznm"] Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.745371 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-jpznm" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.757133 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.777347 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-jpznm"] Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.840909 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/701369d5-716b-4e3d-b3ad-c896ffba2f49-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-snn92\" (UID: \"701369d5-716b-4e3d-b3ad-c896ffba2f49\") " pod="openstack/dnsmasq-dns-5bf47b49b7-snn92" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.840977 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/701369d5-716b-4e3d-b3ad-c896ffba2f49-config\") pod \"dnsmasq-dns-5bf47b49b7-snn92\" (UID: \"701369d5-716b-4e3d-b3ad-c896ffba2f49\") " pod="openstack/dnsmasq-dns-5bf47b49b7-snn92" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.840999 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6r4x\" (UniqueName: \"kubernetes.io/projected/701369d5-716b-4e3d-b3ad-c896ffba2f49-kube-api-access-v6r4x\") pod \"dnsmasq-dns-5bf47b49b7-snn92\" (UID: \"701369d5-716b-4e3d-b3ad-c896ffba2f49\") " pod="openstack/dnsmasq-dns-5bf47b49b7-snn92" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.841080 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lknw\" (UniqueName: \"kubernetes.io/projected/dd0771e6-147c-4069-88e1-c9ae122dc55d-kube-api-access-2lknw\") pod \"dnsmasq-dns-8554648995-jpznm\" (UID: \"dd0771e6-147c-4069-88e1-c9ae122dc55d\") " pod="openstack/dnsmasq-dns-8554648995-jpznm" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.841112 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dd0771e6-147c-4069-88e1-c9ae122dc55d-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-jpznm\" (UID: \"dd0771e6-147c-4069-88e1-c9ae122dc55d\") " pod="openstack/dnsmasq-dns-8554648995-jpznm" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.841132 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/701369d5-716b-4e3d-b3ad-c896ffba2f49-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-snn92\" (UID: \"701369d5-716b-4e3d-b3ad-c896ffba2f49\") " pod="openstack/dnsmasq-dns-5bf47b49b7-snn92" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.841172 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dd0771e6-147c-4069-88e1-c9ae122dc55d-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-jpznm\" (UID: \"dd0771e6-147c-4069-88e1-c9ae122dc55d\") " pod="openstack/dnsmasq-dns-8554648995-jpznm" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.841206 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd0771e6-147c-4069-88e1-c9ae122dc55d-config\") pod \"dnsmasq-dns-8554648995-jpznm\" (UID: \"dd0771e6-147c-4069-88e1-c9ae122dc55d\") " pod="openstack/dnsmasq-dns-8554648995-jpznm" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.841242 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd0771e6-147c-4069-88e1-c9ae122dc55d-dns-svc\") pod \"dnsmasq-dns-8554648995-jpznm\" (UID: \"dd0771e6-147c-4069-88e1-c9ae122dc55d\") " pod="openstack/dnsmasq-dns-8554648995-jpznm" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.842113 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/701369d5-716b-4e3d-b3ad-c896ffba2f49-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-snn92\" (UID: \"701369d5-716b-4e3d-b3ad-c896ffba2f49\") " pod="openstack/dnsmasq-dns-5bf47b49b7-snn92" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.842691 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/701369d5-716b-4e3d-b3ad-c896ffba2f49-config\") pod \"dnsmasq-dns-5bf47b49b7-snn92\" (UID: \"701369d5-716b-4e3d-b3ad-c896ffba2f49\") " pod="openstack/dnsmasq-dns-5bf47b49b7-snn92" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.843382 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/701369d5-716b-4e3d-b3ad-c896ffba2f49-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-snn92\" (UID: \"701369d5-716b-4e3d-b3ad-c896ffba2f49\") " pod="openstack/dnsmasq-dns-5bf47b49b7-snn92" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.863616 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6r4x\" (UniqueName: \"kubernetes.io/projected/701369d5-716b-4e3d-b3ad-c896ffba2f49-kube-api-access-v6r4x\") pod \"dnsmasq-dns-5bf47b49b7-snn92\" (UID: \"701369d5-716b-4e3d-b3ad-c896ffba2f49\") " pod="openstack/dnsmasq-dns-5bf47b49b7-snn92" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.896608 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-snn92" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.942915 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lknw\" (UniqueName: \"kubernetes.io/projected/dd0771e6-147c-4069-88e1-c9ae122dc55d-kube-api-access-2lknw\") pod \"dnsmasq-dns-8554648995-jpznm\" (UID: \"dd0771e6-147c-4069-88e1-c9ae122dc55d\") " pod="openstack/dnsmasq-dns-8554648995-jpznm" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.942983 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dd0771e6-147c-4069-88e1-c9ae122dc55d-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-jpznm\" (UID: \"dd0771e6-147c-4069-88e1-c9ae122dc55d\") " pod="openstack/dnsmasq-dns-8554648995-jpznm" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.943045 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dd0771e6-147c-4069-88e1-c9ae122dc55d-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-jpznm\" (UID: \"dd0771e6-147c-4069-88e1-c9ae122dc55d\") " pod="openstack/dnsmasq-dns-8554648995-jpznm" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.943080 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd0771e6-147c-4069-88e1-c9ae122dc55d-config\") pod \"dnsmasq-dns-8554648995-jpznm\" (UID: \"dd0771e6-147c-4069-88e1-c9ae122dc55d\") " pod="openstack/dnsmasq-dns-8554648995-jpznm" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.943118 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd0771e6-147c-4069-88e1-c9ae122dc55d-dns-svc\") pod \"dnsmasq-dns-8554648995-jpznm\" (UID: \"dd0771e6-147c-4069-88e1-c9ae122dc55d\") " pod="openstack/dnsmasq-dns-8554648995-jpznm" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.945246 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dd0771e6-147c-4069-88e1-c9ae122dc55d-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-jpznm\" (UID: \"dd0771e6-147c-4069-88e1-c9ae122dc55d\") " pod="openstack/dnsmasq-dns-8554648995-jpznm" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.945370 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd0771e6-147c-4069-88e1-c9ae122dc55d-dns-svc\") pod \"dnsmasq-dns-8554648995-jpznm\" (UID: \"dd0771e6-147c-4069-88e1-c9ae122dc55d\") " pod="openstack/dnsmasq-dns-8554648995-jpznm" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.945647 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd0771e6-147c-4069-88e1-c9ae122dc55d-config\") pod \"dnsmasq-dns-8554648995-jpznm\" (UID: \"dd0771e6-147c-4069-88e1-c9ae122dc55d\") " pod="openstack/dnsmasq-dns-8554648995-jpznm" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.945875 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dd0771e6-147c-4069-88e1-c9ae122dc55d-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-jpznm\" (UID: \"dd0771e6-147c-4069-88e1-c9ae122dc55d\") " pod="openstack/dnsmasq-dns-8554648995-jpznm" Feb 17 13:20:53 crc kubenswrapper[4955]: I0217 13:20:53.961880 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lknw\" (UniqueName: \"kubernetes.io/projected/dd0771e6-147c-4069-88e1-c9ae122dc55d-kube-api-access-2lknw\") pod \"dnsmasq-dns-8554648995-jpznm\" (UID: \"dd0771e6-147c-4069-88e1-c9ae122dc55d\") " pod="openstack/dnsmasq-dns-8554648995-jpznm" Feb 17 13:20:54 crc kubenswrapper[4955]: I0217 13:20:54.082198 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-jpznm" Feb 17 13:20:54 crc kubenswrapper[4955]: I0217 13:20:54.785900 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-f5xt5" podUID="f7eef812-b3ad-4b76-885d-c8d474cd43e0" containerName="dnsmasq-dns" containerID="cri-o://977218e0be442828b357f3cabb51b2efffd7ac393678959c05c0b2f99da8b6aa" gracePeriod=10 Feb 17 13:20:54 crc kubenswrapper[4955]: I0217 13:20:54.786054 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-gm68k" podUID="2ced654f-0b1a-4485-ab1d-c98d73e7b8d5" containerName="dnsmasq-dns" containerID="cri-o://5dd819e05d1a527aa01c6b9b8e67b683ad3cc4155f24d81c540081a888accde1" gracePeriod=10 Feb 17 13:20:55 crc kubenswrapper[4955]: I0217 13:20:55.799716 4955 generic.go:334] "Generic (PLEG): container finished" podID="2ced654f-0b1a-4485-ab1d-c98d73e7b8d5" containerID="5dd819e05d1a527aa01c6b9b8e67b683ad3cc4155f24d81c540081a888accde1" exitCode=0 Feb 17 13:20:55 crc kubenswrapper[4955]: I0217 13:20:55.799978 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-gm68k" event={"ID":"2ced654f-0b1a-4485-ab1d-c98d73e7b8d5","Type":"ContainerDied","Data":"5dd819e05d1a527aa01c6b9b8e67b683ad3cc4155f24d81c540081a888accde1"} Feb 17 13:20:55 crc kubenswrapper[4955]: I0217 13:20:55.802740 4955 generic.go:334] "Generic (PLEG): container finished" podID="f7eef812-b3ad-4b76-885d-c8d474cd43e0" containerID="977218e0be442828b357f3cabb51b2efffd7ac393678959c05c0b2f99da8b6aa" exitCode=0 Feb 17 13:20:55 crc kubenswrapper[4955]: I0217 13:20:55.802800 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-f5xt5" event={"ID":"f7eef812-b3ad-4b76-885d-c8d474cd43e0","Type":"ContainerDied","Data":"977218e0be442828b357f3cabb51b2efffd7ac393678959c05c0b2f99da8b6aa"} Feb 17 13:20:57 crc kubenswrapper[4955]: I0217 13:20:57.348897 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-gm68k" Feb 17 13:20:57 crc kubenswrapper[4955]: I0217 13:20:57.496831 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ced654f-0b1a-4485-ab1d-c98d73e7b8d5-config\") pod \"2ced654f-0b1a-4485-ab1d-c98d73e7b8d5\" (UID: \"2ced654f-0b1a-4485-ab1d-c98d73e7b8d5\") " Feb 17 13:20:57 crc kubenswrapper[4955]: I0217 13:20:57.497258 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxrwg\" (UniqueName: \"kubernetes.io/projected/2ced654f-0b1a-4485-ab1d-c98d73e7b8d5-kube-api-access-cxrwg\") pod \"2ced654f-0b1a-4485-ab1d-c98d73e7b8d5\" (UID: \"2ced654f-0b1a-4485-ab1d-c98d73e7b8d5\") " Feb 17 13:20:57 crc kubenswrapper[4955]: I0217 13:20:57.497289 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ced654f-0b1a-4485-ab1d-c98d73e7b8d5-dns-svc\") pod \"2ced654f-0b1a-4485-ab1d-c98d73e7b8d5\" (UID: \"2ced654f-0b1a-4485-ab1d-c98d73e7b8d5\") " Feb 17 13:20:57 crc kubenswrapper[4955]: I0217 13:20:57.504027 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ced654f-0b1a-4485-ab1d-c98d73e7b8d5-kube-api-access-cxrwg" (OuterVolumeSpecName: "kube-api-access-cxrwg") pod "2ced654f-0b1a-4485-ab1d-c98d73e7b8d5" (UID: "2ced654f-0b1a-4485-ab1d-c98d73e7b8d5"). InnerVolumeSpecName "kube-api-access-cxrwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:20:57 crc kubenswrapper[4955]: I0217 13:20:57.532199 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ced654f-0b1a-4485-ab1d-c98d73e7b8d5-config" (OuterVolumeSpecName: "config") pod "2ced654f-0b1a-4485-ab1d-c98d73e7b8d5" (UID: "2ced654f-0b1a-4485-ab1d-c98d73e7b8d5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:20:57 crc kubenswrapper[4955]: I0217 13:20:57.566703 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ced654f-0b1a-4485-ab1d-c98d73e7b8d5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2ced654f-0b1a-4485-ab1d-c98d73e7b8d5" (UID: "2ced654f-0b1a-4485-ab1d-c98d73e7b8d5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:20:57 crc kubenswrapper[4955]: I0217 13:20:57.599755 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxrwg\" (UniqueName: \"kubernetes.io/projected/2ced654f-0b1a-4485-ab1d-c98d73e7b8d5-kube-api-access-cxrwg\") on node \"crc\" DevicePath \"\"" Feb 17 13:20:57 crc kubenswrapper[4955]: I0217 13:20:57.599808 4955 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ced654f-0b1a-4485-ab1d-c98d73e7b8d5-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 17 13:20:57 crc kubenswrapper[4955]: I0217 13:20:57.599821 4955 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ced654f-0b1a-4485-ab1d-c98d73e7b8d5-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:20:57 crc kubenswrapper[4955]: I0217 13:20:57.823375 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-gm68k" event={"ID":"2ced654f-0b1a-4485-ab1d-c98d73e7b8d5","Type":"ContainerDied","Data":"f9bbc2b6db45f390b0ab2c836561c5073bc73ea9c9db27367b0a0c873ec9c300"} Feb 17 13:20:57 crc kubenswrapper[4955]: I0217 13:20:57.823460 4955 scope.go:117] "RemoveContainer" containerID="5dd819e05d1a527aa01c6b9b8e67b683ad3cc4155f24d81c540081a888accde1" Feb 17 13:20:57 crc kubenswrapper[4955]: I0217 13:20:57.823466 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-gm68k" Feb 17 13:20:57 crc kubenswrapper[4955]: I0217 13:20:57.854090 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-gm68k"] Feb 17 13:20:57 crc kubenswrapper[4955]: I0217 13:20:57.859468 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-gm68k"] Feb 17 13:20:58 crc kubenswrapper[4955]: I0217 13:20:58.232499 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ced654f-0b1a-4485-ab1d-c98d73e7b8d5" path="/var/lib/kubelet/pods/2ced654f-0b1a-4485-ab1d-c98d73e7b8d5/volumes" Feb 17 13:21:00 crc kubenswrapper[4955]: I0217 13:21:00.859985 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-f5xt5" event={"ID":"f7eef812-b3ad-4b76-885d-c8d474cd43e0","Type":"ContainerDied","Data":"b51bff11b1d3b806cf7981806717ab805b59ed5bcf2ccf515625950b46687882"} Feb 17 13:21:00 crc kubenswrapper[4955]: I0217 13:21:00.860267 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b51bff11b1d3b806cf7981806717ab805b59ed5bcf2ccf515625950b46687882" Feb 17 13:21:00 crc kubenswrapper[4955]: I0217 13:21:00.876065 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-f5xt5" Feb 17 13:21:01 crc kubenswrapper[4955]: I0217 13:21:01.013947 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-666b6646f7-gm68k" podUID="2ced654f-0b1a-4485-ab1d-c98d73e7b8d5" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.96:5353: i/o timeout" Feb 17 13:21:01 crc kubenswrapper[4955]: I0217 13:21:01.063963 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjddn\" (UniqueName: \"kubernetes.io/projected/f7eef812-b3ad-4b76-885d-c8d474cd43e0-kube-api-access-pjddn\") pod \"f7eef812-b3ad-4b76-885d-c8d474cd43e0\" (UID: \"f7eef812-b3ad-4b76-885d-c8d474cd43e0\") " Feb 17 13:21:01 crc kubenswrapper[4955]: I0217 13:21:01.064385 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f7eef812-b3ad-4b76-885d-c8d474cd43e0-dns-svc\") pod \"f7eef812-b3ad-4b76-885d-c8d474cd43e0\" (UID: \"f7eef812-b3ad-4b76-885d-c8d474cd43e0\") " Feb 17 13:21:01 crc kubenswrapper[4955]: I0217 13:21:01.064554 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7eef812-b3ad-4b76-885d-c8d474cd43e0-config\") pod \"f7eef812-b3ad-4b76-885d-c8d474cd43e0\" (UID: \"f7eef812-b3ad-4b76-885d-c8d474cd43e0\") " Feb 17 13:21:01 crc kubenswrapper[4955]: I0217 13:21:01.068942 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7eef812-b3ad-4b76-885d-c8d474cd43e0-kube-api-access-pjddn" (OuterVolumeSpecName: "kube-api-access-pjddn") pod "f7eef812-b3ad-4b76-885d-c8d474cd43e0" (UID: "f7eef812-b3ad-4b76-885d-c8d474cd43e0"). InnerVolumeSpecName "kube-api-access-pjddn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:21:01 crc kubenswrapper[4955]: I0217 13:21:01.102003 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7eef812-b3ad-4b76-885d-c8d474cd43e0-config" (OuterVolumeSpecName: "config") pod "f7eef812-b3ad-4b76-885d-c8d474cd43e0" (UID: "f7eef812-b3ad-4b76-885d-c8d474cd43e0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:21:01 crc kubenswrapper[4955]: I0217 13:21:01.116776 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7eef812-b3ad-4b76-885d-c8d474cd43e0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f7eef812-b3ad-4b76-885d-c8d474cd43e0" (UID: "f7eef812-b3ad-4b76-885d-c8d474cd43e0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:21:01 crc kubenswrapper[4955]: I0217 13:21:01.166657 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjddn\" (UniqueName: \"kubernetes.io/projected/f7eef812-b3ad-4b76-885d-c8d474cd43e0-kube-api-access-pjddn\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:01 crc kubenswrapper[4955]: I0217 13:21:01.166749 4955 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f7eef812-b3ad-4b76-885d-c8d474cd43e0-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:01 crc kubenswrapper[4955]: I0217 13:21:01.166762 4955 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7eef812-b3ad-4b76-885d-c8d474cd43e0-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:01 crc kubenswrapper[4955]: I0217 13:21:01.647478 4955 scope.go:117] "RemoveContainer" containerID="d0cc4b7ef54033dce204df622bce2e57e0d7fa2b05f6484130ffc4a9da0f6376" Feb 17 13:21:01 crc kubenswrapper[4955]: I0217 13:21:01.867071 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-f5xt5" Feb 17 13:21:01 crc kubenswrapper[4955]: I0217 13:21:01.907831 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-f5xt5"] Feb 17 13:21:01 crc kubenswrapper[4955]: I0217 13:21:01.915419 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-f5xt5"] Feb 17 13:21:02 crc kubenswrapper[4955]: I0217 13:21:02.110425 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-snn92"] Feb 17 13:21:02 crc kubenswrapper[4955]: I0217 13:21:02.165966 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-jpznm"] Feb 17 13:21:02 crc kubenswrapper[4955]: I0217 13:21:02.240504 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7eef812-b3ad-4b76-885d-c8d474cd43e0" path="/var/lib/kubelet/pods/f7eef812-b3ad-4b76-885d-c8d474cd43e0/volumes" Feb 17 13:21:02 crc kubenswrapper[4955]: W0217 13:21:02.329628 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod701369d5_716b_4e3d_b3ad_c896ffba2f49.slice/crio-bcb7dc4aa5ad956a47ba9bc786fe0ed26c7874adba181c463828e96dce746ad1 WatchSource:0}: Error finding container bcb7dc4aa5ad956a47ba9bc786fe0ed26c7874adba181c463828e96dce746ad1: Status 404 returned error can't find the container with id bcb7dc4aa5ad956a47ba9bc786fe0ed26c7874adba181c463828e96dce746ad1 Feb 17 13:21:02 crc kubenswrapper[4955]: I0217 13:21:02.353572 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-2s2tl"] Feb 17 13:21:02 crc kubenswrapper[4955]: W0217 13:21:02.783536 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2a9d7a6_de50_4a56_b18a_ed69f82ea656.slice/crio-30db8cc245e0bb6ddcfe22f673dfcf8eb11e1bcdd8d83b3767f0c4645c207a5f WatchSource:0}: Error finding container 30db8cc245e0bb6ddcfe22f673dfcf8eb11e1bcdd8d83b3767f0c4645c207a5f: Status 404 returned error can't find the container with id 30db8cc245e0bb6ddcfe22f673dfcf8eb11e1bcdd8d83b3767f0c4645c207a5f Feb 17 13:21:02 crc kubenswrapper[4955]: I0217 13:21:02.886769 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-snn92" event={"ID":"701369d5-716b-4e3d-b3ad-c896ffba2f49","Type":"ContainerStarted","Data":"bcb7dc4aa5ad956a47ba9bc786fe0ed26c7874adba181c463828e96dce746ad1"} Feb 17 13:21:02 crc kubenswrapper[4955]: I0217 13:21:02.890315 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-2s2tl" event={"ID":"c2a9d7a6-de50-4a56-b18a-ed69f82ea656","Type":"ContainerStarted","Data":"30db8cc245e0bb6ddcfe22f673dfcf8eb11e1bcdd8d83b3767f0c4645c207a5f"} Feb 17 13:21:02 crc kubenswrapper[4955]: I0217 13:21:02.892240 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-jpznm" event={"ID":"dd0771e6-147c-4069-88e1-c9ae122dc55d","Type":"ContainerStarted","Data":"dd637fd908e2b69a43093810df7623aaaebda64cfccc3a0b1aec87cc9a1af775"} Feb 17 13:21:03 crc kubenswrapper[4955]: I0217 13:21:03.903582 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c631bd4b-26ac-44e3-87aa-8878c42e6ce2","Type":"ContainerStarted","Data":"4a097e5724bf1f20c0f6234932ca8c7b9bcd1c3f18c46c0839ea88f05514bba0"} Feb 17 13:21:03 crc kubenswrapper[4955]: I0217 13:21:03.903991 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Feb 17 13:21:03 crc kubenswrapper[4955]: I0217 13:21:03.907845 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"aa892f09-0663-4fb6-b3ad-a15c300b5b65","Type":"ContainerStarted","Data":"962d7307eba151c5e4ad0abb095c7a34e1b7ae5dc1470e644ad4e046ac036c43"} Feb 17 13:21:03 crc kubenswrapper[4955]: I0217 13:21:03.909626 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"7dab7b98-7d1a-4940-a4c3-8bbbd57fe121","Type":"ContainerStarted","Data":"41b6627c88a1f2e0284fc361652d731a1555083409a7b37226cd4f11de7da8d9"} Feb 17 13:21:03 crc kubenswrapper[4955]: I0217 13:21:03.909773 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Feb 17 13:21:03 crc kubenswrapper[4955]: I0217 13:21:03.911336 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"b3384848-891d-46ff-a869-61d657304b74","Type":"ContainerStarted","Data":"b3e4eb310d9898d8c83b4d0d7d394127efd8c4905300b849836d9d777543f17e"} Feb 17 13:21:03 crc kubenswrapper[4955]: I0217 13:21:03.916255 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-jpznm" event={"ID":"dd0771e6-147c-4069-88e1-c9ae122dc55d","Type":"ContainerStarted","Data":"c08cce887e36aaef58e4a5e04814b2e233ede2d92804b418dbcd45477d0427d8"} Feb 17 13:21:03 crc kubenswrapper[4955]: I0217 13:21:03.935209 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=10.464837423 podStartE2EDuration="21.935185669s" podCreationTimestamp="2026-02-17 13:20:42 +0000 UTC" firstStartedPulling="2026-02-17 13:20:51.452122386 +0000 UTC m=+989.974851929" lastFinishedPulling="2026-02-17 13:21:02.922470632 +0000 UTC m=+1001.445200175" observedRunningTime="2026-02-17 13:21:03.923352762 +0000 UTC m=+1002.446082315" watchObservedRunningTime="2026-02-17 13:21:03.935185669 +0000 UTC m=+1002.457915212" Feb 17 13:21:03 crc kubenswrapper[4955]: I0217 13:21:03.940296 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1c93012c-0538-49a7-b414-6b01bb6c6dce","Type":"ContainerStarted","Data":"09eebb5430a4ddb98d0c852216062a45cc435b09fff25be33ce6e1bf4fb6fa08"} Feb 17 13:21:03 crc kubenswrapper[4955]: I0217 13:21:03.942125 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-m27b4" event={"ID":"9534cd01-b88e-48a4-a2ca-dd968e045dbc","Type":"ContainerStarted","Data":"ba53d920b5f1920a8bceba705b0dec98f32017344154428bb3a8d1d39c3fc21e"} Feb 17 13:21:03 crc kubenswrapper[4955]: I0217 13:21:03.943550 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"def1308a-11ee-4465-89f0-e78ffc4b8b51","Type":"ContainerStarted","Data":"83440afae7ffa2213c6bd44becee9c0e72651142225fc199faaabfbec332fc41"} Feb 17 13:21:03 crc kubenswrapper[4955]: I0217 13:21:03.944988 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-f8cjz" event={"ID":"452bb98a-815b-4154-84c5-6b31338fca29","Type":"ContainerStarted","Data":"b318ea8dcffd2c8c18879244565e7d6106212f757761d0a3b2f02334f42bf245"} Feb 17 13:21:03 crc kubenswrapper[4955]: I0217 13:21:03.945590 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-f8cjz" Feb 17 13:21:03 crc kubenswrapper[4955]: I0217 13:21:03.947773 4955 generic.go:334] "Generic (PLEG): container finished" podID="701369d5-716b-4e3d-b3ad-c896ffba2f49" containerID="f91188b614f542c5e42ccef060451702618f50fd955f3b01cdc649a1324958fc" exitCode=0 Feb 17 13:21:03 crc kubenswrapper[4955]: I0217 13:21:03.947864 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-snn92" event={"ID":"701369d5-716b-4e3d-b3ad-c896ffba2f49","Type":"ContainerDied","Data":"f91188b614f542c5e42ccef060451702618f50fd955f3b01cdc649a1324958fc"} Feb 17 13:21:03 crc kubenswrapper[4955]: I0217 13:21:03.962632 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=13.410276502 podStartE2EDuration="23.962614162s" podCreationTimestamp="2026-02-17 13:20:40 +0000 UTC" firstStartedPulling="2026-02-17 13:20:51.210402531 +0000 UTC m=+989.733132074" lastFinishedPulling="2026-02-17 13:21:01.762740181 +0000 UTC m=+1000.285469734" observedRunningTime="2026-02-17 13:21:03.960584414 +0000 UTC m=+1002.483313947" watchObservedRunningTime="2026-02-17 13:21:03.962614162 +0000 UTC m=+1002.485343705" Feb 17 13:21:04 crc kubenswrapper[4955]: I0217 13:21:04.079562 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-f8cjz" podStartSLOduration=8.566882248 podStartE2EDuration="19.079540676s" podCreationTimestamp="2026-02-17 13:20:45 +0000 UTC" firstStartedPulling="2026-02-17 13:20:51.453605898 +0000 UTC m=+989.976335451" lastFinishedPulling="2026-02-17 13:21:01.966264336 +0000 UTC m=+1000.488993879" observedRunningTime="2026-02-17 13:21:04.074483702 +0000 UTC m=+1002.597213265" watchObservedRunningTime="2026-02-17 13:21:04.079540676 +0000 UTC m=+1002.602270219" Feb 17 13:21:04 crc kubenswrapper[4955]: I0217 13:21:04.959167 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-snn92" event={"ID":"701369d5-716b-4e3d-b3ad-c896ffba2f49","Type":"ContainerStarted","Data":"b9912d8124616cd9ecc844aac7abfea134563de45fe2da2b2d28965254c35572"} Feb 17 13:21:04 crc kubenswrapper[4955]: I0217 13:21:04.959277 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5bf47b49b7-snn92" Feb 17 13:21:04 crc kubenswrapper[4955]: I0217 13:21:04.965297 4955 generic.go:334] "Generic (PLEG): container finished" podID="dd0771e6-147c-4069-88e1-c9ae122dc55d" containerID="c08cce887e36aaef58e4a5e04814b2e233ede2d92804b418dbcd45477d0427d8" exitCode=0 Feb 17 13:21:04 crc kubenswrapper[4955]: I0217 13:21:04.965461 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-jpznm" event={"ID":"dd0771e6-147c-4069-88e1-c9ae122dc55d","Type":"ContainerDied","Data":"c08cce887e36aaef58e4a5e04814b2e233ede2d92804b418dbcd45477d0427d8"} Feb 17 13:21:04 crc kubenswrapper[4955]: I0217 13:21:04.967534 4955 generic.go:334] "Generic (PLEG): container finished" podID="9534cd01-b88e-48a4-a2ca-dd968e045dbc" containerID="ba53d920b5f1920a8bceba705b0dec98f32017344154428bb3a8d1d39c3fc21e" exitCode=0 Feb 17 13:21:04 crc kubenswrapper[4955]: I0217 13:21:04.967738 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-m27b4" event={"ID":"9534cd01-b88e-48a4-a2ca-dd968e045dbc","Type":"ContainerDied","Data":"ba53d920b5f1920a8bceba705b0dec98f32017344154428bb3a8d1d39c3fc21e"} Feb 17 13:21:04 crc kubenswrapper[4955]: I0217 13:21:04.969887 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c241677c-9f24-4cc0-8d15-e166174e4cd0","Type":"ContainerStarted","Data":"cd19a92beba58ac2a112c23c7364fb146fdc4af6d6854544bdce5d7e774494a2"} Feb 17 13:21:04 crc kubenswrapper[4955]: I0217 13:21:04.972569 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"34552e31-68c3-4156-addf-37ea6fdf0b6c","Type":"ContainerStarted","Data":"a8a6c6d9a2920a8907fcd1815d0cabffa7db57ffe67f73f31657d130b21005d5"} Feb 17 13:21:04 crc kubenswrapper[4955]: I0217 13:21:04.982525 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5bf47b49b7-snn92" podStartSLOduration=11.982499153 podStartE2EDuration="11.982499153s" podCreationTimestamp="2026-02-17 13:20:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:21:04.978752136 +0000 UTC m=+1003.501481689" watchObservedRunningTime="2026-02-17 13:21:04.982499153 +0000 UTC m=+1003.505228696" Feb 17 13:21:05 crc kubenswrapper[4955]: I0217 13:21:05.982750 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-m27b4" event={"ID":"9534cd01-b88e-48a4-a2ca-dd968e045dbc","Type":"ContainerStarted","Data":"77eae0e6da75837ecc677692a9266f225cbadef89c09b6011e5b683511418485"} Feb 17 13:21:05 crc kubenswrapper[4955]: I0217 13:21:05.983412 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-m27b4" event={"ID":"9534cd01-b88e-48a4-a2ca-dd968e045dbc","Type":"ContainerStarted","Data":"1bc5182758f1c2ab7dc9cba1d43971ecbc286bc07bde43057bfd8e1641b76e86"} Feb 17 13:21:05 crc kubenswrapper[4955]: I0217 13:21:05.986257 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"aa892f09-0663-4fb6-b3ad-a15c300b5b65","Type":"ContainerStarted","Data":"56894367e9d990cc7c68dc0ceb3c0eaa41400b95b0f516b43a630c1b95a801e6"} Feb 17 13:21:05 crc kubenswrapper[4955]: I0217 13:21:05.988242 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"b3384848-891d-46ff-a869-61d657304b74","Type":"ContainerStarted","Data":"9816633596caceb851f65b01f33554e83b7d4705d4eabedd5ec232dc2b3b3559"} Feb 17 13:21:05 crc kubenswrapper[4955]: I0217 13:21:05.989809 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-2s2tl" event={"ID":"c2a9d7a6-de50-4a56-b18a-ed69f82ea656","Type":"ContainerStarted","Data":"42e3ab9e7577c6547a9e65ec9375a4b1f3fcae2dedfd48a6af65a32bd5467000"} Feb 17 13:21:05 crc kubenswrapper[4955]: I0217 13:21:05.991978 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-jpznm" event={"ID":"dd0771e6-147c-4069-88e1-c9ae122dc55d","Type":"ContainerStarted","Data":"d0afe951ba8e2e8c737ca5d913a77d6e9b224faca55462a95c93bd441480fded"} Feb 17 13:21:05 crc kubenswrapper[4955]: I0217 13:21:05.992257 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8554648995-jpznm" Feb 17 13:21:06 crc kubenswrapper[4955]: I0217 13:21:06.010285 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-m27b4" podStartSLOduration=10.930546181 podStartE2EDuration="21.01026954s" podCreationTimestamp="2026-02-17 13:20:45 +0000 UTC" firstStartedPulling="2026-02-17 13:20:51.682820576 +0000 UTC m=+990.205550119" lastFinishedPulling="2026-02-17 13:21:01.762543935 +0000 UTC m=+1000.285273478" observedRunningTime="2026-02-17 13:21:06.008396656 +0000 UTC m=+1004.531126199" watchObservedRunningTime="2026-02-17 13:21:06.01026954 +0000 UTC m=+1004.532999083" Feb 17 13:21:06 crc kubenswrapper[4955]: I0217 13:21:06.030628 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=8.633625698 podStartE2EDuration="22.0306083s" podCreationTimestamp="2026-02-17 13:20:44 +0000 UTC" firstStartedPulling="2026-02-17 13:20:51.58443727 +0000 UTC m=+990.107166813" lastFinishedPulling="2026-02-17 13:21:04.981419872 +0000 UTC m=+1003.504149415" observedRunningTime="2026-02-17 13:21:06.025281948 +0000 UTC m=+1004.548011491" watchObservedRunningTime="2026-02-17 13:21:06.0306083 +0000 UTC m=+1004.553337833" Feb 17 13:21:06 crc kubenswrapper[4955]: I0217 13:21:06.047591 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=4.605245369 podStartE2EDuration="18.047572714s" podCreationTimestamp="2026-02-17 13:20:48 +0000 UTC" firstStartedPulling="2026-02-17 13:20:51.42420955 +0000 UTC m=+989.946939093" lastFinishedPulling="2026-02-17 13:21:04.866536905 +0000 UTC m=+1003.389266438" observedRunningTime="2026-02-17 13:21:06.045863905 +0000 UTC m=+1004.568593468" watchObservedRunningTime="2026-02-17 13:21:06.047572714 +0000 UTC m=+1004.570302257" Feb 17 13:21:06 crc kubenswrapper[4955]: I0217 13:21:06.047985 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Feb 17 13:21:06 crc kubenswrapper[4955]: I0217 13:21:06.068421 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8554648995-jpznm" podStartSLOduration=13.068405078 podStartE2EDuration="13.068405078s" podCreationTimestamp="2026-02-17 13:20:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:21:06.066979967 +0000 UTC m=+1004.589709510" watchObservedRunningTime="2026-02-17 13:21:06.068405078 +0000 UTC m=+1004.591134621" Feb 17 13:21:06 crc kubenswrapper[4955]: I0217 13:21:06.085603 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-2s2tl" podStartSLOduration=11.0996296 podStartE2EDuration="13.085586878s" podCreationTimestamp="2026-02-17 13:20:53 +0000 UTC" firstStartedPulling="2026-02-17 13:21:02.849069528 +0000 UTC m=+1001.371799071" lastFinishedPulling="2026-02-17 13:21:04.835026806 +0000 UTC m=+1003.357756349" observedRunningTime="2026-02-17 13:21:06.080904105 +0000 UTC m=+1004.603633658" watchObservedRunningTime="2026-02-17 13:21:06.085586878 +0000 UTC m=+1004.608316421" Feb 17 13:21:07 crc kubenswrapper[4955]: I0217 13:21:07.001805 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-m27b4" Feb 17 13:21:07 crc kubenswrapper[4955]: I0217 13:21:07.001856 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-m27b4" Feb 17 13:21:07 crc kubenswrapper[4955]: I0217 13:21:07.048399 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Feb 17 13:21:07 crc kubenswrapper[4955]: I0217 13:21:07.106589 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Feb 17 13:21:07 crc kubenswrapper[4955]: I0217 13:21:07.592637 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Feb 17 13:21:07 crc kubenswrapper[4955]: I0217 13:21:07.638237 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Feb 17 13:21:08 crc kubenswrapper[4955]: I0217 13:21:08.010898 4955 generic.go:334] "Generic (PLEG): container finished" podID="def1308a-11ee-4465-89f0-e78ffc4b8b51" containerID="83440afae7ffa2213c6bd44becee9c0e72651142225fc199faaabfbec332fc41" exitCode=0 Feb 17 13:21:08 crc kubenswrapper[4955]: I0217 13:21:08.010991 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"def1308a-11ee-4465-89f0-e78ffc4b8b51","Type":"ContainerDied","Data":"83440afae7ffa2213c6bd44becee9c0e72651142225fc199faaabfbec332fc41"} Feb 17 13:21:08 crc kubenswrapper[4955]: I0217 13:21:08.012974 4955 generic.go:334] "Generic (PLEG): container finished" podID="1c93012c-0538-49a7-b414-6b01bb6c6dce" containerID="09eebb5430a4ddb98d0c852216062a45cc435b09fff25be33ce6e1bf4fb6fa08" exitCode=0 Feb 17 13:21:08 crc kubenswrapper[4955]: I0217 13:21:08.013325 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1c93012c-0538-49a7-b414-6b01bb6c6dce","Type":"ContainerDied","Data":"09eebb5430a4ddb98d0c852216062a45cc435b09fff25be33ce6e1bf4fb6fa08"} Feb 17 13:21:08 crc kubenswrapper[4955]: I0217 13:21:08.013616 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Feb 17 13:21:08 crc kubenswrapper[4955]: I0217 13:21:08.056661 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Feb 17 13:21:08 crc kubenswrapper[4955]: I0217 13:21:08.066246 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Feb 17 13:21:08 crc kubenswrapper[4955]: I0217 13:21:08.420368 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Feb 17 13:21:08 crc kubenswrapper[4955]: E0217 13:21:08.420954 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7eef812-b3ad-4b76-885d-c8d474cd43e0" containerName="init" Feb 17 13:21:08 crc kubenswrapper[4955]: I0217 13:21:08.420972 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7eef812-b3ad-4b76-885d-c8d474cd43e0" containerName="init" Feb 17 13:21:08 crc kubenswrapper[4955]: E0217 13:21:08.420995 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7eef812-b3ad-4b76-885d-c8d474cd43e0" containerName="dnsmasq-dns" Feb 17 13:21:08 crc kubenswrapper[4955]: I0217 13:21:08.421001 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7eef812-b3ad-4b76-885d-c8d474cd43e0" containerName="dnsmasq-dns" Feb 17 13:21:08 crc kubenswrapper[4955]: E0217 13:21:08.421028 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ced654f-0b1a-4485-ab1d-c98d73e7b8d5" containerName="init" Feb 17 13:21:08 crc kubenswrapper[4955]: I0217 13:21:08.421035 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ced654f-0b1a-4485-ab1d-c98d73e7b8d5" containerName="init" Feb 17 13:21:08 crc kubenswrapper[4955]: E0217 13:21:08.421059 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ced654f-0b1a-4485-ab1d-c98d73e7b8d5" containerName="dnsmasq-dns" Feb 17 13:21:08 crc kubenswrapper[4955]: I0217 13:21:08.421064 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ced654f-0b1a-4485-ab1d-c98d73e7b8d5" containerName="dnsmasq-dns" Feb 17 13:21:08 crc kubenswrapper[4955]: I0217 13:21:08.421255 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ced654f-0b1a-4485-ab1d-c98d73e7b8d5" containerName="dnsmasq-dns" Feb 17 13:21:08 crc kubenswrapper[4955]: I0217 13:21:08.421265 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7eef812-b3ad-4b76-885d-c8d474cd43e0" containerName="dnsmasq-dns" Feb 17 13:21:08 crc kubenswrapper[4955]: I0217 13:21:08.422053 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Feb 17 13:21:08 crc kubenswrapper[4955]: I0217 13:21:08.424427 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Feb 17 13:21:08 crc kubenswrapper[4955]: I0217 13:21:08.424474 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Feb 17 13:21:08 crc kubenswrapper[4955]: I0217 13:21:08.424716 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Feb 17 13:21:08 crc kubenswrapper[4955]: I0217 13:21:08.428135 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-pbwvj" Feb 17 13:21:08 crc kubenswrapper[4955]: I0217 13:21:08.433694 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Feb 17 13:21:08 crc kubenswrapper[4955]: I0217 13:21:08.487855 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5c949\" (UniqueName: \"kubernetes.io/projected/94e88c0f-4478-4de1-9340-99904aac77e9-kube-api-access-5c949\") pod \"ovn-northd-0\" (UID: \"94e88c0f-4478-4de1-9340-99904aac77e9\") " pod="openstack/ovn-northd-0" Feb 17 13:21:08 crc kubenswrapper[4955]: I0217 13:21:08.487914 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/94e88c0f-4478-4de1-9340-99904aac77e9-scripts\") pod \"ovn-northd-0\" (UID: \"94e88c0f-4478-4de1-9340-99904aac77e9\") " pod="openstack/ovn-northd-0" Feb 17 13:21:08 crc kubenswrapper[4955]: I0217 13:21:08.487943 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/94e88c0f-4478-4de1-9340-99904aac77e9-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"94e88c0f-4478-4de1-9340-99904aac77e9\") " pod="openstack/ovn-northd-0" Feb 17 13:21:08 crc kubenswrapper[4955]: I0217 13:21:08.488081 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94e88c0f-4478-4de1-9340-99904aac77e9-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"94e88c0f-4478-4de1-9340-99904aac77e9\") " pod="openstack/ovn-northd-0" Feb 17 13:21:08 crc kubenswrapper[4955]: I0217 13:21:08.488106 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/94e88c0f-4478-4de1-9340-99904aac77e9-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"94e88c0f-4478-4de1-9340-99904aac77e9\") " pod="openstack/ovn-northd-0" Feb 17 13:21:08 crc kubenswrapper[4955]: I0217 13:21:08.488169 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94e88c0f-4478-4de1-9340-99904aac77e9-config\") pod \"ovn-northd-0\" (UID: \"94e88c0f-4478-4de1-9340-99904aac77e9\") " pod="openstack/ovn-northd-0" Feb 17 13:21:08 crc kubenswrapper[4955]: I0217 13:21:08.488201 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/94e88c0f-4478-4de1-9340-99904aac77e9-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"94e88c0f-4478-4de1-9340-99904aac77e9\") " pod="openstack/ovn-northd-0" Feb 17 13:21:08 crc kubenswrapper[4955]: I0217 13:21:08.589915 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5c949\" (UniqueName: \"kubernetes.io/projected/94e88c0f-4478-4de1-9340-99904aac77e9-kube-api-access-5c949\") pod \"ovn-northd-0\" (UID: \"94e88c0f-4478-4de1-9340-99904aac77e9\") " pod="openstack/ovn-northd-0" Feb 17 13:21:08 crc kubenswrapper[4955]: I0217 13:21:08.589964 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/94e88c0f-4478-4de1-9340-99904aac77e9-scripts\") pod \"ovn-northd-0\" (UID: \"94e88c0f-4478-4de1-9340-99904aac77e9\") " pod="openstack/ovn-northd-0" Feb 17 13:21:08 crc kubenswrapper[4955]: I0217 13:21:08.589991 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/94e88c0f-4478-4de1-9340-99904aac77e9-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"94e88c0f-4478-4de1-9340-99904aac77e9\") " pod="openstack/ovn-northd-0" Feb 17 13:21:08 crc kubenswrapper[4955]: I0217 13:21:08.590051 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94e88c0f-4478-4de1-9340-99904aac77e9-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"94e88c0f-4478-4de1-9340-99904aac77e9\") " pod="openstack/ovn-northd-0" Feb 17 13:21:08 crc kubenswrapper[4955]: I0217 13:21:08.590066 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/94e88c0f-4478-4de1-9340-99904aac77e9-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"94e88c0f-4478-4de1-9340-99904aac77e9\") " pod="openstack/ovn-northd-0" Feb 17 13:21:08 crc kubenswrapper[4955]: I0217 13:21:08.590105 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94e88c0f-4478-4de1-9340-99904aac77e9-config\") pod \"ovn-northd-0\" (UID: \"94e88c0f-4478-4de1-9340-99904aac77e9\") " pod="openstack/ovn-northd-0" Feb 17 13:21:08 crc kubenswrapper[4955]: I0217 13:21:08.590130 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/94e88c0f-4478-4de1-9340-99904aac77e9-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"94e88c0f-4478-4de1-9340-99904aac77e9\") " pod="openstack/ovn-northd-0" Feb 17 13:21:08 crc kubenswrapper[4955]: I0217 13:21:08.590607 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/94e88c0f-4478-4de1-9340-99904aac77e9-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"94e88c0f-4478-4de1-9340-99904aac77e9\") " pod="openstack/ovn-northd-0" Feb 17 13:21:08 crc kubenswrapper[4955]: I0217 13:21:08.591194 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94e88c0f-4478-4de1-9340-99904aac77e9-config\") pod \"ovn-northd-0\" (UID: \"94e88c0f-4478-4de1-9340-99904aac77e9\") " pod="openstack/ovn-northd-0" Feb 17 13:21:08 crc kubenswrapper[4955]: I0217 13:21:08.591531 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/94e88c0f-4478-4de1-9340-99904aac77e9-scripts\") pod \"ovn-northd-0\" (UID: \"94e88c0f-4478-4de1-9340-99904aac77e9\") " pod="openstack/ovn-northd-0" Feb 17 13:21:08 crc kubenswrapper[4955]: I0217 13:21:08.595232 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/94e88c0f-4478-4de1-9340-99904aac77e9-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"94e88c0f-4478-4de1-9340-99904aac77e9\") " pod="openstack/ovn-northd-0" Feb 17 13:21:08 crc kubenswrapper[4955]: I0217 13:21:08.604078 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/94e88c0f-4478-4de1-9340-99904aac77e9-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"94e88c0f-4478-4de1-9340-99904aac77e9\") " pod="openstack/ovn-northd-0" Feb 17 13:21:08 crc kubenswrapper[4955]: I0217 13:21:08.605924 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94e88c0f-4478-4de1-9340-99904aac77e9-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"94e88c0f-4478-4de1-9340-99904aac77e9\") " pod="openstack/ovn-northd-0" Feb 17 13:21:08 crc kubenswrapper[4955]: I0217 13:21:08.607660 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5c949\" (UniqueName: \"kubernetes.io/projected/94e88c0f-4478-4de1-9340-99904aac77e9-kube-api-access-5c949\") pod \"ovn-northd-0\" (UID: \"94e88c0f-4478-4de1-9340-99904aac77e9\") " pod="openstack/ovn-northd-0" Feb 17 13:21:08 crc kubenswrapper[4955]: I0217 13:21:08.740210 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Feb 17 13:21:09 crc kubenswrapper[4955]: I0217 13:21:09.026645 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"def1308a-11ee-4465-89f0-e78ffc4b8b51","Type":"ContainerStarted","Data":"2c85f39f2a5bc6b8925f942914d2b59bf1ab955fa87eecf99b744c5174fd3881"} Feb 17 13:21:09 crc kubenswrapper[4955]: I0217 13:21:09.033186 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1c93012c-0538-49a7-b414-6b01bb6c6dce","Type":"ContainerStarted","Data":"bb685d355cb91727527d74ec64cee553a61d3278d0554f2f829f60752d8a408e"} Feb 17 13:21:09 crc kubenswrapper[4955]: I0217 13:21:09.062355 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=20.816429014 podStartE2EDuration="32.062334138s" podCreationTimestamp="2026-02-17 13:20:37 +0000 UTC" firstStartedPulling="2026-02-17 13:20:50.263693386 +0000 UTC m=+988.786422929" lastFinishedPulling="2026-02-17 13:21:01.50959851 +0000 UTC m=+1000.032328053" observedRunningTime="2026-02-17 13:21:09.056755729 +0000 UTC m=+1007.579485282" watchObservedRunningTime="2026-02-17 13:21:09.062334138 +0000 UTC m=+1007.585063701" Feb 17 13:21:09 crc kubenswrapper[4955]: I0217 13:21:09.077525 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=19.767312822 podStartE2EDuration="31.07750368s" podCreationTimestamp="2026-02-17 13:20:38 +0000 UTC" firstStartedPulling="2026-02-17 13:20:51.022934883 +0000 UTC m=+989.545664426" lastFinishedPulling="2026-02-17 13:21:02.333125741 +0000 UTC m=+1000.855855284" observedRunningTime="2026-02-17 13:21:09.077215352 +0000 UTC m=+1007.599944915" watchObservedRunningTime="2026-02-17 13:21:09.07750368 +0000 UTC m=+1007.600233223" Feb 17 13:21:09 crc kubenswrapper[4955]: I0217 13:21:09.189517 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Feb 17 13:21:09 crc kubenswrapper[4955]: W0217 13:21:09.195216 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94e88c0f_4478_4de1_9340_99904aac77e9.slice/crio-28301012dcfedb63211b49499a7be69d891005fd00da6efa11a29242d1751967 WatchSource:0}: Error finding container 28301012dcfedb63211b49499a7be69d891005fd00da6efa11a29242d1751967: Status 404 returned error can't find the container with id 28301012dcfedb63211b49499a7be69d891005fd00da6efa11a29242d1751967 Feb 17 13:21:10 crc kubenswrapper[4955]: I0217 13:21:10.043162 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"94e88c0f-4478-4de1-9340-99904aac77e9","Type":"ContainerStarted","Data":"28301012dcfedb63211b49499a7be69d891005fd00da6efa11a29242d1751967"} Feb 17 13:21:10 crc kubenswrapper[4955]: I0217 13:21:10.050087 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Feb 17 13:21:10 crc kubenswrapper[4955]: I0217 13:21:10.050372 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Feb 17 13:21:10 crc kubenswrapper[4955]: I0217 13:21:10.402714 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Feb 17 13:21:11 crc kubenswrapper[4955]: I0217 13:21:11.055667 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"94e88c0f-4478-4de1-9340-99904aac77e9","Type":"ContainerStarted","Data":"bbbd36a2e584bef8d309f9197d4deb9e476acad4f3b287a1b5abed44c7cbbd54"} Feb 17 13:21:11 crc kubenswrapper[4955]: I0217 13:21:11.056131 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Feb 17 13:21:11 crc kubenswrapper[4955]: I0217 13:21:11.056151 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"94e88c0f-4478-4de1-9340-99904aac77e9","Type":"ContainerStarted","Data":"27e8f0c804451ab934c25e4be862afb1493f4481a5afa2466ca59e6ab07c4593"} Feb 17 13:21:11 crc kubenswrapper[4955]: I0217 13:21:11.076404 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=1.8329526889999999 podStartE2EDuration="3.076384378s" podCreationTimestamp="2026-02-17 13:21:08 +0000 UTC" firstStartedPulling="2026-02-17 13:21:09.198187573 +0000 UTC m=+1007.720917106" lastFinishedPulling="2026-02-17 13:21:10.441619242 +0000 UTC m=+1008.964348795" observedRunningTime="2026-02-17 13:21:11.07399326 +0000 UTC m=+1009.596722813" watchObservedRunningTime="2026-02-17 13:21:11.076384378 +0000 UTC m=+1009.599113911" Feb 17 13:21:12 crc kubenswrapper[4955]: I0217 13:21:12.756714 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-snn92"] Feb 17 13:21:12 crc kubenswrapper[4955]: I0217 13:21:12.757154 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5bf47b49b7-snn92" podUID="701369d5-716b-4e3d-b3ad-c896ffba2f49" containerName="dnsmasq-dns" containerID="cri-o://b9912d8124616cd9ecc844aac7abfea134563de45fe2da2b2d28965254c35572" gracePeriod=10 Feb 17 13:21:12 crc kubenswrapper[4955]: I0217 13:21:12.759055 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5bf47b49b7-snn92" Feb 17 13:21:12 crc kubenswrapper[4955]: I0217 13:21:12.765101 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Feb 17 13:21:12 crc kubenswrapper[4955]: I0217 13:21:12.797952 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-dwwb9"] Feb 17 13:21:12 crc kubenswrapper[4955]: I0217 13:21:12.806556 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-dwwb9" Feb 17 13:21:12 crc kubenswrapper[4955]: I0217 13:21:12.825572 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-dwwb9"] Feb 17 13:21:12 crc kubenswrapper[4955]: I0217 13:21:12.953393 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61f387f2-8694-4f14-a7a7-f996a7ed8efe-config\") pod \"dnsmasq-dns-b8fbc5445-dwwb9\" (UID: \"61f387f2-8694-4f14-a7a7-f996a7ed8efe\") " pod="openstack/dnsmasq-dns-b8fbc5445-dwwb9" Feb 17 13:21:12 crc kubenswrapper[4955]: I0217 13:21:12.953443 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-px2ht\" (UniqueName: \"kubernetes.io/projected/61f387f2-8694-4f14-a7a7-f996a7ed8efe-kube-api-access-px2ht\") pod \"dnsmasq-dns-b8fbc5445-dwwb9\" (UID: \"61f387f2-8694-4f14-a7a7-f996a7ed8efe\") " pod="openstack/dnsmasq-dns-b8fbc5445-dwwb9" Feb 17 13:21:12 crc kubenswrapper[4955]: I0217 13:21:12.953507 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61f387f2-8694-4f14-a7a7-f996a7ed8efe-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-dwwb9\" (UID: \"61f387f2-8694-4f14-a7a7-f996a7ed8efe\") " pod="openstack/dnsmasq-dns-b8fbc5445-dwwb9" Feb 17 13:21:12 crc kubenswrapper[4955]: I0217 13:21:12.953541 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/61f387f2-8694-4f14-a7a7-f996a7ed8efe-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-dwwb9\" (UID: \"61f387f2-8694-4f14-a7a7-f996a7ed8efe\") " pod="openstack/dnsmasq-dns-b8fbc5445-dwwb9" Feb 17 13:21:12 crc kubenswrapper[4955]: I0217 13:21:12.953583 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/61f387f2-8694-4f14-a7a7-f996a7ed8efe-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-dwwb9\" (UID: \"61f387f2-8694-4f14-a7a7-f996a7ed8efe\") " pod="openstack/dnsmasq-dns-b8fbc5445-dwwb9" Feb 17 13:21:13 crc kubenswrapper[4955]: I0217 13:21:13.057248 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/61f387f2-8694-4f14-a7a7-f996a7ed8efe-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-dwwb9\" (UID: \"61f387f2-8694-4f14-a7a7-f996a7ed8efe\") " pod="openstack/dnsmasq-dns-b8fbc5445-dwwb9" Feb 17 13:21:13 crc kubenswrapper[4955]: I0217 13:21:13.057472 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/61f387f2-8694-4f14-a7a7-f996a7ed8efe-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-dwwb9\" (UID: \"61f387f2-8694-4f14-a7a7-f996a7ed8efe\") " pod="openstack/dnsmasq-dns-b8fbc5445-dwwb9" Feb 17 13:21:13 crc kubenswrapper[4955]: I0217 13:21:13.057559 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61f387f2-8694-4f14-a7a7-f996a7ed8efe-config\") pod \"dnsmasq-dns-b8fbc5445-dwwb9\" (UID: \"61f387f2-8694-4f14-a7a7-f996a7ed8efe\") " pod="openstack/dnsmasq-dns-b8fbc5445-dwwb9" Feb 17 13:21:13 crc kubenswrapper[4955]: I0217 13:21:13.057815 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-px2ht\" (UniqueName: \"kubernetes.io/projected/61f387f2-8694-4f14-a7a7-f996a7ed8efe-kube-api-access-px2ht\") pod \"dnsmasq-dns-b8fbc5445-dwwb9\" (UID: \"61f387f2-8694-4f14-a7a7-f996a7ed8efe\") " pod="openstack/dnsmasq-dns-b8fbc5445-dwwb9" Feb 17 13:21:13 crc kubenswrapper[4955]: I0217 13:21:13.057955 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61f387f2-8694-4f14-a7a7-f996a7ed8efe-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-dwwb9\" (UID: \"61f387f2-8694-4f14-a7a7-f996a7ed8efe\") " pod="openstack/dnsmasq-dns-b8fbc5445-dwwb9" Feb 17 13:21:13 crc kubenswrapper[4955]: I0217 13:21:13.058771 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61f387f2-8694-4f14-a7a7-f996a7ed8efe-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-dwwb9\" (UID: \"61f387f2-8694-4f14-a7a7-f996a7ed8efe\") " pod="openstack/dnsmasq-dns-b8fbc5445-dwwb9" Feb 17 13:21:13 crc kubenswrapper[4955]: I0217 13:21:13.059525 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/61f387f2-8694-4f14-a7a7-f996a7ed8efe-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-dwwb9\" (UID: \"61f387f2-8694-4f14-a7a7-f996a7ed8efe\") " pod="openstack/dnsmasq-dns-b8fbc5445-dwwb9" Feb 17 13:21:13 crc kubenswrapper[4955]: I0217 13:21:13.060241 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/61f387f2-8694-4f14-a7a7-f996a7ed8efe-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-dwwb9\" (UID: \"61f387f2-8694-4f14-a7a7-f996a7ed8efe\") " pod="openstack/dnsmasq-dns-b8fbc5445-dwwb9" Feb 17 13:21:13 crc kubenswrapper[4955]: I0217 13:21:13.060303 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61f387f2-8694-4f14-a7a7-f996a7ed8efe-config\") pod \"dnsmasq-dns-b8fbc5445-dwwb9\" (UID: \"61f387f2-8694-4f14-a7a7-f996a7ed8efe\") " pod="openstack/dnsmasq-dns-b8fbc5445-dwwb9" Feb 17 13:21:13 crc kubenswrapper[4955]: I0217 13:21:13.087015 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-px2ht\" (UniqueName: \"kubernetes.io/projected/61f387f2-8694-4f14-a7a7-f996a7ed8efe-kube-api-access-px2ht\") pod \"dnsmasq-dns-b8fbc5445-dwwb9\" (UID: \"61f387f2-8694-4f14-a7a7-f996a7ed8efe\") " pod="openstack/dnsmasq-dns-b8fbc5445-dwwb9" Feb 17 13:21:13 crc kubenswrapper[4955]: I0217 13:21:13.118089 4955 generic.go:334] "Generic (PLEG): container finished" podID="701369d5-716b-4e3d-b3ad-c896ffba2f49" containerID="b9912d8124616cd9ecc844aac7abfea134563de45fe2da2b2d28965254c35572" exitCode=0 Feb 17 13:21:13 crc kubenswrapper[4955]: I0217 13:21:13.118410 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-snn92" event={"ID":"701369d5-716b-4e3d-b3ad-c896ffba2f49","Type":"ContainerDied","Data":"b9912d8124616cd9ecc844aac7abfea134563de45fe2da2b2d28965254c35572"} Feb 17 13:21:13 crc kubenswrapper[4955]: I0217 13:21:13.206190 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-dwwb9" Feb 17 13:21:13 crc kubenswrapper[4955]: I0217 13:21:13.438371 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-snn92" Feb 17 13:21:13 crc kubenswrapper[4955]: I0217 13:21:13.566930 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6r4x\" (UniqueName: \"kubernetes.io/projected/701369d5-716b-4e3d-b3ad-c896ffba2f49-kube-api-access-v6r4x\") pod \"701369d5-716b-4e3d-b3ad-c896ffba2f49\" (UID: \"701369d5-716b-4e3d-b3ad-c896ffba2f49\") " Feb 17 13:21:13 crc kubenswrapper[4955]: I0217 13:21:13.567981 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/701369d5-716b-4e3d-b3ad-c896ffba2f49-config\") pod \"701369d5-716b-4e3d-b3ad-c896ffba2f49\" (UID: \"701369d5-716b-4e3d-b3ad-c896ffba2f49\") " Feb 17 13:21:13 crc kubenswrapper[4955]: I0217 13:21:13.568156 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/701369d5-716b-4e3d-b3ad-c896ffba2f49-dns-svc\") pod \"701369d5-716b-4e3d-b3ad-c896ffba2f49\" (UID: \"701369d5-716b-4e3d-b3ad-c896ffba2f49\") " Feb 17 13:21:13 crc kubenswrapper[4955]: I0217 13:21:13.568177 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/701369d5-716b-4e3d-b3ad-c896ffba2f49-ovsdbserver-nb\") pod \"701369d5-716b-4e3d-b3ad-c896ffba2f49\" (UID: \"701369d5-716b-4e3d-b3ad-c896ffba2f49\") " Feb 17 13:21:13 crc kubenswrapper[4955]: I0217 13:21:13.572903 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/701369d5-716b-4e3d-b3ad-c896ffba2f49-kube-api-access-v6r4x" (OuterVolumeSpecName: "kube-api-access-v6r4x") pod "701369d5-716b-4e3d-b3ad-c896ffba2f49" (UID: "701369d5-716b-4e3d-b3ad-c896ffba2f49"). InnerVolumeSpecName "kube-api-access-v6r4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:21:13 crc kubenswrapper[4955]: I0217 13:21:13.611516 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/701369d5-716b-4e3d-b3ad-c896ffba2f49-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "701369d5-716b-4e3d-b3ad-c896ffba2f49" (UID: "701369d5-716b-4e3d-b3ad-c896ffba2f49"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:21:13 crc kubenswrapper[4955]: I0217 13:21:13.619614 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/701369d5-716b-4e3d-b3ad-c896ffba2f49-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "701369d5-716b-4e3d-b3ad-c896ffba2f49" (UID: "701369d5-716b-4e3d-b3ad-c896ffba2f49"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:21:13 crc kubenswrapper[4955]: I0217 13:21:13.630047 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/701369d5-716b-4e3d-b3ad-c896ffba2f49-config" (OuterVolumeSpecName: "config") pod "701369d5-716b-4e3d-b3ad-c896ffba2f49" (UID: "701369d5-716b-4e3d-b3ad-c896ffba2f49"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:21:13 crc kubenswrapper[4955]: I0217 13:21:13.667350 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-dwwb9"] Feb 17 13:21:13 crc kubenswrapper[4955]: I0217 13:21:13.670519 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6r4x\" (UniqueName: \"kubernetes.io/projected/701369d5-716b-4e3d-b3ad-c896ffba2f49-kube-api-access-v6r4x\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:13 crc kubenswrapper[4955]: I0217 13:21:13.670547 4955 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/701369d5-716b-4e3d-b3ad-c896ffba2f49-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:13 crc kubenswrapper[4955]: I0217 13:21:13.670558 4955 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/701369d5-716b-4e3d-b3ad-c896ffba2f49-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:13 crc kubenswrapper[4955]: I0217 13:21:13.670567 4955 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/701369d5-716b-4e3d-b3ad-c896ffba2f49-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:13 crc kubenswrapper[4955]: I0217 13:21:13.920154 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Feb 17 13:21:13 crc kubenswrapper[4955]: E0217 13:21:13.920569 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="701369d5-716b-4e3d-b3ad-c896ffba2f49" containerName="init" Feb 17 13:21:13 crc kubenswrapper[4955]: I0217 13:21:13.920590 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="701369d5-716b-4e3d-b3ad-c896ffba2f49" containerName="init" Feb 17 13:21:13 crc kubenswrapper[4955]: E0217 13:21:13.920628 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="701369d5-716b-4e3d-b3ad-c896ffba2f49" containerName="dnsmasq-dns" Feb 17 13:21:13 crc kubenswrapper[4955]: I0217 13:21:13.920637 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="701369d5-716b-4e3d-b3ad-c896ffba2f49" containerName="dnsmasq-dns" Feb 17 13:21:13 crc kubenswrapper[4955]: I0217 13:21:13.920983 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="701369d5-716b-4e3d-b3ad-c896ffba2f49" containerName="dnsmasq-dns" Feb 17 13:21:13 crc kubenswrapper[4955]: I0217 13:21:13.946649 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Feb 17 13:21:13 crc kubenswrapper[4955]: I0217 13:21:13.950686 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Feb 17 13:21:13 crc kubenswrapper[4955]: I0217 13:21:13.951146 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Feb 17 13:21:13 crc kubenswrapper[4955]: I0217 13:21:13.951180 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Feb 17 13:21:13 crc kubenswrapper[4955]: I0217 13:21:13.951322 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-2l4mf" Feb 17 13:21:13 crc kubenswrapper[4955]: I0217 13:21:13.959566 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Feb 17 13:21:14 crc kubenswrapper[4955]: I0217 13:21:14.080593 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/fb6d9434-4476-438b-90e4-4781bd85b0bb-cache\") pod \"swift-storage-0\" (UID: \"fb6d9434-4476-438b-90e4-4781bd85b0bb\") " pod="openstack/swift-storage-0" Feb 17 13:21:14 crc kubenswrapper[4955]: I0217 13:21:14.080711 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/fb6d9434-4476-438b-90e4-4781bd85b0bb-lock\") pod \"swift-storage-0\" (UID: \"fb6d9434-4476-438b-90e4-4781bd85b0bb\") " pod="openstack/swift-storage-0" Feb 17 13:21:14 crc kubenswrapper[4955]: I0217 13:21:14.080737 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zv4d8\" (UniqueName: \"kubernetes.io/projected/fb6d9434-4476-438b-90e4-4781bd85b0bb-kube-api-access-zv4d8\") pod \"swift-storage-0\" (UID: \"fb6d9434-4476-438b-90e4-4781bd85b0bb\") " pod="openstack/swift-storage-0" Feb 17 13:21:14 crc kubenswrapper[4955]: I0217 13:21:14.081124 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb6d9434-4476-438b-90e4-4781bd85b0bb-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"fb6d9434-4476-438b-90e4-4781bd85b0bb\") " pod="openstack/swift-storage-0" Feb 17 13:21:14 crc kubenswrapper[4955]: I0217 13:21:14.081198 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"fb6d9434-4476-438b-90e4-4781bd85b0bb\") " pod="openstack/swift-storage-0" Feb 17 13:21:14 crc kubenswrapper[4955]: I0217 13:21:14.081326 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/fb6d9434-4476-438b-90e4-4781bd85b0bb-etc-swift\") pod \"swift-storage-0\" (UID: \"fb6d9434-4476-438b-90e4-4781bd85b0bb\") " pod="openstack/swift-storage-0" Feb 17 13:21:14 crc kubenswrapper[4955]: I0217 13:21:14.084896 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8554648995-jpznm" Feb 17 13:21:14 crc kubenswrapper[4955]: I0217 13:21:14.130621 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-snn92" event={"ID":"701369d5-716b-4e3d-b3ad-c896ffba2f49","Type":"ContainerDied","Data":"bcb7dc4aa5ad956a47ba9bc786fe0ed26c7874adba181c463828e96dce746ad1"} Feb 17 13:21:14 crc kubenswrapper[4955]: I0217 13:21:14.130720 4955 scope.go:117] "RemoveContainer" containerID="b9912d8124616cd9ecc844aac7abfea134563de45fe2da2b2d28965254c35572" Feb 17 13:21:14 crc kubenswrapper[4955]: I0217 13:21:14.130649 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-snn92" Feb 17 13:21:14 crc kubenswrapper[4955]: I0217 13:21:14.133127 4955 generic.go:334] "Generic (PLEG): container finished" podID="61f387f2-8694-4f14-a7a7-f996a7ed8efe" containerID="f2b4cbe4ffcb0fdc840eb2b1cbf94fb31dd007c9ce03477751ed41ca9969e12c" exitCode=0 Feb 17 13:21:14 crc kubenswrapper[4955]: I0217 13:21:14.133183 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-dwwb9" event={"ID":"61f387f2-8694-4f14-a7a7-f996a7ed8efe","Type":"ContainerDied","Data":"f2b4cbe4ffcb0fdc840eb2b1cbf94fb31dd007c9ce03477751ed41ca9969e12c"} Feb 17 13:21:14 crc kubenswrapper[4955]: I0217 13:21:14.133219 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-dwwb9" event={"ID":"61f387f2-8694-4f14-a7a7-f996a7ed8efe","Type":"ContainerStarted","Data":"8c46dd48f25cc39cb6b50f19f918a7aa1bfd09bcb04e9e36799ff7039cf0bf84"} Feb 17 13:21:14 crc kubenswrapper[4955]: I0217 13:21:14.157812 4955 scope.go:117] "RemoveContainer" containerID="f91188b614f542c5e42ccef060451702618f50fd955f3b01cdc649a1324958fc" Feb 17 13:21:14 crc kubenswrapper[4955]: I0217 13:21:14.182609 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/fb6d9434-4476-438b-90e4-4781bd85b0bb-cache\") pod \"swift-storage-0\" (UID: \"fb6d9434-4476-438b-90e4-4781bd85b0bb\") " pod="openstack/swift-storage-0" Feb 17 13:21:14 crc kubenswrapper[4955]: I0217 13:21:14.182730 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/fb6d9434-4476-438b-90e4-4781bd85b0bb-lock\") pod \"swift-storage-0\" (UID: \"fb6d9434-4476-438b-90e4-4781bd85b0bb\") " pod="openstack/swift-storage-0" Feb 17 13:21:14 crc kubenswrapper[4955]: I0217 13:21:14.182748 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zv4d8\" (UniqueName: \"kubernetes.io/projected/fb6d9434-4476-438b-90e4-4781bd85b0bb-kube-api-access-zv4d8\") pod \"swift-storage-0\" (UID: \"fb6d9434-4476-438b-90e4-4781bd85b0bb\") " pod="openstack/swift-storage-0" Feb 17 13:21:14 crc kubenswrapper[4955]: I0217 13:21:14.182774 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb6d9434-4476-438b-90e4-4781bd85b0bb-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"fb6d9434-4476-438b-90e4-4781bd85b0bb\") " pod="openstack/swift-storage-0" Feb 17 13:21:14 crc kubenswrapper[4955]: I0217 13:21:14.182811 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"fb6d9434-4476-438b-90e4-4781bd85b0bb\") " pod="openstack/swift-storage-0" Feb 17 13:21:14 crc kubenswrapper[4955]: I0217 13:21:14.182844 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/fb6d9434-4476-438b-90e4-4781bd85b0bb-etc-swift\") pod \"swift-storage-0\" (UID: \"fb6d9434-4476-438b-90e4-4781bd85b0bb\") " pod="openstack/swift-storage-0" Feb 17 13:21:14 crc kubenswrapper[4955]: I0217 13:21:14.183499 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/fb6d9434-4476-438b-90e4-4781bd85b0bb-cache\") pod \"swift-storage-0\" (UID: \"fb6d9434-4476-438b-90e4-4781bd85b0bb\") " pod="openstack/swift-storage-0" Feb 17 13:21:14 crc kubenswrapper[4955]: I0217 13:21:14.184628 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/fb6d9434-4476-438b-90e4-4781bd85b0bb-lock\") pod \"swift-storage-0\" (UID: \"fb6d9434-4476-438b-90e4-4781bd85b0bb\") " pod="openstack/swift-storage-0" Feb 17 13:21:14 crc kubenswrapper[4955]: E0217 13:21:14.185329 4955 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 17 13:21:14 crc kubenswrapper[4955]: E0217 13:21:14.185382 4955 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 17 13:21:14 crc kubenswrapper[4955]: I0217 13:21:14.185448 4955 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"fb6d9434-4476-438b-90e4-4781bd85b0bb\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/swift-storage-0" Feb 17 13:21:14 crc kubenswrapper[4955]: E0217 13:21:14.185459 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/fb6d9434-4476-438b-90e4-4781bd85b0bb-etc-swift podName:fb6d9434-4476-438b-90e4-4781bd85b0bb nodeName:}" failed. No retries permitted until 2026-02-17 13:21:14.685422081 +0000 UTC m=+1013.208151624 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/fb6d9434-4476-438b-90e4-4781bd85b0bb-etc-swift") pod "swift-storage-0" (UID: "fb6d9434-4476-438b-90e4-4781bd85b0bb") : configmap "swift-ring-files" not found Feb 17 13:21:14 crc kubenswrapper[4955]: I0217 13:21:14.193368 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb6d9434-4476-438b-90e4-4781bd85b0bb-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"fb6d9434-4476-438b-90e4-4781bd85b0bb\") " pod="openstack/swift-storage-0" Feb 17 13:21:14 crc kubenswrapper[4955]: I0217 13:21:14.208441 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zv4d8\" (UniqueName: \"kubernetes.io/projected/fb6d9434-4476-438b-90e4-4781bd85b0bb-kube-api-access-zv4d8\") pod \"swift-storage-0\" (UID: \"fb6d9434-4476-438b-90e4-4781bd85b0bb\") " pod="openstack/swift-storage-0" Feb 17 13:21:14 crc kubenswrapper[4955]: I0217 13:21:14.212299 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"fb6d9434-4476-438b-90e4-4781bd85b0bb\") " pod="openstack/swift-storage-0" Feb 17 13:21:14 crc kubenswrapper[4955]: I0217 13:21:14.689333 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/fb6d9434-4476-438b-90e4-4781bd85b0bb-etc-swift\") pod \"swift-storage-0\" (UID: \"fb6d9434-4476-438b-90e4-4781bd85b0bb\") " pod="openstack/swift-storage-0" Feb 17 13:21:14 crc kubenswrapper[4955]: E0217 13:21:14.689581 4955 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 17 13:21:14 crc kubenswrapper[4955]: E0217 13:21:14.689753 4955 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 17 13:21:14 crc kubenswrapper[4955]: E0217 13:21:14.689830 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/fb6d9434-4476-438b-90e4-4781bd85b0bb-etc-swift podName:fb6d9434-4476-438b-90e4-4781bd85b0bb nodeName:}" failed. No retries permitted until 2026-02-17 13:21:15.689809579 +0000 UTC m=+1014.212539132 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/fb6d9434-4476-438b-90e4-4781bd85b0bb-etc-swift") pod "swift-storage-0" (UID: "fb6d9434-4476-438b-90e4-4781bd85b0bb") : configmap "swift-ring-files" not found Feb 17 13:21:15 crc kubenswrapper[4955]: I0217 13:21:15.142076 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-dwwb9" event={"ID":"61f387f2-8694-4f14-a7a7-f996a7ed8efe","Type":"ContainerStarted","Data":"aee79c390be2fc0b558e8b284c1bbef59caa978e0391cb988dc7b7af84f05fe1"} Feb 17 13:21:15 crc kubenswrapper[4955]: I0217 13:21:15.142400 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-dwwb9" Feb 17 13:21:15 crc kubenswrapper[4955]: I0217 13:21:15.170503 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-dwwb9" podStartSLOduration=3.170483969 podStartE2EDuration="3.170483969s" podCreationTimestamp="2026-02-17 13:21:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:21:15.161605886 +0000 UTC m=+1013.684335439" watchObservedRunningTime="2026-02-17 13:21:15.170483969 +0000 UTC m=+1013.693213512" Feb 17 13:21:15 crc kubenswrapper[4955]: I0217 13:21:15.703350 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/fb6d9434-4476-438b-90e4-4781bd85b0bb-etc-swift\") pod \"swift-storage-0\" (UID: \"fb6d9434-4476-438b-90e4-4781bd85b0bb\") " pod="openstack/swift-storage-0" Feb 17 13:21:15 crc kubenswrapper[4955]: E0217 13:21:15.703577 4955 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 17 13:21:15 crc kubenswrapper[4955]: E0217 13:21:15.703740 4955 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 17 13:21:15 crc kubenswrapper[4955]: E0217 13:21:15.703827 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/fb6d9434-4476-438b-90e4-4781bd85b0bb-etc-swift podName:fb6d9434-4476-438b-90e4-4781bd85b0bb nodeName:}" failed. No retries permitted until 2026-02-17 13:21:17.703799032 +0000 UTC m=+1016.226528605 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/fb6d9434-4476-438b-90e4-4781bd85b0bb-etc-swift") pod "swift-storage-0" (UID: "fb6d9434-4476-438b-90e4-4781bd85b0bb") : configmap "swift-ring-files" not found Feb 17 13:21:16 crc kubenswrapper[4955]: I0217 13:21:16.146231 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Feb 17 13:21:16 crc kubenswrapper[4955]: I0217 13:21:16.243674 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Feb 17 13:21:17 crc kubenswrapper[4955]: I0217 13:21:17.759620 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/fb6d9434-4476-438b-90e4-4781bd85b0bb-etc-swift\") pod \"swift-storage-0\" (UID: \"fb6d9434-4476-438b-90e4-4781bd85b0bb\") " pod="openstack/swift-storage-0" Feb 17 13:21:17 crc kubenswrapper[4955]: E0217 13:21:17.759871 4955 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 17 13:21:17 crc kubenswrapper[4955]: E0217 13:21:17.760025 4955 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 17 13:21:17 crc kubenswrapper[4955]: E0217 13:21:17.760082 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/fb6d9434-4476-438b-90e4-4781bd85b0bb-etc-swift podName:fb6d9434-4476-438b-90e4-4781bd85b0bb nodeName:}" failed. No retries permitted until 2026-02-17 13:21:21.760065566 +0000 UTC m=+1020.282795109 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/fb6d9434-4476-438b-90e4-4781bd85b0bb-etc-swift") pod "swift-storage-0" (UID: "fb6d9434-4476-438b-90e4-4781bd85b0bb") : configmap "swift-ring-files" not found Feb 17 13:21:17 crc kubenswrapper[4955]: I0217 13:21:17.876560 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-56p7b"] Feb 17 13:21:17 crc kubenswrapper[4955]: I0217 13:21:17.877702 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-56p7b" Feb 17 13:21:17 crc kubenswrapper[4955]: I0217 13:21:17.880135 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Feb 17 13:21:17 crc kubenswrapper[4955]: I0217 13:21:17.880506 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Feb 17 13:21:17 crc kubenswrapper[4955]: I0217 13:21:17.880521 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Feb 17 13:21:17 crc kubenswrapper[4955]: I0217 13:21:17.896598 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-56p7b"] Feb 17 13:21:18 crc kubenswrapper[4955]: I0217 13:21:18.064172 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46beefa1-21de-4818-a70f-e9f840a920ec-combined-ca-bundle\") pod \"swift-ring-rebalance-56p7b\" (UID: \"46beefa1-21de-4818-a70f-e9f840a920ec\") " pod="openstack/swift-ring-rebalance-56p7b" Feb 17 13:21:18 crc kubenswrapper[4955]: I0217 13:21:18.064231 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/46beefa1-21de-4818-a70f-e9f840a920ec-scripts\") pod \"swift-ring-rebalance-56p7b\" (UID: \"46beefa1-21de-4818-a70f-e9f840a920ec\") " pod="openstack/swift-ring-rebalance-56p7b" Feb 17 13:21:18 crc kubenswrapper[4955]: I0217 13:21:18.064272 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/46beefa1-21de-4818-a70f-e9f840a920ec-ring-data-devices\") pod \"swift-ring-rebalance-56p7b\" (UID: \"46beefa1-21de-4818-a70f-e9f840a920ec\") " pod="openstack/swift-ring-rebalance-56p7b" Feb 17 13:21:18 crc kubenswrapper[4955]: I0217 13:21:18.064408 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sz7pl\" (UniqueName: \"kubernetes.io/projected/46beefa1-21de-4818-a70f-e9f840a920ec-kube-api-access-sz7pl\") pod \"swift-ring-rebalance-56p7b\" (UID: \"46beefa1-21de-4818-a70f-e9f840a920ec\") " pod="openstack/swift-ring-rebalance-56p7b" Feb 17 13:21:18 crc kubenswrapper[4955]: I0217 13:21:18.064525 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/46beefa1-21de-4818-a70f-e9f840a920ec-etc-swift\") pod \"swift-ring-rebalance-56p7b\" (UID: \"46beefa1-21de-4818-a70f-e9f840a920ec\") " pod="openstack/swift-ring-rebalance-56p7b" Feb 17 13:21:18 crc kubenswrapper[4955]: I0217 13:21:18.064616 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/46beefa1-21de-4818-a70f-e9f840a920ec-dispersionconf\") pod \"swift-ring-rebalance-56p7b\" (UID: \"46beefa1-21de-4818-a70f-e9f840a920ec\") " pod="openstack/swift-ring-rebalance-56p7b" Feb 17 13:21:18 crc kubenswrapper[4955]: I0217 13:21:18.064834 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/46beefa1-21de-4818-a70f-e9f840a920ec-swiftconf\") pod \"swift-ring-rebalance-56p7b\" (UID: \"46beefa1-21de-4818-a70f-e9f840a920ec\") " pod="openstack/swift-ring-rebalance-56p7b" Feb 17 13:21:18 crc kubenswrapper[4955]: I0217 13:21:18.166335 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/46beefa1-21de-4818-a70f-e9f840a920ec-ring-data-devices\") pod \"swift-ring-rebalance-56p7b\" (UID: \"46beefa1-21de-4818-a70f-e9f840a920ec\") " pod="openstack/swift-ring-rebalance-56p7b" Feb 17 13:21:18 crc kubenswrapper[4955]: I0217 13:21:18.166394 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sz7pl\" (UniqueName: \"kubernetes.io/projected/46beefa1-21de-4818-a70f-e9f840a920ec-kube-api-access-sz7pl\") pod \"swift-ring-rebalance-56p7b\" (UID: \"46beefa1-21de-4818-a70f-e9f840a920ec\") " pod="openstack/swift-ring-rebalance-56p7b" Feb 17 13:21:18 crc kubenswrapper[4955]: I0217 13:21:18.166431 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/46beefa1-21de-4818-a70f-e9f840a920ec-etc-swift\") pod \"swift-ring-rebalance-56p7b\" (UID: \"46beefa1-21de-4818-a70f-e9f840a920ec\") " pod="openstack/swift-ring-rebalance-56p7b" Feb 17 13:21:18 crc kubenswrapper[4955]: I0217 13:21:18.166467 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/46beefa1-21de-4818-a70f-e9f840a920ec-dispersionconf\") pod \"swift-ring-rebalance-56p7b\" (UID: \"46beefa1-21de-4818-a70f-e9f840a920ec\") " pod="openstack/swift-ring-rebalance-56p7b" Feb 17 13:21:18 crc kubenswrapper[4955]: I0217 13:21:18.166547 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/46beefa1-21de-4818-a70f-e9f840a920ec-swiftconf\") pod \"swift-ring-rebalance-56p7b\" (UID: \"46beefa1-21de-4818-a70f-e9f840a920ec\") " pod="openstack/swift-ring-rebalance-56p7b" Feb 17 13:21:18 crc kubenswrapper[4955]: I0217 13:21:18.166574 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46beefa1-21de-4818-a70f-e9f840a920ec-combined-ca-bundle\") pod \"swift-ring-rebalance-56p7b\" (UID: \"46beefa1-21de-4818-a70f-e9f840a920ec\") " pod="openstack/swift-ring-rebalance-56p7b" Feb 17 13:21:18 crc kubenswrapper[4955]: I0217 13:21:18.166591 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/46beefa1-21de-4818-a70f-e9f840a920ec-scripts\") pod \"swift-ring-rebalance-56p7b\" (UID: \"46beefa1-21de-4818-a70f-e9f840a920ec\") " pod="openstack/swift-ring-rebalance-56p7b" Feb 17 13:21:18 crc kubenswrapper[4955]: I0217 13:21:18.167344 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/46beefa1-21de-4818-a70f-e9f840a920ec-scripts\") pod \"swift-ring-rebalance-56p7b\" (UID: \"46beefa1-21de-4818-a70f-e9f840a920ec\") " pod="openstack/swift-ring-rebalance-56p7b" Feb 17 13:21:18 crc kubenswrapper[4955]: I0217 13:21:18.167914 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/46beefa1-21de-4818-a70f-e9f840a920ec-etc-swift\") pod \"swift-ring-rebalance-56p7b\" (UID: \"46beefa1-21de-4818-a70f-e9f840a920ec\") " pod="openstack/swift-ring-rebalance-56p7b" Feb 17 13:21:18 crc kubenswrapper[4955]: I0217 13:21:18.168262 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/46beefa1-21de-4818-a70f-e9f840a920ec-ring-data-devices\") pod \"swift-ring-rebalance-56p7b\" (UID: \"46beefa1-21de-4818-a70f-e9f840a920ec\") " pod="openstack/swift-ring-rebalance-56p7b" Feb 17 13:21:18 crc kubenswrapper[4955]: I0217 13:21:18.175406 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/46beefa1-21de-4818-a70f-e9f840a920ec-swiftconf\") pod \"swift-ring-rebalance-56p7b\" (UID: \"46beefa1-21de-4818-a70f-e9f840a920ec\") " pod="openstack/swift-ring-rebalance-56p7b" Feb 17 13:21:18 crc kubenswrapper[4955]: I0217 13:21:18.175591 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/46beefa1-21de-4818-a70f-e9f840a920ec-dispersionconf\") pod \"swift-ring-rebalance-56p7b\" (UID: \"46beefa1-21de-4818-a70f-e9f840a920ec\") " pod="openstack/swift-ring-rebalance-56p7b" Feb 17 13:21:18 crc kubenswrapper[4955]: I0217 13:21:18.175988 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46beefa1-21de-4818-a70f-e9f840a920ec-combined-ca-bundle\") pod \"swift-ring-rebalance-56p7b\" (UID: \"46beefa1-21de-4818-a70f-e9f840a920ec\") " pod="openstack/swift-ring-rebalance-56p7b" Feb 17 13:21:18 crc kubenswrapper[4955]: I0217 13:21:18.188626 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sz7pl\" (UniqueName: \"kubernetes.io/projected/46beefa1-21de-4818-a70f-e9f840a920ec-kube-api-access-sz7pl\") pod \"swift-ring-rebalance-56p7b\" (UID: \"46beefa1-21de-4818-a70f-e9f840a920ec\") " pod="openstack/swift-ring-rebalance-56p7b" Feb 17 13:21:18 crc kubenswrapper[4955]: I0217 13:21:18.194578 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-56p7b" Feb 17 13:21:18 crc kubenswrapper[4955]: I0217 13:21:18.654808 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-56p7b"] Feb 17 13:21:18 crc kubenswrapper[4955]: I0217 13:21:18.782096 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-cqgbw"] Feb 17 13:21:18 crc kubenswrapper[4955]: I0217 13:21:18.783306 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-cqgbw" Feb 17 13:21:18 crc kubenswrapper[4955]: I0217 13:21:18.784882 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Feb 17 13:21:18 crc kubenswrapper[4955]: I0217 13:21:18.793531 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-cqgbw"] Feb 17 13:21:18 crc kubenswrapper[4955]: I0217 13:21:18.877274 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5fe5a03-2bb2-450f-8258-7cf8ad985475-operator-scripts\") pod \"root-account-create-update-cqgbw\" (UID: \"e5fe5a03-2bb2-450f-8258-7cf8ad985475\") " pod="openstack/root-account-create-update-cqgbw" Feb 17 13:21:18 crc kubenswrapper[4955]: I0217 13:21:18.877368 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpf98\" (UniqueName: \"kubernetes.io/projected/e5fe5a03-2bb2-450f-8258-7cf8ad985475-kube-api-access-lpf98\") pod \"root-account-create-update-cqgbw\" (UID: \"e5fe5a03-2bb2-450f-8258-7cf8ad985475\") " pod="openstack/root-account-create-update-cqgbw" Feb 17 13:21:18 crc kubenswrapper[4955]: I0217 13:21:18.962501 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Feb 17 13:21:18 crc kubenswrapper[4955]: I0217 13:21:18.962554 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Feb 17 13:21:18 crc kubenswrapper[4955]: I0217 13:21:18.978313 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpf98\" (UniqueName: \"kubernetes.io/projected/e5fe5a03-2bb2-450f-8258-7cf8ad985475-kube-api-access-lpf98\") pod \"root-account-create-update-cqgbw\" (UID: \"e5fe5a03-2bb2-450f-8258-7cf8ad985475\") " pod="openstack/root-account-create-update-cqgbw" Feb 17 13:21:18 crc kubenswrapper[4955]: I0217 13:21:18.978452 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5fe5a03-2bb2-450f-8258-7cf8ad985475-operator-scripts\") pod \"root-account-create-update-cqgbw\" (UID: \"e5fe5a03-2bb2-450f-8258-7cf8ad985475\") " pod="openstack/root-account-create-update-cqgbw" Feb 17 13:21:18 crc kubenswrapper[4955]: I0217 13:21:18.979236 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5fe5a03-2bb2-450f-8258-7cf8ad985475-operator-scripts\") pod \"root-account-create-update-cqgbw\" (UID: \"e5fe5a03-2bb2-450f-8258-7cf8ad985475\") " pod="openstack/root-account-create-update-cqgbw" Feb 17 13:21:18 crc kubenswrapper[4955]: I0217 13:21:18.998779 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpf98\" (UniqueName: \"kubernetes.io/projected/e5fe5a03-2bb2-450f-8258-7cf8ad985475-kube-api-access-lpf98\") pod \"root-account-create-update-cqgbw\" (UID: \"e5fe5a03-2bb2-450f-8258-7cf8ad985475\") " pod="openstack/root-account-create-update-cqgbw" Feb 17 13:21:19 crc kubenswrapper[4955]: I0217 13:21:19.047719 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Feb 17 13:21:19 crc kubenswrapper[4955]: I0217 13:21:19.109045 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-cqgbw" Feb 17 13:21:19 crc kubenswrapper[4955]: I0217 13:21:19.169875 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-56p7b" event={"ID":"46beefa1-21de-4818-a70f-e9f840a920ec","Type":"ContainerStarted","Data":"4c5bb6dc3c345b1a51f563296d4c182af165466f26f7a79b1b3e017cba8194aa"} Feb 17 13:21:19 crc kubenswrapper[4955]: I0217 13:21:19.255121 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Feb 17 13:21:19 crc kubenswrapper[4955]: I0217 13:21:19.575321 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-cqgbw"] Feb 17 13:21:19 crc kubenswrapper[4955]: W0217 13:21:19.576863 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5fe5a03_2bb2_450f_8258_7cf8ad985475.slice/crio-ac2c649738823cbb12b3590ef249623edd73791ef58a55ebc17f0c127643a6cb WatchSource:0}: Error finding container ac2c649738823cbb12b3590ef249623edd73791ef58a55ebc17f0c127643a6cb: Status 404 returned error can't find the container with id ac2c649738823cbb12b3590ef249623edd73791ef58a55ebc17f0c127643a6cb Feb 17 13:21:20 crc kubenswrapper[4955]: I0217 13:21:20.180477 4955 generic.go:334] "Generic (PLEG): container finished" podID="e5fe5a03-2bb2-450f-8258-7cf8ad985475" containerID="a91687d4b5b34d2b759f39ea553342f9cc3938a66b65c3b62191bbffbeebb176" exitCode=0 Feb 17 13:21:20 crc kubenswrapper[4955]: I0217 13:21:20.180535 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-cqgbw" event={"ID":"e5fe5a03-2bb2-450f-8258-7cf8ad985475","Type":"ContainerDied","Data":"a91687d4b5b34d2b759f39ea553342f9cc3938a66b65c3b62191bbffbeebb176"} Feb 17 13:21:20 crc kubenswrapper[4955]: I0217 13:21:20.180920 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-cqgbw" event={"ID":"e5fe5a03-2bb2-450f-8258-7cf8ad985475","Type":"ContainerStarted","Data":"ac2c649738823cbb12b3590ef249623edd73791ef58a55ebc17f0c127643a6cb"} Feb 17 13:21:20 crc kubenswrapper[4955]: I0217 13:21:20.559564 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-hvwtj"] Feb 17 13:21:20 crc kubenswrapper[4955]: I0217 13:21:20.562189 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-hvwtj" Feb 17 13:21:20 crc kubenswrapper[4955]: I0217 13:21:20.568586 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-hvwtj"] Feb 17 13:21:20 crc kubenswrapper[4955]: I0217 13:21:20.690256 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-be28-account-create-update-xx2j6"] Feb 17 13:21:20 crc kubenswrapper[4955]: I0217 13:21:20.691594 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-be28-account-create-update-xx2j6" Feb 17 13:21:20 crc kubenswrapper[4955]: I0217 13:21:20.697967 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Feb 17 13:21:20 crc kubenswrapper[4955]: I0217 13:21:20.716871 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-be28-account-create-update-xx2j6"] Feb 17 13:21:20 crc kubenswrapper[4955]: I0217 13:21:20.725015 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/966dab54-42a6-452a-8459-e76ec889c9ee-operator-scripts\") pod \"glance-db-create-hvwtj\" (UID: \"966dab54-42a6-452a-8459-e76ec889c9ee\") " pod="openstack/glance-db-create-hvwtj" Feb 17 13:21:20 crc kubenswrapper[4955]: I0217 13:21:20.725160 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thppx\" (UniqueName: \"kubernetes.io/projected/966dab54-42a6-452a-8459-e76ec889c9ee-kube-api-access-thppx\") pod \"glance-db-create-hvwtj\" (UID: \"966dab54-42a6-452a-8459-e76ec889c9ee\") " pod="openstack/glance-db-create-hvwtj" Feb 17 13:21:20 crc kubenswrapper[4955]: I0217 13:21:20.827260 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4f5tf\" (UniqueName: \"kubernetes.io/projected/7a2f748d-5c00-4975-8888-cdea65d3ab36-kube-api-access-4f5tf\") pod \"glance-be28-account-create-update-xx2j6\" (UID: \"7a2f748d-5c00-4975-8888-cdea65d3ab36\") " pod="openstack/glance-be28-account-create-update-xx2j6" Feb 17 13:21:20 crc kubenswrapper[4955]: I0217 13:21:20.827369 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thppx\" (UniqueName: \"kubernetes.io/projected/966dab54-42a6-452a-8459-e76ec889c9ee-kube-api-access-thppx\") pod \"glance-db-create-hvwtj\" (UID: \"966dab54-42a6-452a-8459-e76ec889c9ee\") " pod="openstack/glance-db-create-hvwtj" Feb 17 13:21:20 crc kubenswrapper[4955]: I0217 13:21:20.827426 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/966dab54-42a6-452a-8459-e76ec889c9ee-operator-scripts\") pod \"glance-db-create-hvwtj\" (UID: \"966dab54-42a6-452a-8459-e76ec889c9ee\") " pod="openstack/glance-db-create-hvwtj" Feb 17 13:21:20 crc kubenswrapper[4955]: I0217 13:21:20.827511 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7a2f748d-5c00-4975-8888-cdea65d3ab36-operator-scripts\") pod \"glance-be28-account-create-update-xx2j6\" (UID: \"7a2f748d-5c00-4975-8888-cdea65d3ab36\") " pod="openstack/glance-be28-account-create-update-xx2j6" Feb 17 13:21:20 crc kubenswrapper[4955]: I0217 13:21:20.829766 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/966dab54-42a6-452a-8459-e76ec889c9ee-operator-scripts\") pod \"glance-db-create-hvwtj\" (UID: \"966dab54-42a6-452a-8459-e76ec889c9ee\") " pod="openstack/glance-db-create-hvwtj" Feb 17 13:21:20 crc kubenswrapper[4955]: I0217 13:21:20.851398 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thppx\" (UniqueName: \"kubernetes.io/projected/966dab54-42a6-452a-8459-e76ec889c9ee-kube-api-access-thppx\") pod \"glance-db-create-hvwtj\" (UID: \"966dab54-42a6-452a-8459-e76ec889c9ee\") " pod="openstack/glance-db-create-hvwtj" Feb 17 13:21:20 crc kubenswrapper[4955]: I0217 13:21:20.928764 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7a2f748d-5c00-4975-8888-cdea65d3ab36-operator-scripts\") pod \"glance-be28-account-create-update-xx2j6\" (UID: \"7a2f748d-5c00-4975-8888-cdea65d3ab36\") " pod="openstack/glance-be28-account-create-update-xx2j6" Feb 17 13:21:20 crc kubenswrapper[4955]: I0217 13:21:20.928856 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4f5tf\" (UniqueName: \"kubernetes.io/projected/7a2f748d-5c00-4975-8888-cdea65d3ab36-kube-api-access-4f5tf\") pod \"glance-be28-account-create-update-xx2j6\" (UID: \"7a2f748d-5c00-4975-8888-cdea65d3ab36\") " pod="openstack/glance-be28-account-create-update-xx2j6" Feb 17 13:21:20 crc kubenswrapper[4955]: I0217 13:21:20.930049 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7a2f748d-5c00-4975-8888-cdea65d3ab36-operator-scripts\") pod \"glance-be28-account-create-update-xx2j6\" (UID: \"7a2f748d-5c00-4975-8888-cdea65d3ab36\") " pod="openstack/glance-be28-account-create-update-xx2j6" Feb 17 13:21:20 crc kubenswrapper[4955]: I0217 13:21:20.944966 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4f5tf\" (UniqueName: \"kubernetes.io/projected/7a2f748d-5c00-4975-8888-cdea65d3ab36-kube-api-access-4f5tf\") pod \"glance-be28-account-create-update-xx2j6\" (UID: \"7a2f748d-5c00-4975-8888-cdea65d3ab36\") " pod="openstack/glance-be28-account-create-update-xx2j6" Feb 17 13:21:20 crc kubenswrapper[4955]: I0217 13:21:20.951826 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-hvwtj" Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.005622 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-be28-account-create-update-xx2j6" Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.306708 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-nbjcj"] Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.307917 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-nbjcj" Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.318893 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-nbjcj"] Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.377753 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-9f71-account-create-update-fhns5"] Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.379353 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-9f71-account-create-update-fhns5" Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.382151 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.398512 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-9f71-account-create-update-fhns5"] Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.439061 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmnhn\" (UniqueName: \"kubernetes.io/projected/96770935-7ba8-440a-8947-28a74cd86f08-kube-api-access-xmnhn\") pod \"keystone-db-create-nbjcj\" (UID: \"96770935-7ba8-440a-8947-28a74cd86f08\") " pod="openstack/keystone-db-create-nbjcj" Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.439281 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/96770935-7ba8-440a-8947-28a74cd86f08-operator-scripts\") pod \"keystone-db-create-nbjcj\" (UID: \"96770935-7ba8-440a-8947-28a74cd86f08\") " pod="openstack/keystone-db-create-nbjcj" Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.540862 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmnhn\" (UniqueName: \"kubernetes.io/projected/96770935-7ba8-440a-8947-28a74cd86f08-kube-api-access-xmnhn\") pod \"keystone-db-create-nbjcj\" (UID: \"96770935-7ba8-440a-8947-28a74cd86f08\") " pod="openstack/keystone-db-create-nbjcj" Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.540963 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/19e2a114-5254-4281-9d4d-209c26dcb1d6-operator-scripts\") pod \"keystone-9f71-account-create-update-fhns5\" (UID: \"19e2a114-5254-4281-9d4d-209c26dcb1d6\") " pod="openstack/keystone-9f71-account-create-update-fhns5" Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.541035 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/96770935-7ba8-440a-8947-28a74cd86f08-operator-scripts\") pod \"keystone-db-create-nbjcj\" (UID: \"96770935-7ba8-440a-8947-28a74cd86f08\") " pod="openstack/keystone-db-create-nbjcj" Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.541072 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbb65\" (UniqueName: \"kubernetes.io/projected/19e2a114-5254-4281-9d4d-209c26dcb1d6-kube-api-access-dbb65\") pod \"keystone-9f71-account-create-update-fhns5\" (UID: \"19e2a114-5254-4281-9d4d-209c26dcb1d6\") " pod="openstack/keystone-9f71-account-create-update-fhns5" Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.546008 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/96770935-7ba8-440a-8947-28a74cd86f08-operator-scripts\") pod \"keystone-db-create-nbjcj\" (UID: \"96770935-7ba8-440a-8947-28a74cd86f08\") " pod="openstack/keystone-db-create-nbjcj" Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.559824 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-7gmhj"] Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.561072 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-7gmhj" Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.575835 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmnhn\" (UniqueName: \"kubernetes.io/projected/96770935-7ba8-440a-8947-28a74cd86f08-kube-api-access-xmnhn\") pod \"keystone-db-create-nbjcj\" (UID: \"96770935-7ba8-440a-8947-28a74cd86f08\") " pod="openstack/keystone-db-create-nbjcj" Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.584278 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-7gmhj"] Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.639653 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-nbjcj" Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.644027 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7g6cd\" (UniqueName: \"kubernetes.io/projected/2d8307b9-b38d-44ed-830f-c0cb10538496-kube-api-access-7g6cd\") pod \"placement-db-create-7gmhj\" (UID: \"2d8307b9-b38d-44ed-830f-c0cb10538496\") " pod="openstack/placement-db-create-7gmhj" Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.644108 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbb65\" (UniqueName: \"kubernetes.io/projected/19e2a114-5254-4281-9d4d-209c26dcb1d6-kube-api-access-dbb65\") pod \"keystone-9f71-account-create-update-fhns5\" (UID: \"19e2a114-5254-4281-9d4d-209c26dcb1d6\") " pod="openstack/keystone-9f71-account-create-update-fhns5" Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.644211 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d8307b9-b38d-44ed-830f-c0cb10538496-operator-scripts\") pod \"placement-db-create-7gmhj\" (UID: \"2d8307b9-b38d-44ed-830f-c0cb10538496\") " pod="openstack/placement-db-create-7gmhj" Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.644417 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/19e2a114-5254-4281-9d4d-209c26dcb1d6-operator-scripts\") pod \"keystone-9f71-account-create-update-fhns5\" (UID: \"19e2a114-5254-4281-9d4d-209c26dcb1d6\") " pod="openstack/keystone-9f71-account-create-update-fhns5" Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.645665 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/19e2a114-5254-4281-9d4d-209c26dcb1d6-operator-scripts\") pod \"keystone-9f71-account-create-update-fhns5\" (UID: \"19e2a114-5254-4281-9d4d-209c26dcb1d6\") " pod="openstack/keystone-9f71-account-create-update-fhns5" Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.677496 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-509d-account-create-update-pfzj7"] Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.679480 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-509d-account-create-update-pfzj7" Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.684013 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbb65\" (UniqueName: \"kubernetes.io/projected/19e2a114-5254-4281-9d4d-209c26dcb1d6-kube-api-access-dbb65\") pod \"keystone-9f71-account-create-update-fhns5\" (UID: \"19e2a114-5254-4281-9d4d-209c26dcb1d6\") " pod="openstack/keystone-9f71-account-create-update-fhns5" Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.684207 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.698583 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-509d-account-create-update-pfzj7"] Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.707381 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-9f71-account-create-update-fhns5" Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.746459 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7g6cd\" (UniqueName: \"kubernetes.io/projected/2d8307b9-b38d-44ed-830f-c0cb10538496-kube-api-access-7g6cd\") pod \"placement-db-create-7gmhj\" (UID: \"2d8307b9-b38d-44ed-830f-c0cb10538496\") " pod="openstack/placement-db-create-7gmhj" Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.746567 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d8307b9-b38d-44ed-830f-c0cb10538496-operator-scripts\") pod \"placement-db-create-7gmhj\" (UID: \"2d8307b9-b38d-44ed-830f-c0cb10538496\") " pod="openstack/placement-db-create-7gmhj" Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.747302 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d8307b9-b38d-44ed-830f-c0cb10538496-operator-scripts\") pod \"placement-db-create-7gmhj\" (UID: \"2d8307b9-b38d-44ed-830f-c0cb10538496\") " pod="openstack/placement-db-create-7gmhj" Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.762332 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7g6cd\" (UniqueName: \"kubernetes.io/projected/2d8307b9-b38d-44ed-830f-c0cb10538496-kube-api-access-7g6cd\") pod \"placement-db-create-7gmhj\" (UID: \"2d8307b9-b38d-44ed-830f-c0cb10538496\") " pod="openstack/placement-db-create-7gmhj" Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.848604 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wkpl\" (UniqueName: \"kubernetes.io/projected/f54e2d0a-6914-42a1-a9d0-f8c6c5e207b0-kube-api-access-2wkpl\") pod \"placement-509d-account-create-update-pfzj7\" (UID: \"f54e2d0a-6914-42a1-a9d0-f8c6c5e207b0\") " pod="openstack/placement-509d-account-create-update-pfzj7" Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.848764 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f54e2d0a-6914-42a1-a9d0-f8c6c5e207b0-operator-scripts\") pod \"placement-509d-account-create-update-pfzj7\" (UID: \"f54e2d0a-6914-42a1-a9d0-f8c6c5e207b0\") " pod="openstack/placement-509d-account-create-update-pfzj7" Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.848834 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/fb6d9434-4476-438b-90e4-4781bd85b0bb-etc-swift\") pod \"swift-storage-0\" (UID: \"fb6d9434-4476-438b-90e4-4781bd85b0bb\") " pod="openstack/swift-storage-0" Feb 17 13:21:21 crc kubenswrapper[4955]: E0217 13:21:21.849054 4955 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 17 13:21:21 crc kubenswrapper[4955]: E0217 13:21:21.849076 4955 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 17 13:21:21 crc kubenswrapper[4955]: E0217 13:21:21.849125 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/fb6d9434-4476-438b-90e4-4781bd85b0bb-etc-swift podName:fb6d9434-4476-438b-90e4-4781bd85b0bb nodeName:}" failed. No retries permitted until 2026-02-17 13:21:29.849107934 +0000 UTC m=+1028.371837477 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/fb6d9434-4476-438b-90e4-4781bd85b0bb-etc-swift") pod "swift-storage-0" (UID: "fb6d9434-4476-438b-90e4-4781bd85b0bb") : configmap "swift-ring-files" not found Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.906867 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-7gmhj" Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.950880 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wkpl\" (UniqueName: \"kubernetes.io/projected/f54e2d0a-6914-42a1-a9d0-f8c6c5e207b0-kube-api-access-2wkpl\") pod \"placement-509d-account-create-update-pfzj7\" (UID: \"f54e2d0a-6914-42a1-a9d0-f8c6c5e207b0\") " pod="openstack/placement-509d-account-create-update-pfzj7" Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.950993 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f54e2d0a-6914-42a1-a9d0-f8c6c5e207b0-operator-scripts\") pod \"placement-509d-account-create-update-pfzj7\" (UID: \"f54e2d0a-6914-42a1-a9d0-f8c6c5e207b0\") " pod="openstack/placement-509d-account-create-update-pfzj7" Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.951807 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f54e2d0a-6914-42a1-a9d0-f8c6c5e207b0-operator-scripts\") pod \"placement-509d-account-create-update-pfzj7\" (UID: \"f54e2d0a-6914-42a1-a9d0-f8c6c5e207b0\") " pod="openstack/placement-509d-account-create-update-pfzj7" Feb 17 13:21:21 crc kubenswrapper[4955]: I0217 13:21:21.969088 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wkpl\" (UniqueName: \"kubernetes.io/projected/f54e2d0a-6914-42a1-a9d0-f8c6c5e207b0-kube-api-access-2wkpl\") pod \"placement-509d-account-create-update-pfzj7\" (UID: \"f54e2d0a-6914-42a1-a9d0-f8c6c5e207b0\") " pod="openstack/placement-509d-account-create-update-pfzj7" Feb 17 13:21:22 crc kubenswrapper[4955]: I0217 13:21:22.019566 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-509d-account-create-update-pfzj7" Feb 17 13:21:22 crc kubenswrapper[4955]: I0217 13:21:22.727753 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-cqgbw" Feb 17 13:21:22 crc kubenswrapper[4955]: I0217 13:21:22.864135 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lpf98\" (UniqueName: \"kubernetes.io/projected/e5fe5a03-2bb2-450f-8258-7cf8ad985475-kube-api-access-lpf98\") pod \"e5fe5a03-2bb2-450f-8258-7cf8ad985475\" (UID: \"e5fe5a03-2bb2-450f-8258-7cf8ad985475\") " Feb 17 13:21:22 crc kubenswrapper[4955]: I0217 13:21:22.864280 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5fe5a03-2bb2-450f-8258-7cf8ad985475-operator-scripts\") pod \"e5fe5a03-2bb2-450f-8258-7cf8ad985475\" (UID: \"e5fe5a03-2bb2-450f-8258-7cf8ad985475\") " Feb 17 13:21:22 crc kubenswrapper[4955]: I0217 13:21:22.864970 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5fe5a03-2bb2-450f-8258-7cf8ad985475-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e5fe5a03-2bb2-450f-8258-7cf8ad985475" (UID: "e5fe5a03-2bb2-450f-8258-7cf8ad985475"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:21:22 crc kubenswrapper[4955]: I0217 13:21:22.868077 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5fe5a03-2bb2-450f-8258-7cf8ad985475-kube-api-access-lpf98" (OuterVolumeSpecName: "kube-api-access-lpf98") pod "e5fe5a03-2bb2-450f-8258-7cf8ad985475" (UID: "e5fe5a03-2bb2-450f-8258-7cf8ad985475"). InnerVolumeSpecName "kube-api-access-lpf98". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:21:22 crc kubenswrapper[4955]: I0217 13:21:22.965934 4955 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5fe5a03-2bb2-450f-8258-7cf8ad985475-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:22 crc kubenswrapper[4955]: I0217 13:21:22.966231 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lpf98\" (UniqueName: \"kubernetes.io/projected/e5fe5a03-2bb2-450f-8258-7cf8ad985475-kube-api-access-lpf98\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:23 crc kubenswrapper[4955]: I0217 13:21:23.207846 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-cqgbw" Feb 17 13:21:23 crc kubenswrapper[4955]: I0217 13:21:23.207868 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-cqgbw" event={"ID":"e5fe5a03-2bb2-450f-8258-7cf8ad985475","Type":"ContainerDied","Data":"ac2c649738823cbb12b3590ef249623edd73791ef58a55ebc17f0c127643a6cb"} Feb 17 13:21:23 crc kubenswrapper[4955]: I0217 13:21:23.208508 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ac2c649738823cbb12b3590ef249623edd73791ef58a55ebc17f0c127643a6cb" Feb 17 13:21:23 crc kubenswrapper[4955]: I0217 13:21:23.208607 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-dwwb9" Feb 17 13:21:23 crc kubenswrapper[4955]: I0217 13:21:23.209464 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-56p7b" event={"ID":"46beefa1-21de-4818-a70f-e9f840a920ec","Type":"ContainerStarted","Data":"5f190e937676a3c7f39f40f4b14cf50b9c4c117f66d82dde7fa030590da7ccef"} Feb 17 13:21:23 crc kubenswrapper[4955]: I0217 13:21:23.253403 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-56p7b" podStartSLOduration=1.999372376 podStartE2EDuration="6.253382589s" podCreationTimestamp="2026-02-17 13:21:17 +0000 UTC" firstStartedPulling="2026-02-17 13:21:18.678317828 +0000 UTC m=+1017.201047371" lastFinishedPulling="2026-02-17 13:21:22.932328041 +0000 UTC m=+1021.455057584" observedRunningTime="2026-02-17 13:21:23.252848134 +0000 UTC m=+1021.775577697" watchObservedRunningTime="2026-02-17 13:21:23.253382589 +0000 UTC m=+1021.776112142" Feb 17 13:21:23 crc kubenswrapper[4955]: I0217 13:21:23.275637 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-jpznm"] Feb 17 13:21:23 crc kubenswrapper[4955]: I0217 13:21:23.275901 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8554648995-jpznm" podUID="dd0771e6-147c-4069-88e1-c9ae122dc55d" containerName="dnsmasq-dns" containerID="cri-o://d0afe951ba8e2e8c737ca5d913a77d6e9b224faca55462a95c93bd441480fded" gracePeriod=10 Feb 17 13:21:23 crc kubenswrapper[4955]: I0217 13:21:23.356899 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-9f71-account-create-update-fhns5"] Feb 17 13:21:23 crc kubenswrapper[4955]: I0217 13:21:23.394496 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Feb 17 13:21:23 crc kubenswrapper[4955]: I0217 13:21:23.394663 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Feb 17 13:21:23 crc kubenswrapper[4955]: I0217 13:21:23.398853 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-509d-account-create-update-pfzj7"] Feb 17 13:21:23 crc kubenswrapper[4955]: I0217 13:21:23.706766 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-nbjcj"] Feb 17 13:21:23 crc kubenswrapper[4955]: I0217 13:21:23.737770 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-hvwtj"] Feb 17 13:21:23 crc kubenswrapper[4955]: I0217 13:21:23.771133 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-7gmhj"] Feb 17 13:21:23 crc kubenswrapper[4955]: W0217 13:21:23.797012 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod966dab54_42a6_452a_8459_e76ec889c9ee.slice/crio-3e0d343a7853a6d70356f65601f0246273fc64f41ebb2ba9fdd2f21aec2d070a WatchSource:0}: Error finding container 3e0d343a7853a6d70356f65601f0246273fc64f41ebb2ba9fdd2f21aec2d070a: Status 404 returned error can't find the container with id 3e0d343a7853a6d70356f65601f0246273fc64f41ebb2ba9fdd2f21aec2d070a Feb 17 13:21:23 crc kubenswrapper[4955]: I0217 13:21:23.800882 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-be28-account-create-update-xx2j6"] Feb 17 13:21:23 crc kubenswrapper[4955]: I0217 13:21:23.818269 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Feb 17 13:21:23 crc kubenswrapper[4955]: I0217 13:21:23.898290 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-jpznm" Feb 17 13:21:24 crc kubenswrapper[4955]: I0217 13:21:24.003433 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd0771e6-147c-4069-88e1-c9ae122dc55d-config\") pod \"dd0771e6-147c-4069-88e1-c9ae122dc55d\" (UID: \"dd0771e6-147c-4069-88e1-c9ae122dc55d\") " Feb 17 13:21:24 crc kubenswrapper[4955]: I0217 13:21:24.003495 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dd0771e6-147c-4069-88e1-c9ae122dc55d-ovsdbserver-sb\") pod \"dd0771e6-147c-4069-88e1-c9ae122dc55d\" (UID: \"dd0771e6-147c-4069-88e1-c9ae122dc55d\") " Feb 17 13:21:24 crc kubenswrapper[4955]: I0217 13:21:24.003516 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dd0771e6-147c-4069-88e1-c9ae122dc55d-ovsdbserver-nb\") pod \"dd0771e6-147c-4069-88e1-c9ae122dc55d\" (UID: \"dd0771e6-147c-4069-88e1-c9ae122dc55d\") " Feb 17 13:21:24 crc kubenswrapper[4955]: I0217 13:21:24.003538 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd0771e6-147c-4069-88e1-c9ae122dc55d-dns-svc\") pod \"dd0771e6-147c-4069-88e1-c9ae122dc55d\" (UID: \"dd0771e6-147c-4069-88e1-c9ae122dc55d\") " Feb 17 13:21:24 crc kubenswrapper[4955]: I0217 13:21:24.003629 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2lknw\" (UniqueName: \"kubernetes.io/projected/dd0771e6-147c-4069-88e1-c9ae122dc55d-kube-api-access-2lknw\") pod \"dd0771e6-147c-4069-88e1-c9ae122dc55d\" (UID: \"dd0771e6-147c-4069-88e1-c9ae122dc55d\") " Feb 17 13:21:24 crc kubenswrapper[4955]: I0217 13:21:24.025791 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd0771e6-147c-4069-88e1-c9ae122dc55d-kube-api-access-2lknw" (OuterVolumeSpecName: "kube-api-access-2lknw") pod "dd0771e6-147c-4069-88e1-c9ae122dc55d" (UID: "dd0771e6-147c-4069-88e1-c9ae122dc55d"). InnerVolumeSpecName "kube-api-access-2lknw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:21:24 crc kubenswrapper[4955]: I0217 13:21:24.105609 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd0771e6-147c-4069-88e1-c9ae122dc55d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "dd0771e6-147c-4069-88e1-c9ae122dc55d" (UID: "dd0771e6-147c-4069-88e1-c9ae122dc55d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:21:24 crc kubenswrapper[4955]: I0217 13:21:24.106679 4955 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dd0771e6-147c-4069-88e1-c9ae122dc55d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:24 crc kubenswrapper[4955]: I0217 13:21:24.106816 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2lknw\" (UniqueName: \"kubernetes.io/projected/dd0771e6-147c-4069-88e1-c9ae122dc55d-kube-api-access-2lknw\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:24 crc kubenswrapper[4955]: I0217 13:21:24.118468 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd0771e6-147c-4069-88e1-c9ae122dc55d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "dd0771e6-147c-4069-88e1-c9ae122dc55d" (UID: "dd0771e6-147c-4069-88e1-c9ae122dc55d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:21:24 crc kubenswrapper[4955]: I0217 13:21:24.120165 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd0771e6-147c-4069-88e1-c9ae122dc55d-config" (OuterVolumeSpecName: "config") pod "dd0771e6-147c-4069-88e1-c9ae122dc55d" (UID: "dd0771e6-147c-4069-88e1-c9ae122dc55d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:21:24 crc kubenswrapper[4955]: I0217 13:21:24.132702 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd0771e6-147c-4069-88e1-c9ae122dc55d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "dd0771e6-147c-4069-88e1-c9ae122dc55d" (UID: "dd0771e6-147c-4069-88e1-c9ae122dc55d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:21:24 crc kubenswrapper[4955]: I0217 13:21:24.208513 4955 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dd0771e6-147c-4069-88e1-c9ae122dc55d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:24 crc kubenswrapper[4955]: I0217 13:21:24.208576 4955 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd0771e6-147c-4069-88e1-c9ae122dc55d-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:24 crc kubenswrapper[4955]: I0217 13:21:24.208592 4955 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd0771e6-147c-4069-88e1-c9ae122dc55d-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:24 crc kubenswrapper[4955]: I0217 13:21:24.223194 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-hvwtj" event={"ID":"966dab54-42a6-452a-8459-e76ec889c9ee","Type":"ContainerStarted","Data":"3e0d343a7853a6d70356f65601f0246273fc64f41ebb2ba9fdd2f21aec2d070a"} Feb 17 13:21:24 crc kubenswrapper[4955]: I0217 13:21:24.225217 4955 generic.go:334] "Generic (PLEG): container finished" podID="dd0771e6-147c-4069-88e1-c9ae122dc55d" containerID="d0afe951ba8e2e8c737ca5d913a77d6e9b224faca55462a95c93bd441480fded" exitCode=0 Feb 17 13:21:24 crc kubenswrapper[4955]: I0217 13:21:24.225348 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-jpznm" Feb 17 13:21:24 crc kubenswrapper[4955]: I0217 13:21:24.233922 4955 generic.go:334] "Generic (PLEG): container finished" podID="f54e2d0a-6914-42a1-a9d0-f8c6c5e207b0" containerID="f5994e3d18d3af598e883e39ab79069f680f8acf585814862c87251b0d202062" exitCode=0 Feb 17 13:21:24 crc kubenswrapper[4955]: I0217 13:21:24.242503 4955 generic.go:334] "Generic (PLEG): container finished" podID="19e2a114-5254-4281-9d4d-209c26dcb1d6" containerID="bb5673a151e591e332e71278606624a994c4a495305921eb69fc3e8fe412fcb5" exitCode=0 Feb 17 13:21:24 crc kubenswrapper[4955]: I0217 13:21:24.248406 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-jpznm" event={"ID":"dd0771e6-147c-4069-88e1-c9ae122dc55d","Type":"ContainerDied","Data":"d0afe951ba8e2e8c737ca5d913a77d6e9b224faca55462a95c93bd441480fded"} Feb 17 13:21:24 crc kubenswrapper[4955]: I0217 13:21:24.248457 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-jpznm" event={"ID":"dd0771e6-147c-4069-88e1-c9ae122dc55d","Type":"ContainerDied","Data":"dd637fd908e2b69a43093810df7623aaaebda64cfccc3a0b1aec87cc9a1af775"} Feb 17 13:21:24 crc kubenswrapper[4955]: I0217 13:21:24.248470 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-509d-account-create-update-pfzj7" event={"ID":"f54e2d0a-6914-42a1-a9d0-f8c6c5e207b0","Type":"ContainerDied","Data":"f5994e3d18d3af598e883e39ab79069f680f8acf585814862c87251b0d202062"} Feb 17 13:21:24 crc kubenswrapper[4955]: I0217 13:21:24.248484 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-509d-account-create-update-pfzj7" event={"ID":"f54e2d0a-6914-42a1-a9d0-f8c6c5e207b0","Type":"ContainerStarted","Data":"6adf9563db64e90e0e5d2ff82bebb8a8a41d3eb8e994a701418a307b7d59f234"} Feb 17 13:21:24 crc kubenswrapper[4955]: I0217 13:21:24.248494 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-be28-account-create-update-xx2j6" event={"ID":"7a2f748d-5c00-4975-8888-cdea65d3ab36","Type":"ContainerStarted","Data":"04c8cf2fb0cdb10eb8ec8fc8920721e50490e64dc925d5eec0b2cad40553a688"} Feb 17 13:21:24 crc kubenswrapper[4955]: I0217 13:21:24.248506 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-be28-account-create-update-xx2j6" event={"ID":"7a2f748d-5c00-4975-8888-cdea65d3ab36","Type":"ContainerStarted","Data":"7138254bf4d126537f29a027ca2d7246fab42e34c840a781d99a1a1e8aec8b61"} Feb 17 13:21:24 crc kubenswrapper[4955]: I0217 13:21:24.248517 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-7gmhj" event={"ID":"2d8307b9-b38d-44ed-830f-c0cb10538496","Type":"ContainerStarted","Data":"05c87d393167e736f7b09dc84dbdd5af5657783de253f846a36d99cbf6457afa"} Feb 17 13:21:24 crc kubenswrapper[4955]: I0217 13:21:24.248531 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-7gmhj" event={"ID":"2d8307b9-b38d-44ed-830f-c0cb10538496","Type":"ContainerStarted","Data":"31650270abaf72b5c7fbdbb4e80d7fd1a358a89f2658c7107f59a53a7f7d36e0"} Feb 17 13:21:24 crc kubenswrapper[4955]: I0217 13:21:24.248542 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-nbjcj" event={"ID":"96770935-7ba8-440a-8947-28a74cd86f08","Type":"ContainerStarted","Data":"db6ad3bcc9953fcc4a859c6bd4fe41a71d2e8ddd5a03484b89669ef344b6c13b"} Feb 17 13:21:24 crc kubenswrapper[4955]: I0217 13:21:24.248554 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-nbjcj" event={"ID":"96770935-7ba8-440a-8947-28a74cd86f08","Type":"ContainerStarted","Data":"8e436febb5d5fef37b4e50ec221617c7d6f4c6d56c107d451082bcc528e0104a"} Feb 17 13:21:24 crc kubenswrapper[4955]: I0217 13:21:24.248566 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-9f71-account-create-update-fhns5" event={"ID":"19e2a114-5254-4281-9d4d-209c26dcb1d6","Type":"ContainerDied","Data":"bb5673a151e591e332e71278606624a994c4a495305921eb69fc3e8fe412fcb5"} Feb 17 13:21:24 crc kubenswrapper[4955]: I0217 13:21:24.248577 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-9f71-account-create-update-fhns5" event={"ID":"19e2a114-5254-4281-9d4d-209c26dcb1d6","Type":"ContainerStarted","Data":"ba04061f717fd9779fc543c7e59f7868d8dff4786f00399580ab6f8a3723d862"} Feb 17 13:21:24 crc kubenswrapper[4955]: I0217 13:21:24.249060 4955 scope.go:117] "RemoveContainer" containerID="d0afe951ba8e2e8c737ca5d913a77d6e9b224faca55462a95c93bd441480fded" Feb 17 13:21:24 crc kubenswrapper[4955]: I0217 13:21:24.304402 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-7gmhj" podStartSLOduration=3.304342918 podStartE2EDuration="3.304342918s" podCreationTimestamp="2026-02-17 13:21:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:21:24.292422598 +0000 UTC m=+1022.815152141" watchObservedRunningTime="2026-02-17 13:21:24.304342918 +0000 UTC m=+1022.827072471" Feb 17 13:21:24 crc kubenswrapper[4955]: I0217 13:21:24.311299 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-be28-account-create-update-xx2j6" podStartSLOduration=4.311270535 podStartE2EDuration="4.311270535s" podCreationTimestamp="2026-02-17 13:21:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:21:24.31038206 +0000 UTC m=+1022.833111603" watchObservedRunningTime="2026-02-17 13:21:24.311270535 +0000 UTC m=+1022.834000078" Feb 17 13:21:24 crc kubenswrapper[4955]: I0217 13:21:24.503176 4955 scope.go:117] "RemoveContainer" containerID="c08cce887e36aaef58e4a5e04814b2e233ede2d92804b418dbcd45477d0427d8" Feb 17 13:21:24 crc kubenswrapper[4955]: I0217 13:21:24.533412 4955 scope.go:117] "RemoveContainer" containerID="d0afe951ba8e2e8c737ca5d913a77d6e9b224faca55462a95c93bd441480fded" Feb 17 13:21:24 crc kubenswrapper[4955]: E0217 13:21:24.534462 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0afe951ba8e2e8c737ca5d913a77d6e9b224faca55462a95c93bd441480fded\": container with ID starting with d0afe951ba8e2e8c737ca5d913a77d6e9b224faca55462a95c93bd441480fded not found: ID does not exist" containerID="d0afe951ba8e2e8c737ca5d913a77d6e9b224faca55462a95c93bd441480fded" Feb 17 13:21:24 crc kubenswrapper[4955]: I0217 13:21:24.534512 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0afe951ba8e2e8c737ca5d913a77d6e9b224faca55462a95c93bd441480fded"} err="failed to get container status \"d0afe951ba8e2e8c737ca5d913a77d6e9b224faca55462a95c93bd441480fded\": rpc error: code = NotFound desc = could not find container \"d0afe951ba8e2e8c737ca5d913a77d6e9b224faca55462a95c93bd441480fded\": container with ID starting with d0afe951ba8e2e8c737ca5d913a77d6e9b224faca55462a95c93bd441480fded not found: ID does not exist" Feb 17 13:21:24 crc kubenswrapper[4955]: I0217 13:21:24.534544 4955 scope.go:117] "RemoveContainer" containerID="c08cce887e36aaef58e4a5e04814b2e233ede2d92804b418dbcd45477d0427d8" Feb 17 13:21:24 crc kubenswrapper[4955]: E0217 13:21:24.534947 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c08cce887e36aaef58e4a5e04814b2e233ede2d92804b418dbcd45477d0427d8\": container with ID starting with c08cce887e36aaef58e4a5e04814b2e233ede2d92804b418dbcd45477d0427d8 not found: ID does not exist" containerID="c08cce887e36aaef58e4a5e04814b2e233ede2d92804b418dbcd45477d0427d8" Feb 17 13:21:24 crc kubenswrapper[4955]: I0217 13:21:24.534974 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c08cce887e36aaef58e4a5e04814b2e233ede2d92804b418dbcd45477d0427d8"} err="failed to get container status \"c08cce887e36aaef58e4a5e04814b2e233ede2d92804b418dbcd45477d0427d8\": rpc error: code = NotFound desc = could not find container \"c08cce887e36aaef58e4a5e04814b2e233ede2d92804b418dbcd45477d0427d8\": container with ID starting with c08cce887e36aaef58e4a5e04814b2e233ede2d92804b418dbcd45477d0427d8 not found: ID does not exist" Feb 17 13:21:25 crc kubenswrapper[4955]: I0217 13:21:25.256112 4955 generic.go:334] "Generic (PLEG): container finished" podID="7a2f748d-5c00-4975-8888-cdea65d3ab36" containerID="04c8cf2fb0cdb10eb8ec8fc8920721e50490e64dc925d5eec0b2cad40553a688" exitCode=0 Feb 17 13:21:25 crc kubenswrapper[4955]: I0217 13:21:25.256618 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-be28-account-create-update-xx2j6" event={"ID":"7a2f748d-5c00-4975-8888-cdea65d3ab36","Type":"ContainerDied","Data":"04c8cf2fb0cdb10eb8ec8fc8920721e50490e64dc925d5eec0b2cad40553a688"} Feb 17 13:21:25 crc kubenswrapper[4955]: I0217 13:21:25.261913 4955 generic.go:334] "Generic (PLEG): container finished" podID="2d8307b9-b38d-44ed-830f-c0cb10538496" containerID="05c87d393167e736f7b09dc84dbdd5af5657783de253f846a36d99cbf6457afa" exitCode=0 Feb 17 13:21:25 crc kubenswrapper[4955]: I0217 13:21:25.262074 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-7gmhj" event={"ID":"2d8307b9-b38d-44ed-830f-c0cb10538496","Type":"ContainerDied","Data":"05c87d393167e736f7b09dc84dbdd5af5657783de253f846a36d99cbf6457afa"} Feb 17 13:21:25 crc kubenswrapper[4955]: I0217 13:21:25.263674 4955 generic.go:334] "Generic (PLEG): container finished" podID="96770935-7ba8-440a-8947-28a74cd86f08" containerID="db6ad3bcc9953fcc4a859c6bd4fe41a71d2e8ddd5a03484b89669ef344b6c13b" exitCode=0 Feb 17 13:21:25 crc kubenswrapper[4955]: I0217 13:21:25.263749 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-nbjcj" event={"ID":"96770935-7ba8-440a-8947-28a74cd86f08","Type":"ContainerDied","Data":"db6ad3bcc9953fcc4a859c6bd4fe41a71d2e8ddd5a03484b89669ef344b6c13b"} Feb 17 13:21:25 crc kubenswrapper[4955]: I0217 13:21:25.274006 4955 generic.go:334] "Generic (PLEG): container finished" podID="966dab54-42a6-452a-8459-e76ec889c9ee" containerID="4a2cd1a1e878c8b153deec8babc4bfeb329265cf49676349093f4f2c18eb192d" exitCode=0 Feb 17 13:21:25 crc kubenswrapper[4955]: I0217 13:21:25.274162 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-hvwtj" event={"ID":"966dab54-42a6-452a-8459-e76ec889c9ee","Type":"ContainerDied","Data":"4a2cd1a1e878c8b153deec8babc4bfeb329265cf49676349093f4f2c18eb192d"} Feb 17 13:21:25 crc kubenswrapper[4955]: I0217 13:21:25.763851 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-9f71-account-create-update-fhns5" Feb 17 13:21:25 crc kubenswrapper[4955]: I0217 13:21:25.854193 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/19e2a114-5254-4281-9d4d-209c26dcb1d6-operator-scripts\") pod \"19e2a114-5254-4281-9d4d-209c26dcb1d6\" (UID: \"19e2a114-5254-4281-9d4d-209c26dcb1d6\") " Feb 17 13:21:25 crc kubenswrapper[4955]: I0217 13:21:25.854241 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbb65\" (UniqueName: \"kubernetes.io/projected/19e2a114-5254-4281-9d4d-209c26dcb1d6-kube-api-access-dbb65\") pod \"19e2a114-5254-4281-9d4d-209c26dcb1d6\" (UID: \"19e2a114-5254-4281-9d4d-209c26dcb1d6\") " Feb 17 13:21:25 crc kubenswrapper[4955]: I0217 13:21:25.854891 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19e2a114-5254-4281-9d4d-209c26dcb1d6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "19e2a114-5254-4281-9d4d-209c26dcb1d6" (UID: "19e2a114-5254-4281-9d4d-209c26dcb1d6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:21:25 crc kubenswrapper[4955]: I0217 13:21:25.857693 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-509d-account-create-update-pfzj7" Feb 17 13:21:25 crc kubenswrapper[4955]: I0217 13:21:25.859115 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19e2a114-5254-4281-9d4d-209c26dcb1d6-kube-api-access-dbb65" (OuterVolumeSpecName: "kube-api-access-dbb65") pod "19e2a114-5254-4281-9d4d-209c26dcb1d6" (UID: "19e2a114-5254-4281-9d4d-209c26dcb1d6"). InnerVolumeSpecName "kube-api-access-dbb65". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:21:25 crc kubenswrapper[4955]: I0217 13:21:25.893310 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-nbjcj" Feb 17 13:21:25 crc kubenswrapper[4955]: I0217 13:21:25.955831 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2wkpl\" (UniqueName: \"kubernetes.io/projected/f54e2d0a-6914-42a1-a9d0-f8c6c5e207b0-kube-api-access-2wkpl\") pod \"f54e2d0a-6914-42a1-a9d0-f8c6c5e207b0\" (UID: \"f54e2d0a-6914-42a1-a9d0-f8c6c5e207b0\") " Feb 17 13:21:25 crc kubenswrapper[4955]: I0217 13:21:25.955904 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f54e2d0a-6914-42a1-a9d0-f8c6c5e207b0-operator-scripts\") pod \"f54e2d0a-6914-42a1-a9d0-f8c6c5e207b0\" (UID: \"f54e2d0a-6914-42a1-a9d0-f8c6c5e207b0\") " Feb 17 13:21:25 crc kubenswrapper[4955]: I0217 13:21:25.955973 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/96770935-7ba8-440a-8947-28a74cd86f08-operator-scripts\") pod \"96770935-7ba8-440a-8947-28a74cd86f08\" (UID: \"96770935-7ba8-440a-8947-28a74cd86f08\") " Feb 17 13:21:25 crc kubenswrapper[4955]: I0217 13:21:25.956025 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmnhn\" (UniqueName: \"kubernetes.io/projected/96770935-7ba8-440a-8947-28a74cd86f08-kube-api-access-xmnhn\") pod \"96770935-7ba8-440a-8947-28a74cd86f08\" (UID: \"96770935-7ba8-440a-8947-28a74cd86f08\") " Feb 17 13:21:25 crc kubenswrapper[4955]: I0217 13:21:25.956438 4955 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/19e2a114-5254-4281-9d4d-209c26dcb1d6-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:25 crc kubenswrapper[4955]: I0217 13:21:25.956470 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbb65\" (UniqueName: \"kubernetes.io/projected/19e2a114-5254-4281-9d4d-209c26dcb1d6-kube-api-access-dbb65\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:25 crc kubenswrapper[4955]: I0217 13:21:25.956459 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96770935-7ba8-440a-8947-28a74cd86f08-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "96770935-7ba8-440a-8947-28a74cd86f08" (UID: "96770935-7ba8-440a-8947-28a74cd86f08"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:21:25 crc kubenswrapper[4955]: I0217 13:21:25.956881 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f54e2d0a-6914-42a1-a9d0-f8c6c5e207b0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f54e2d0a-6914-42a1-a9d0-f8c6c5e207b0" (UID: "f54e2d0a-6914-42a1-a9d0-f8c6c5e207b0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:21:25 crc kubenswrapper[4955]: I0217 13:21:25.959207 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f54e2d0a-6914-42a1-a9d0-f8c6c5e207b0-kube-api-access-2wkpl" (OuterVolumeSpecName: "kube-api-access-2wkpl") pod "f54e2d0a-6914-42a1-a9d0-f8c6c5e207b0" (UID: "f54e2d0a-6914-42a1-a9d0-f8c6c5e207b0"). InnerVolumeSpecName "kube-api-access-2wkpl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:21:25 crc kubenswrapper[4955]: I0217 13:21:25.959914 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96770935-7ba8-440a-8947-28a74cd86f08-kube-api-access-xmnhn" (OuterVolumeSpecName: "kube-api-access-xmnhn") pod "96770935-7ba8-440a-8947-28a74cd86f08" (UID: "96770935-7ba8-440a-8947-28a74cd86f08"). InnerVolumeSpecName "kube-api-access-xmnhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:21:26 crc kubenswrapper[4955]: I0217 13:21:26.057151 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2wkpl\" (UniqueName: \"kubernetes.io/projected/f54e2d0a-6914-42a1-a9d0-f8c6c5e207b0-kube-api-access-2wkpl\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:26 crc kubenswrapper[4955]: I0217 13:21:26.057188 4955 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f54e2d0a-6914-42a1-a9d0-f8c6c5e207b0-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:26 crc kubenswrapper[4955]: I0217 13:21:26.057200 4955 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/96770935-7ba8-440a-8947-28a74cd86f08-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:26 crc kubenswrapper[4955]: I0217 13:21:26.057211 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmnhn\" (UniqueName: \"kubernetes.io/projected/96770935-7ba8-440a-8947-28a74cd86f08-kube-api-access-xmnhn\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:26 crc kubenswrapper[4955]: I0217 13:21:26.282647 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-509d-account-create-update-pfzj7" Feb 17 13:21:26 crc kubenswrapper[4955]: I0217 13:21:26.283357 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-509d-account-create-update-pfzj7" event={"ID":"f54e2d0a-6914-42a1-a9d0-f8c6c5e207b0","Type":"ContainerDied","Data":"6adf9563db64e90e0e5d2ff82bebb8a8a41d3eb8e994a701418a307b7d59f234"} Feb 17 13:21:26 crc kubenswrapper[4955]: I0217 13:21:26.283386 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6adf9563db64e90e0e5d2ff82bebb8a8a41d3eb8e994a701418a307b7d59f234" Feb 17 13:21:26 crc kubenswrapper[4955]: I0217 13:21:26.285171 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-nbjcj" event={"ID":"96770935-7ba8-440a-8947-28a74cd86f08","Type":"ContainerDied","Data":"8e436febb5d5fef37b4e50ec221617c7d6f4c6d56c107d451082bcc528e0104a"} Feb 17 13:21:26 crc kubenswrapper[4955]: I0217 13:21:26.285221 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8e436febb5d5fef37b4e50ec221617c7d6f4c6d56c107d451082bcc528e0104a" Feb 17 13:21:26 crc kubenswrapper[4955]: I0217 13:21:26.285348 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-nbjcj" Feb 17 13:21:26 crc kubenswrapper[4955]: I0217 13:21:26.290273 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-9f71-account-create-update-fhns5" event={"ID":"19e2a114-5254-4281-9d4d-209c26dcb1d6","Type":"ContainerDied","Data":"ba04061f717fd9779fc543c7e59f7868d8dff4786f00399580ab6f8a3723d862"} Feb 17 13:21:26 crc kubenswrapper[4955]: I0217 13:21:26.290340 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba04061f717fd9779fc543c7e59f7868d8dff4786f00399580ab6f8a3723d862" Feb 17 13:21:26 crc kubenswrapper[4955]: I0217 13:21:26.290374 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-9f71-account-create-update-fhns5" Feb 17 13:21:26 crc kubenswrapper[4955]: I0217 13:21:26.648499 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-hvwtj" Feb 17 13:21:26 crc kubenswrapper[4955]: I0217 13:21:26.771904 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/966dab54-42a6-452a-8459-e76ec889c9ee-operator-scripts\") pod \"966dab54-42a6-452a-8459-e76ec889c9ee\" (UID: \"966dab54-42a6-452a-8459-e76ec889c9ee\") " Feb 17 13:21:26 crc kubenswrapper[4955]: I0217 13:21:26.772099 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thppx\" (UniqueName: \"kubernetes.io/projected/966dab54-42a6-452a-8459-e76ec889c9ee-kube-api-access-thppx\") pod \"966dab54-42a6-452a-8459-e76ec889c9ee\" (UID: \"966dab54-42a6-452a-8459-e76ec889c9ee\") " Feb 17 13:21:26 crc kubenswrapper[4955]: I0217 13:21:26.773086 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/966dab54-42a6-452a-8459-e76ec889c9ee-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "966dab54-42a6-452a-8459-e76ec889c9ee" (UID: "966dab54-42a6-452a-8459-e76ec889c9ee"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:21:26 crc kubenswrapper[4955]: I0217 13:21:26.778069 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/966dab54-42a6-452a-8459-e76ec889c9ee-kube-api-access-thppx" (OuterVolumeSpecName: "kube-api-access-thppx") pod "966dab54-42a6-452a-8459-e76ec889c9ee" (UID: "966dab54-42a6-452a-8459-e76ec889c9ee"). InnerVolumeSpecName "kube-api-access-thppx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:21:26 crc kubenswrapper[4955]: I0217 13:21:26.848367 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-be28-account-create-update-xx2j6" Feb 17 13:21:26 crc kubenswrapper[4955]: I0217 13:21:26.854894 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-7gmhj" Feb 17 13:21:26 crc kubenswrapper[4955]: I0217 13:21:26.874114 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thppx\" (UniqueName: \"kubernetes.io/projected/966dab54-42a6-452a-8459-e76ec889c9ee-kube-api-access-thppx\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:26 crc kubenswrapper[4955]: I0217 13:21:26.874138 4955 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/966dab54-42a6-452a-8459-e76ec889c9ee-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:26 crc kubenswrapper[4955]: I0217 13:21:26.975505 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d8307b9-b38d-44ed-830f-c0cb10538496-operator-scripts\") pod \"2d8307b9-b38d-44ed-830f-c0cb10538496\" (UID: \"2d8307b9-b38d-44ed-830f-c0cb10538496\") " Feb 17 13:21:26 crc kubenswrapper[4955]: I0217 13:21:26.975544 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4f5tf\" (UniqueName: \"kubernetes.io/projected/7a2f748d-5c00-4975-8888-cdea65d3ab36-kube-api-access-4f5tf\") pod \"7a2f748d-5c00-4975-8888-cdea65d3ab36\" (UID: \"7a2f748d-5c00-4975-8888-cdea65d3ab36\") " Feb 17 13:21:26 crc kubenswrapper[4955]: I0217 13:21:26.975581 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7a2f748d-5c00-4975-8888-cdea65d3ab36-operator-scripts\") pod \"7a2f748d-5c00-4975-8888-cdea65d3ab36\" (UID: \"7a2f748d-5c00-4975-8888-cdea65d3ab36\") " Feb 17 13:21:26 crc kubenswrapper[4955]: I0217 13:21:26.975620 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7g6cd\" (UniqueName: \"kubernetes.io/projected/2d8307b9-b38d-44ed-830f-c0cb10538496-kube-api-access-7g6cd\") pod \"2d8307b9-b38d-44ed-830f-c0cb10538496\" (UID: \"2d8307b9-b38d-44ed-830f-c0cb10538496\") " Feb 17 13:21:26 crc kubenswrapper[4955]: I0217 13:21:26.977091 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d8307b9-b38d-44ed-830f-c0cb10538496-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2d8307b9-b38d-44ed-830f-c0cb10538496" (UID: "2d8307b9-b38d-44ed-830f-c0cb10538496"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:21:26 crc kubenswrapper[4955]: I0217 13:21:26.977276 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a2f748d-5c00-4975-8888-cdea65d3ab36-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7a2f748d-5c00-4975-8888-cdea65d3ab36" (UID: "7a2f748d-5c00-4975-8888-cdea65d3ab36"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:21:26 crc kubenswrapper[4955]: I0217 13:21:26.979115 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d8307b9-b38d-44ed-830f-c0cb10538496-kube-api-access-7g6cd" (OuterVolumeSpecName: "kube-api-access-7g6cd") pod "2d8307b9-b38d-44ed-830f-c0cb10538496" (UID: "2d8307b9-b38d-44ed-830f-c0cb10538496"). InnerVolumeSpecName "kube-api-access-7g6cd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:21:26 crc kubenswrapper[4955]: I0217 13:21:26.979159 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a2f748d-5c00-4975-8888-cdea65d3ab36-kube-api-access-4f5tf" (OuterVolumeSpecName: "kube-api-access-4f5tf") pod "7a2f748d-5c00-4975-8888-cdea65d3ab36" (UID: "7a2f748d-5c00-4975-8888-cdea65d3ab36"). InnerVolumeSpecName "kube-api-access-4f5tf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.077973 4955 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d8307b9-b38d-44ed-830f-c0cb10538496-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.078039 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4f5tf\" (UniqueName: \"kubernetes.io/projected/7a2f748d-5c00-4975-8888-cdea65d3ab36-kube-api-access-4f5tf\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.078061 4955 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7a2f748d-5c00-4975-8888-cdea65d3ab36-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.078080 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7g6cd\" (UniqueName: \"kubernetes.io/projected/2d8307b9-b38d-44ed-830f-c0cb10538496-kube-api-access-7g6cd\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.300109 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-7gmhj" event={"ID":"2d8307b9-b38d-44ed-830f-c0cb10538496","Type":"ContainerDied","Data":"31650270abaf72b5c7fbdbb4e80d7fd1a358a89f2658c7107f59a53a7f7d36e0"} Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.300158 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31650270abaf72b5c7fbdbb4e80d7fd1a358a89f2658c7107f59a53a7f7d36e0" Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.300206 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-7gmhj" Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.308534 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-hvwtj" event={"ID":"966dab54-42a6-452a-8459-e76ec889c9ee","Type":"ContainerDied","Data":"3e0d343a7853a6d70356f65601f0246273fc64f41ebb2ba9fdd2f21aec2d070a"} Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.308573 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e0d343a7853a6d70356f65601f0246273fc64f41ebb2ba9fdd2f21aec2d070a" Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.308594 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-hvwtj" Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.317338 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-be28-account-create-update-xx2j6" event={"ID":"7a2f748d-5c00-4975-8888-cdea65d3ab36","Type":"ContainerDied","Data":"7138254bf4d126537f29a027ca2d7246fab42e34c840a781d99a1a1e8aec8b61"} Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.317401 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7138254bf4d126537f29a027ca2d7246fab42e34c840a781d99a1a1e8aec8b61" Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.317774 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-be28-account-create-update-xx2j6" Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.318151 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-cqgbw"] Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.328648 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-cqgbw"] Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.385903 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-hwwtj"] Feb 17 13:21:27 crc kubenswrapper[4955]: E0217 13:21:27.386319 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f54e2d0a-6914-42a1-a9d0-f8c6c5e207b0" containerName="mariadb-account-create-update" Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.386340 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="f54e2d0a-6914-42a1-a9d0-f8c6c5e207b0" containerName="mariadb-account-create-update" Feb 17 13:21:27 crc kubenswrapper[4955]: E0217 13:21:27.386354 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd0771e6-147c-4069-88e1-c9ae122dc55d" containerName="dnsmasq-dns" Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.386362 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd0771e6-147c-4069-88e1-c9ae122dc55d" containerName="dnsmasq-dns" Feb 17 13:21:27 crc kubenswrapper[4955]: E0217 13:21:27.386385 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96770935-7ba8-440a-8947-28a74cd86f08" containerName="mariadb-database-create" Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.386393 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="96770935-7ba8-440a-8947-28a74cd86f08" containerName="mariadb-database-create" Feb 17 13:21:27 crc kubenswrapper[4955]: E0217 13:21:27.386409 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d8307b9-b38d-44ed-830f-c0cb10538496" containerName="mariadb-database-create" Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.386416 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d8307b9-b38d-44ed-830f-c0cb10538496" containerName="mariadb-database-create" Feb 17 13:21:27 crc kubenswrapper[4955]: E0217 13:21:27.386425 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5fe5a03-2bb2-450f-8258-7cf8ad985475" containerName="mariadb-account-create-update" Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.386432 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5fe5a03-2bb2-450f-8258-7cf8ad985475" containerName="mariadb-account-create-update" Feb 17 13:21:27 crc kubenswrapper[4955]: E0217 13:21:27.386446 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19e2a114-5254-4281-9d4d-209c26dcb1d6" containerName="mariadb-account-create-update" Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.386453 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="19e2a114-5254-4281-9d4d-209c26dcb1d6" containerName="mariadb-account-create-update" Feb 17 13:21:27 crc kubenswrapper[4955]: E0217 13:21:27.386467 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="966dab54-42a6-452a-8459-e76ec889c9ee" containerName="mariadb-database-create" Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.386474 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="966dab54-42a6-452a-8459-e76ec889c9ee" containerName="mariadb-database-create" Feb 17 13:21:27 crc kubenswrapper[4955]: E0217 13:21:27.386489 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd0771e6-147c-4069-88e1-c9ae122dc55d" containerName="init" Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.386496 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd0771e6-147c-4069-88e1-c9ae122dc55d" containerName="init" Feb 17 13:21:27 crc kubenswrapper[4955]: E0217 13:21:27.386510 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a2f748d-5c00-4975-8888-cdea65d3ab36" containerName="mariadb-account-create-update" Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.386517 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a2f748d-5c00-4975-8888-cdea65d3ab36" containerName="mariadb-account-create-update" Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.386688 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="96770935-7ba8-440a-8947-28a74cd86f08" containerName="mariadb-database-create" Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.386707 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5fe5a03-2bb2-450f-8258-7cf8ad985475" containerName="mariadb-account-create-update" Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.386716 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd0771e6-147c-4069-88e1-c9ae122dc55d" containerName="dnsmasq-dns" Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.386731 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="966dab54-42a6-452a-8459-e76ec889c9ee" containerName="mariadb-database-create" Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.386742 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="19e2a114-5254-4281-9d4d-209c26dcb1d6" containerName="mariadb-account-create-update" Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.386755 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d8307b9-b38d-44ed-830f-c0cb10538496" containerName="mariadb-database-create" Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.386768 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="f54e2d0a-6914-42a1-a9d0-f8c6c5e207b0" containerName="mariadb-account-create-update" Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.386795 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a2f748d-5c00-4975-8888-cdea65d3ab36" containerName="mariadb-account-create-update" Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.387267 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-hwwtj" Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.389320 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.392583 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-hwwtj"] Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.586638 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b7293150-5c3f-4348-8f9e-2f46a2373d1d-operator-scripts\") pod \"root-account-create-update-hwwtj\" (UID: \"b7293150-5c3f-4348-8f9e-2f46a2373d1d\") " pod="openstack/root-account-create-update-hwwtj" Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.586801 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srnk7\" (UniqueName: \"kubernetes.io/projected/b7293150-5c3f-4348-8f9e-2f46a2373d1d-kube-api-access-srnk7\") pod \"root-account-create-update-hwwtj\" (UID: \"b7293150-5c3f-4348-8f9e-2f46a2373d1d\") " pod="openstack/root-account-create-update-hwwtj" Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.688068 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srnk7\" (UniqueName: \"kubernetes.io/projected/b7293150-5c3f-4348-8f9e-2f46a2373d1d-kube-api-access-srnk7\") pod \"root-account-create-update-hwwtj\" (UID: \"b7293150-5c3f-4348-8f9e-2f46a2373d1d\") " pod="openstack/root-account-create-update-hwwtj" Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.688187 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b7293150-5c3f-4348-8f9e-2f46a2373d1d-operator-scripts\") pod \"root-account-create-update-hwwtj\" (UID: \"b7293150-5c3f-4348-8f9e-2f46a2373d1d\") " pod="openstack/root-account-create-update-hwwtj" Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.689267 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b7293150-5c3f-4348-8f9e-2f46a2373d1d-operator-scripts\") pod \"root-account-create-update-hwwtj\" (UID: \"b7293150-5c3f-4348-8f9e-2f46a2373d1d\") " pod="openstack/root-account-create-update-hwwtj" Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.703142 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srnk7\" (UniqueName: \"kubernetes.io/projected/b7293150-5c3f-4348-8f9e-2f46a2373d1d-kube-api-access-srnk7\") pod \"root-account-create-update-hwwtj\" (UID: \"b7293150-5c3f-4348-8f9e-2f46a2373d1d\") " pod="openstack/root-account-create-update-hwwtj" Feb 17 13:21:27 crc kubenswrapper[4955]: I0217 13:21:27.705157 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-hwwtj" Feb 17 13:21:28 crc kubenswrapper[4955]: I0217 13:21:28.160953 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-hwwtj"] Feb 17 13:21:28 crc kubenswrapper[4955]: I0217 13:21:28.232494 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5fe5a03-2bb2-450f-8258-7cf8ad985475" path="/var/lib/kubelet/pods/e5fe5a03-2bb2-450f-8258-7cf8ad985475/volumes" Feb 17 13:21:28 crc kubenswrapper[4955]: I0217 13:21:28.325856 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-hwwtj" event={"ID":"b7293150-5c3f-4348-8f9e-2f46a2373d1d","Type":"ContainerStarted","Data":"afc26b8d48036bd9870642c758345deb2628c68cf04e7b3c61281574d4939e96"} Feb 17 13:21:28 crc kubenswrapper[4955]: I0217 13:21:28.801503 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Feb 17 13:21:29 crc kubenswrapper[4955]: I0217 13:21:29.332737 4955 generic.go:334] "Generic (PLEG): container finished" podID="b7293150-5c3f-4348-8f9e-2f46a2373d1d" containerID="9e4eee87ea3f68861eadb36091cc1bcbece301c8473a28c4c868c5197488131a" exitCode=0 Feb 17 13:21:29 crc kubenswrapper[4955]: I0217 13:21:29.333119 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-hwwtj" event={"ID":"b7293150-5c3f-4348-8f9e-2f46a2373d1d","Type":"ContainerDied","Data":"9e4eee87ea3f68861eadb36091cc1bcbece301c8473a28c4c868c5197488131a"} Feb 17 13:21:29 crc kubenswrapper[4955]: I0217 13:21:29.926000 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/fb6d9434-4476-438b-90e4-4781bd85b0bb-etc-swift\") pod \"swift-storage-0\" (UID: \"fb6d9434-4476-438b-90e4-4781bd85b0bb\") " pod="openstack/swift-storage-0" Feb 17 13:21:29 crc kubenswrapper[4955]: E0217 13:21:29.926209 4955 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 17 13:21:29 crc kubenswrapper[4955]: E0217 13:21:29.926458 4955 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 17 13:21:29 crc kubenswrapper[4955]: E0217 13:21:29.926513 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/fb6d9434-4476-438b-90e4-4781bd85b0bb-etc-swift podName:fb6d9434-4476-438b-90e4-4781bd85b0bb nodeName:}" failed. No retries permitted until 2026-02-17 13:21:45.926495816 +0000 UTC m=+1044.449225359 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/fb6d9434-4476-438b-90e4-4781bd85b0bb-etc-swift") pod "swift-storage-0" (UID: "fb6d9434-4476-438b-90e4-4781bd85b0bb") : configmap "swift-ring-files" not found Feb 17 13:21:30 crc kubenswrapper[4955]: I0217 13:21:30.342822 4955 generic.go:334] "Generic (PLEG): container finished" podID="46beefa1-21de-4818-a70f-e9f840a920ec" containerID="5f190e937676a3c7f39f40f4b14cf50b9c4c117f66d82dde7fa030590da7ccef" exitCode=0 Feb 17 13:21:30 crc kubenswrapper[4955]: I0217 13:21:30.342891 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-56p7b" event={"ID":"46beefa1-21de-4818-a70f-e9f840a920ec","Type":"ContainerDied","Data":"5f190e937676a3c7f39f40f4b14cf50b9c4c117f66d82dde7fa030590da7ccef"} Feb 17 13:21:30 crc kubenswrapper[4955]: I0217 13:21:30.697060 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-hwwtj" Feb 17 13:21:30 crc kubenswrapper[4955]: I0217 13:21:30.841395 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b7293150-5c3f-4348-8f9e-2f46a2373d1d-operator-scripts\") pod \"b7293150-5c3f-4348-8f9e-2f46a2373d1d\" (UID: \"b7293150-5c3f-4348-8f9e-2f46a2373d1d\") " Feb 17 13:21:30 crc kubenswrapper[4955]: I0217 13:21:30.841603 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srnk7\" (UniqueName: \"kubernetes.io/projected/b7293150-5c3f-4348-8f9e-2f46a2373d1d-kube-api-access-srnk7\") pod \"b7293150-5c3f-4348-8f9e-2f46a2373d1d\" (UID: \"b7293150-5c3f-4348-8f9e-2f46a2373d1d\") " Feb 17 13:21:30 crc kubenswrapper[4955]: I0217 13:21:30.842274 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7293150-5c3f-4348-8f9e-2f46a2373d1d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b7293150-5c3f-4348-8f9e-2f46a2373d1d" (UID: "b7293150-5c3f-4348-8f9e-2f46a2373d1d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:21:30 crc kubenswrapper[4955]: I0217 13:21:30.848252 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7293150-5c3f-4348-8f9e-2f46a2373d1d-kube-api-access-srnk7" (OuterVolumeSpecName: "kube-api-access-srnk7") pod "b7293150-5c3f-4348-8f9e-2f46a2373d1d" (UID: "b7293150-5c3f-4348-8f9e-2f46a2373d1d"). InnerVolumeSpecName "kube-api-access-srnk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:21:30 crc kubenswrapper[4955]: I0217 13:21:30.894522 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-jwz6b"] Feb 17 13:21:30 crc kubenswrapper[4955]: E0217 13:21:30.895009 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7293150-5c3f-4348-8f9e-2f46a2373d1d" containerName="mariadb-account-create-update" Feb 17 13:21:30 crc kubenswrapper[4955]: I0217 13:21:30.895033 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7293150-5c3f-4348-8f9e-2f46a2373d1d" containerName="mariadb-account-create-update" Feb 17 13:21:30 crc kubenswrapper[4955]: I0217 13:21:30.895252 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7293150-5c3f-4348-8f9e-2f46a2373d1d" containerName="mariadb-account-create-update" Feb 17 13:21:30 crc kubenswrapper[4955]: I0217 13:21:30.895796 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-jwz6b" Feb 17 13:21:30 crc kubenswrapper[4955]: I0217 13:21:30.898167 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-d6l2j" Feb 17 13:21:30 crc kubenswrapper[4955]: I0217 13:21:30.898190 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Feb 17 13:21:30 crc kubenswrapper[4955]: I0217 13:21:30.902850 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-jwz6b"] Feb 17 13:21:30 crc kubenswrapper[4955]: I0217 13:21:30.943888 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srnk7\" (UniqueName: \"kubernetes.io/projected/b7293150-5c3f-4348-8f9e-2f46a2373d1d-kube-api-access-srnk7\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:30 crc kubenswrapper[4955]: I0217 13:21:30.943929 4955 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b7293150-5c3f-4348-8f9e-2f46a2373d1d-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:31 crc kubenswrapper[4955]: I0217 13:21:31.045296 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xpsh\" (UniqueName: \"kubernetes.io/projected/aed5c07b-3d7a-43c1-a11f-4a4220289bef-kube-api-access-2xpsh\") pod \"glance-db-sync-jwz6b\" (UID: \"aed5c07b-3d7a-43c1-a11f-4a4220289bef\") " pod="openstack/glance-db-sync-jwz6b" Feb 17 13:21:31 crc kubenswrapper[4955]: I0217 13:21:31.045369 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aed5c07b-3d7a-43c1-a11f-4a4220289bef-combined-ca-bundle\") pod \"glance-db-sync-jwz6b\" (UID: \"aed5c07b-3d7a-43c1-a11f-4a4220289bef\") " pod="openstack/glance-db-sync-jwz6b" Feb 17 13:21:31 crc kubenswrapper[4955]: I0217 13:21:31.045432 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/aed5c07b-3d7a-43c1-a11f-4a4220289bef-db-sync-config-data\") pod \"glance-db-sync-jwz6b\" (UID: \"aed5c07b-3d7a-43c1-a11f-4a4220289bef\") " pod="openstack/glance-db-sync-jwz6b" Feb 17 13:21:31 crc kubenswrapper[4955]: I0217 13:21:31.045458 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aed5c07b-3d7a-43c1-a11f-4a4220289bef-config-data\") pod \"glance-db-sync-jwz6b\" (UID: \"aed5c07b-3d7a-43c1-a11f-4a4220289bef\") " pod="openstack/glance-db-sync-jwz6b" Feb 17 13:21:31 crc kubenswrapper[4955]: I0217 13:21:31.146381 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/aed5c07b-3d7a-43c1-a11f-4a4220289bef-db-sync-config-data\") pod \"glance-db-sync-jwz6b\" (UID: \"aed5c07b-3d7a-43c1-a11f-4a4220289bef\") " pod="openstack/glance-db-sync-jwz6b" Feb 17 13:21:31 crc kubenswrapper[4955]: I0217 13:21:31.146439 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aed5c07b-3d7a-43c1-a11f-4a4220289bef-config-data\") pod \"glance-db-sync-jwz6b\" (UID: \"aed5c07b-3d7a-43c1-a11f-4a4220289bef\") " pod="openstack/glance-db-sync-jwz6b" Feb 17 13:21:31 crc kubenswrapper[4955]: I0217 13:21:31.146508 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xpsh\" (UniqueName: \"kubernetes.io/projected/aed5c07b-3d7a-43c1-a11f-4a4220289bef-kube-api-access-2xpsh\") pod \"glance-db-sync-jwz6b\" (UID: \"aed5c07b-3d7a-43c1-a11f-4a4220289bef\") " pod="openstack/glance-db-sync-jwz6b" Feb 17 13:21:31 crc kubenswrapper[4955]: I0217 13:21:31.146550 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aed5c07b-3d7a-43c1-a11f-4a4220289bef-combined-ca-bundle\") pod \"glance-db-sync-jwz6b\" (UID: \"aed5c07b-3d7a-43c1-a11f-4a4220289bef\") " pod="openstack/glance-db-sync-jwz6b" Feb 17 13:21:31 crc kubenswrapper[4955]: I0217 13:21:31.150166 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aed5c07b-3d7a-43c1-a11f-4a4220289bef-combined-ca-bundle\") pod \"glance-db-sync-jwz6b\" (UID: \"aed5c07b-3d7a-43c1-a11f-4a4220289bef\") " pod="openstack/glance-db-sync-jwz6b" Feb 17 13:21:31 crc kubenswrapper[4955]: I0217 13:21:31.150184 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/aed5c07b-3d7a-43c1-a11f-4a4220289bef-db-sync-config-data\") pod \"glance-db-sync-jwz6b\" (UID: \"aed5c07b-3d7a-43c1-a11f-4a4220289bef\") " pod="openstack/glance-db-sync-jwz6b" Feb 17 13:21:31 crc kubenswrapper[4955]: I0217 13:21:31.150201 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aed5c07b-3d7a-43c1-a11f-4a4220289bef-config-data\") pod \"glance-db-sync-jwz6b\" (UID: \"aed5c07b-3d7a-43c1-a11f-4a4220289bef\") " pod="openstack/glance-db-sync-jwz6b" Feb 17 13:21:31 crc kubenswrapper[4955]: I0217 13:21:31.163493 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xpsh\" (UniqueName: \"kubernetes.io/projected/aed5c07b-3d7a-43c1-a11f-4a4220289bef-kube-api-access-2xpsh\") pod \"glance-db-sync-jwz6b\" (UID: \"aed5c07b-3d7a-43c1-a11f-4a4220289bef\") " pod="openstack/glance-db-sync-jwz6b" Feb 17 13:21:31 crc kubenswrapper[4955]: I0217 13:21:31.219487 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-jwz6b" Feb 17 13:21:31 crc kubenswrapper[4955]: I0217 13:21:31.351904 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-hwwtj" Feb 17 13:21:31 crc kubenswrapper[4955]: I0217 13:21:31.351900 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-hwwtj" event={"ID":"b7293150-5c3f-4348-8f9e-2f46a2373d1d","Type":"ContainerDied","Data":"afc26b8d48036bd9870642c758345deb2628c68cf04e7b3c61281574d4939e96"} Feb 17 13:21:31 crc kubenswrapper[4955]: I0217 13:21:31.352267 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="afc26b8d48036bd9870642c758345deb2628c68cf04e7b3c61281574d4939e96" Feb 17 13:21:31 crc kubenswrapper[4955]: I0217 13:21:31.601368 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-56p7b" Feb 17 13:21:31 crc kubenswrapper[4955]: I0217 13:21:31.721293 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-jwz6b"] Feb 17 13:21:31 crc kubenswrapper[4955]: W0217 13:21:31.725691 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaed5c07b_3d7a_43c1_a11f_4a4220289bef.slice/crio-6b7e8abcbf0d6f84ebbc3b884eafa2b0b4be79cc92dca2ec4808cb08a9d447d8 WatchSource:0}: Error finding container 6b7e8abcbf0d6f84ebbc3b884eafa2b0b4be79cc92dca2ec4808cb08a9d447d8: Status 404 returned error can't find the container with id 6b7e8abcbf0d6f84ebbc3b884eafa2b0b4be79cc92dca2ec4808cb08a9d447d8 Feb 17 13:21:31 crc kubenswrapper[4955]: I0217 13:21:31.756365 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/46beefa1-21de-4818-a70f-e9f840a920ec-dispersionconf\") pod \"46beefa1-21de-4818-a70f-e9f840a920ec\" (UID: \"46beefa1-21de-4818-a70f-e9f840a920ec\") " Feb 17 13:21:31 crc kubenswrapper[4955]: I0217 13:21:31.756402 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46beefa1-21de-4818-a70f-e9f840a920ec-combined-ca-bundle\") pod \"46beefa1-21de-4818-a70f-e9f840a920ec\" (UID: \"46beefa1-21de-4818-a70f-e9f840a920ec\") " Feb 17 13:21:31 crc kubenswrapper[4955]: I0217 13:21:31.756465 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/46beefa1-21de-4818-a70f-e9f840a920ec-scripts\") pod \"46beefa1-21de-4818-a70f-e9f840a920ec\" (UID: \"46beefa1-21de-4818-a70f-e9f840a920ec\") " Feb 17 13:21:31 crc kubenswrapper[4955]: I0217 13:21:31.756510 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/46beefa1-21de-4818-a70f-e9f840a920ec-etc-swift\") pod \"46beefa1-21de-4818-a70f-e9f840a920ec\" (UID: \"46beefa1-21de-4818-a70f-e9f840a920ec\") " Feb 17 13:21:31 crc kubenswrapper[4955]: I0217 13:21:31.756567 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/46beefa1-21de-4818-a70f-e9f840a920ec-swiftconf\") pod \"46beefa1-21de-4818-a70f-e9f840a920ec\" (UID: \"46beefa1-21de-4818-a70f-e9f840a920ec\") " Feb 17 13:21:31 crc kubenswrapper[4955]: I0217 13:21:31.756657 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sz7pl\" (UniqueName: \"kubernetes.io/projected/46beefa1-21de-4818-a70f-e9f840a920ec-kube-api-access-sz7pl\") pod \"46beefa1-21de-4818-a70f-e9f840a920ec\" (UID: \"46beefa1-21de-4818-a70f-e9f840a920ec\") " Feb 17 13:21:31 crc kubenswrapper[4955]: I0217 13:21:31.756759 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/46beefa1-21de-4818-a70f-e9f840a920ec-ring-data-devices\") pod \"46beefa1-21de-4818-a70f-e9f840a920ec\" (UID: \"46beefa1-21de-4818-a70f-e9f840a920ec\") " Feb 17 13:21:31 crc kubenswrapper[4955]: I0217 13:21:31.757509 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46beefa1-21de-4818-a70f-e9f840a920ec-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "46beefa1-21de-4818-a70f-e9f840a920ec" (UID: "46beefa1-21de-4818-a70f-e9f840a920ec"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:21:31 crc kubenswrapper[4955]: I0217 13:21:31.757740 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46beefa1-21de-4818-a70f-e9f840a920ec-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "46beefa1-21de-4818-a70f-e9f840a920ec" (UID: "46beefa1-21de-4818-a70f-e9f840a920ec"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:21:31 crc kubenswrapper[4955]: I0217 13:21:31.761155 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46beefa1-21de-4818-a70f-e9f840a920ec-kube-api-access-sz7pl" (OuterVolumeSpecName: "kube-api-access-sz7pl") pod "46beefa1-21de-4818-a70f-e9f840a920ec" (UID: "46beefa1-21de-4818-a70f-e9f840a920ec"). InnerVolumeSpecName "kube-api-access-sz7pl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:21:31 crc kubenswrapper[4955]: I0217 13:21:31.764027 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46beefa1-21de-4818-a70f-e9f840a920ec-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "46beefa1-21de-4818-a70f-e9f840a920ec" (UID: "46beefa1-21de-4818-a70f-e9f840a920ec"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:21:31 crc kubenswrapper[4955]: I0217 13:21:31.774449 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46beefa1-21de-4818-a70f-e9f840a920ec-scripts" (OuterVolumeSpecName: "scripts") pod "46beefa1-21de-4818-a70f-e9f840a920ec" (UID: "46beefa1-21de-4818-a70f-e9f840a920ec"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:21:31 crc kubenswrapper[4955]: I0217 13:21:31.777813 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46beefa1-21de-4818-a70f-e9f840a920ec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "46beefa1-21de-4818-a70f-e9f840a920ec" (UID: "46beefa1-21de-4818-a70f-e9f840a920ec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:21:31 crc kubenswrapper[4955]: I0217 13:21:31.780939 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46beefa1-21de-4818-a70f-e9f840a920ec-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "46beefa1-21de-4818-a70f-e9f840a920ec" (UID: "46beefa1-21de-4818-a70f-e9f840a920ec"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:21:31 crc kubenswrapper[4955]: I0217 13:21:31.859020 4955 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/46beefa1-21de-4818-a70f-e9f840a920ec-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:31 crc kubenswrapper[4955]: I0217 13:21:31.859074 4955 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/46beefa1-21de-4818-a70f-e9f840a920ec-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:31 crc kubenswrapper[4955]: I0217 13:21:31.859085 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46beefa1-21de-4818-a70f-e9f840a920ec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:31 crc kubenswrapper[4955]: I0217 13:21:31.859099 4955 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/46beefa1-21de-4818-a70f-e9f840a920ec-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:31 crc kubenswrapper[4955]: I0217 13:21:31.859109 4955 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/46beefa1-21de-4818-a70f-e9f840a920ec-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:31 crc kubenswrapper[4955]: I0217 13:21:31.859119 4955 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/46beefa1-21de-4818-a70f-e9f840a920ec-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:31 crc kubenswrapper[4955]: I0217 13:21:31.859130 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sz7pl\" (UniqueName: \"kubernetes.io/projected/46beefa1-21de-4818-a70f-e9f840a920ec-kube-api-access-sz7pl\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:32 crc kubenswrapper[4955]: I0217 13:21:32.359182 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-56p7b" event={"ID":"46beefa1-21de-4818-a70f-e9f840a920ec","Type":"ContainerDied","Data":"4c5bb6dc3c345b1a51f563296d4c182af165466f26f7a79b1b3e017cba8194aa"} Feb 17 13:21:32 crc kubenswrapper[4955]: I0217 13:21:32.359504 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c5bb6dc3c345b1a51f563296d4c182af165466f26f7a79b1b3e017cba8194aa" Feb 17 13:21:32 crc kubenswrapper[4955]: I0217 13:21:32.359197 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-56p7b" Feb 17 13:21:32 crc kubenswrapper[4955]: I0217 13:21:32.360616 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-jwz6b" event={"ID":"aed5c07b-3d7a-43c1-a11f-4a4220289bef","Type":"ContainerStarted","Data":"6b7e8abcbf0d6f84ebbc3b884eafa2b0b4be79cc92dca2ec4808cb08a9d447d8"} Feb 17 13:21:35 crc kubenswrapper[4955]: I0217 13:21:35.525076 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-f8cjz" podUID="452bb98a-815b-4154-84c5-6b31338fca29" containerName="ovn-controller" probeResult="failure" output=< Feb 17 13:21:35 crc kubenswrapper[4955]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Feb 17 13:21:35 crc kubenswrapper[4955]: > Feb 17 13:21:35 crc kubenswrapper[4955]: I0217 13:21:35.594490 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-m27b4" Feb 17 13:21:35 crc kubenswrapper[4955]: I0217 13:21:35.599155 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-m27b4" Feb 17 13:21:35 crc kubenswrapper[4955]: I0217 13:21:35.829316 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-f8cjz-config-z9pqp"] Feb 17 13:21:35 crc kubenswrapper[4955]: E0217 13:21:35.832265 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46beefa1-21de-4818-a70f-e9f840a920ec" containerName="swift-ring-rebalance" Feb 17 13:21:35 crc kubenswrapper[4955]: I0217 13:21:35.832287 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="46beefa1-21de-4818-a70f-e9f840a920ec" containerName="swift-ring-rebalance" Feb 17 13:21:35 crc kubenswrapper[4955]: I0217 13:21:35.832452 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="46beefa1-21de-4818-a70f-e9f840a920ec" containerName="swift-ring-rebalance" Feb 17 13:21:35 crc kubenswrapper[4955]: I0217 13:21:35.832988 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-f8cjz-config-z9pqp" Feb 17 13:21:35 crc kubenswrapper[4955]: I0217 13:21:35.834840 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Feb 17 13:21:35 crc kubenswrapper[4955]: I0217 13:21:35.850154 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-f8cjz-config-z9pqp"] Feb 17 13:21:35 crc kubenswrapper[4955]: I0217 13:21:35.956594 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlwht\" (UniqueName: \"kubernetes.io/projected/797dd93e-1f77-4edf-a02e-398795307f04-kube-api-access-mlwht\") pod \"ovn-controller-f8cjz-config-z9pqp\" (UID: \"797dd93e-1f77-4edf-a02e-398795307f04\") " pod="openstack/ovn-controller-f8cjz-config-z9pqp" Feb 17 13:21:35 crc kubenswrapper[4955]: I0217 13:21:35.956664 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/797dd93e-1f77-4edf-a02e-398795307f04-var-run-ovn\") pod \"ovn-controller-f8cjz-config-z9pqp\" (UID: \"797dd93e-1f77-4edf-a02e-398795307f04\") " pod="openstack/ovn-controller-f8cjz-config-z9pqp" Feb 17 13:21:35 crc kubenswrapper[4955]: I0217 13:21:35.956736 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/797dd93e-1f77-4edf-a02e-398795307f04-var-run\") pod \"ovn-controller-f8cjz-config-z9pqp\" (UID: \"797dd93e-1f77-4edf-a02e-398795307f04\") " pod="openstack/ovn-controller-f8cjz-config-z9pqp" Feb 17 13:21:35 crc kubenswrapper[4955]: I0217 13:21:35.956795 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/797dd93e-1f77-4edf-a02e-398795307f04-additional-scripts\") pod \"ovn-controller-f8cjz-config-z9pqp\" (UID: \"797dd93e-1f77-4edf-a02e-398795307f04\") " pod="openstack/ovn-controller-f8cjz-config-z9pqp" Feb 17 13:21:35 crc kubenswrapper[4955]: I0217 13:21:35.956854 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/797dd93e-1f77-4edf-a02e-398795307f04-scripts\") pod \"ovn-controller-f8cjz-config-z9pqp\" (UID: \"797dd93e-1f77-4edf-a02e-398795307f04\") " pod="openstack/ovn-controller-f8cjz-config-z9pqp" Feb 17 13:21:35 crc kubenswrapper[4955]: I0217 13:21:35.956887 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/797dd93e-1f77-4edf-a02e-398795307f04-var-log-ovn\") pod \"ovn-controller-f8cjz-config-z9pqp\" (UID: \"797dd93e-1f77-4edf-a02e-398795307f04\") " pod="openstack/ovn-controller-f8cjz-config-z9pqp" Feb 17 13:21:36 crc kubenswrapper[4955]: I0217 13:21:36.058901 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/797dd93e-1f77-4edf-a02e-398795307f04-var-run\") pod \"ovn-controller-f8cjz-config-z9pqp\" (UID: \"797dd93e-1f77-4edf-a02e-398795307f04\") " pod="openstack/ovn-controller-f8cjz-config-z9pqp" Feb 17 13:21:36 crc kubenswrapper[4955]: I0217 13:21:36.058956 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/797dd93e-1f77-4edf-a02e-398795307f04-additional-scripts\") pod \"ovn-controller-f8cjz-config-z9pqp\" (UID: \"797dd93e-1f77-4edf-a02e-398795307f04\") " pod="openstack/ovn-controller-f8cjz-config-z9pqp" Feb 17 13:21:36 crc kubenswrapper[4955]: I0217 13:21:36.059002 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/797dd93e-1f77-4edf-a02e-398795307f04-scripts\") pod \"ovn-controller-f8cjz-config-z9pqp\" (UID: \"797dd93e-1f77-4edf-a02e-398795307f04\") " pod="openstack/ovn-controller-f8cjz-config-z9pqp" Feb 17 13:21:36 crc kubenswrapper[4955]: I0217 13:21:36.059027 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/797dd93e-1f77-4edf-a02e-398795307f04-var-log-ovn\") pod \"ovn-controller-f8cjz-config-z9pqp\" (UID: \"797dd93e-1f77-4edf-a02e-398795307f04\") " pod="openstack/ovn-controller-f8cjz-config-z9pqp" Feb 17 13:21:36 crc kubenswrapper[4955]: I0217 13:21:36.059061 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlwht\" (UniqueName: \"kubernetes.io/projected/797dd93e-1f77-4edf-a02e-398795307f04-kube-api-access-mlwht\") pod \"ovn-controller-f8cjz-config-z9pqp\" (UID: \"797dd93e-1f77-4edf-a02e-398795307f04\") " pod="openstack/ovn-controller-f8cjz-config-z9pqp" Feb 17 13:21:36 crc kubenswrapper[4955]: I0217 13:21:36.059093 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/797dd93e-1f77-4edf-a02e-398795307f04-var-run-ovn\") pod \"ovn-controller-f8cjz-config-z9pqp\" (UID: \"797dd93e-1f77-4edf-a02e-398795307f04\") " pod="openstack/ovn-controller-f8cjz-config-z9pqp" Feb 17 13:21:36 crc kubenswrapper[4955]: I0217 13:21:36.059219 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/797dd93e-1f77-4edf-a02e-398795307f04-var-run-ovn\") pod \"ovn-controller-f8cjz-config-z9pqp\" (UID: \"797dd93e-1f77-4edf-a02e-398795307f04\") " pod="openstack/ovn-controller-f8cjz-config-z9pqp" Feb 17 13:21:36 crc kubenswrapper[4955]: I0217 13:21:36.059231 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/797dd93e-1f77-4edf-a02e-398795307f04-var-run\") pod \"ovn-controller-f8cjz-config-z9pqp\" (UID: \"797dd93e-1f77-4edf-a02e-398795307f04\") " pod="openstack/ovn-controller-f8cjz-config-z9pqp" Feb 17 13:21:36 crc kubenswrapper[4955]: I0217 13:21:36.059303 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/797dd93e-1f77-4edf-a02e-398795307f04-var-log-ovn\") pod \"ovn-controller-f8cjz-config-z9pqp\" (UID: \"797dd93e-1f77-4edf-a02e-398795307f04\") " pod="openstack/ovn-controller-f8cjz-config-z9pqp" Feb 17 13:21:36 crc kubenswrapper[4955]: I0217 13:21:36.060079 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/797dd93e-1f77-4edf-a02e-398795307f04-additional-scripts\") pod \"ovn-controller-f8cjz-config-z9pqp\" (UID: \"797dd93e-1f77-4edf-a02e-398795307f04\") " pod="openstack/ovn-controller-f8cjz-config-z9pqp" Feb 17 13:21:36 crc kubenswrapper[4955]: I0217 13:21:36.062039 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/797dd93e-1f77-4edf-a02e-398795307f04-scripts\") pod \"ovn-controller-f8cjz-config-z9pqp\" (UID: \"797dd93e-1f77-4edf-a02e-398795307f04\") " pod="openstack/ovn-controller-f8cjz-config-z9pqp" Feb 17 13:21:36 crc kubenswrapper[4955]: I0217 13:21:36.084045 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlwht\" (UniqueName: \"kubernetes.io/projected/797dd93e-1f77-4edf-a02e-398795307f04-kube-api-access-mlwht\") pod \"ovn-controller-f8cjz-config-z9pqp\" (UID: \"797dd93e-1f77-4edf-a02e-398795307f04\") " pod="openstack/ovn-controller-f8cjz-config-z9pqp" Feb 17 13:21:36 crc kubenswrapper[4955]: I0217 13:21:36.254520 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-f8cjz-config-z9pqp" Feb 17 13:21:36 crc kubenswrapper[4955]: I0217 13:21:36.415050 4955 generic.go:334] "Generic (PLEG): container finished" podID="34552e31-68c3-4156-addf-37ea6fdf0b6c" containerID="a8a6c6d9a2920a8907fcd1815d0cabffa7db57ffe67f73f31657d130b21005d5" exitCode=0 Feb 17 13:21:36 crc kubenswrapper[4955]: I0217 13:21:36.415118 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"34552e31-68c3-4156-addf-37ea6fdf0b6c","Type":"ContainerDied","Data":"a8a6c6d9a2920a8907fcd1815d0cabffa7db57ffe67f73f31657d130b21005d5"} Feb 17 13:21:36 crc kubenswrapper[4955]: I0217 13:21:36.416998 4955 generic.go:334] "Generic (PLEG): container finished" podID="c241677c-9f24-4cc0-8d15-e166174e4cd0" containerID="cd19a92beba58ac2a112c23c7364fb146fdc4af6d6854544bdce5d7e774494a2" exitCode=0 Feb 17 13:21:36 crc kubenswrapper[4955]: I0217 13:21:36.417057 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c241677c-9f24-4cc0-8d15-e166174e4cd0","Type":"ContainerDied","Data":"cd19a92beba58ac2a112c23c7364fb146fdc4af6d6854544bdce5d7e774494a2"} Feb 17 13:21:40 crc kubenswrapper[4955]: I0217 13:21:40.524172 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-f8cjz" podUID="452bb98a-815b-4154-84c5-6b31338fca29" containerName="ovn-controller" probeResult="failure" output=< Feb 17 13:21:40 crc kubenswrapper[4955]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Feb 17 13:21:40 crc kubenswrapper[4955]: > Feb 17 13:21:42 crc kubenswrapper[4955]: I0217 13:21:42.283211 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-f8cjz-config-z9pqp"] Feb 17 13:21:42 crc kubenswrapper[4955]: W0217 13:21:42.293356 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod797dd93e_1f77_4edf_a02e_398795307f04.slice/crio-1277a30060293d6c24091a2eba6db55289c2cfbb165d20777f0c6b196b9ce008 WatchSource:0}: Error finding container 1277a30060293d6c24091a2eba6db55289c2cfbb165d20777f0c6b196b9ce008: Status 404 returned error can't find the container with id 1277a30060293d6c24091a2eba6db55289c2cfbb165d20777f0c6b196b9ce008 Feb 17 13:21:42 crc kubenswrapper[4955]: I0217 13:21:42.508713 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-f8cjz-config-z9pqp" event={"ID":"797dd93e-1f77-4edf-a02e-398795307f04","Type":"ContainerStarted","Data":"1277a30060293d6c24091a2eba6db55289c2cfbb165d20777f0c6b196b9ce008"} Feb 17 13:21:42 crc kubenswrapper[4955]: I0217 13:21:42.511625 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c241677c-9f24-4cc0-8d15-e166174e4cd0","Type":"ContainerStarted","Data":"30e8197d8a6e3cdbe4f6414abf7e655a8860a1ac94d7c201ed6d3f58a7f1ecf2"} Feb 17 13:21:42 crc kubenswrapper[4955]: I0217 13:21:42.512701 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:21:42 crc kubenswrapper[4955]: I0217 13:21:42.516677 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"34552e31-68c3-4156-addf-37ea6fdf0b6c","Type":"ContainerStarted","Data":"f07eeaceb5a628e2e825faa765c2135da249f0bd9b80274d2fdeafb27d148d57"} Feb 17 13:21:42 crc kubenswrapper[4955]: I0217 13:21:42.517429 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Feb 17 13:21:42 crc kubenswrapper[4955]: I0217 13:21:42.540023 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=56.257244859 podStartE2EDuration="1m6.540003019s" podCreationTimestamp="2026-02-17 13:20:36 +0000 UTC" firstStartedPulling="2026-02-17 13:20:51.22546158 +0000 UTC m=+989.748191113" lastFinishedPulling="2026-02-17 13:21:01.50821974 +0000 UTC m=+1000.030949273" observedRunningTime="2026-02-17 13:21:42.539056082 +0000 UTC m=+1041.061785635" watchObservedRunningTime="2026-02-17 13:21:42.540003019 +0000 UTC m=+1041.062732562" Feb 17 13:21:42 crc kubenswrapper[4955]: I0217 13:21:42.557310 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=56.74543245 podStartE2EDuration="1m7.557292732s" podCreationTimestamp="2026-02-17 13:20:35 +0000 UTC" firstStartedPulling="2026-02-17 13:20:50.949503059 +0000 UTC m=+989.472232622" lastFinishedPulling="2026-02-17 13:21:01.761363361 +0000 UTC m=+1000.284092904" observedRunningTime="2026-02-17 13:21:42.556797118 +0000 UTC m=+1041.079526671" watchObservedRunningTime="2026-02-17 13:21:42.557292732 +0000 UTC m=+1041.080022275" Feb 17 13:21:43 crc kubenswrapper[4955]: I0217 13:21:43.526125 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-jwz6b" event={"ID":"aed5c07b-3d7a-43c1-a11f-4a4220289bef","Type":"ContainerStarted","Data":"b171b99fad766be9125f2dc416c5ba69a27c46624d56bc5d39d482c4c9459ea1"} Feb 17 13:21:43 crc kubenswrapper[4955]: I0217 13:21:43.529192 4955 generic.go:334] "Generic (PLEG): container finished" podID="797dd93e-1f77-4edf-a02e-398795307f04" containerID="86273f17c38a0753b27e85668517d95a43d5eda6373742707eac860fe4973011" exitCode=0 Feb 17 13:21:43 crc kubenswrapper[4955]: I0217 13:21:43.529851 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-f8cjz-config-z9pqp" event={"ID":"797dd93e-1f77-4edf-a02e-398795307f04","Type":"ContainerDied","Data":"86273f17c38a0753b27e85668517d95a43d5eda6373742707eac860fe4973011"} Feb 17 13:21:43 crc kubenswrapper[4955]: I0217 13:21:43.544962 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-jwz6b" podStartSLOduration=3.305341555 podStartE2EDuration="13.544946984s" podCreationTimestamp="2026-02-17 13:21:30 +0000 UTC" firstStartedPulling="2026-02-17 13:21:31.727716655 +0000 UTC m=+1030.250446198" lastFinishedPulling="2026-02-17 13:21:41.967322094 +0000 UTC m=+1040.490051627" observedRunningTime="2026-02-17 13:21:43.543816912 +0000 UTC m=+1042.066546465" watchObservedRunningTime="2026-02-17 13:21:43.544946984 +0000 UTC m=+1042.067676527" Feb 17 13:21:44 crc kubenswrapper[4955]: I0217 13:21:44.282013 4955 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod701369d5-716b-4e3d-b3ad-c896ffba2f49"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod701369d5-716b-4e3d-b3ad-c896ffba2f49] : Timed out while waiting for systemd to remove kubepods-besteffort-pod701369d5_716b_4e3d_b3ad_c896ffba2f49.slice" Feb 17 13:21:44 crc kubenswrapper[4955]: E0217 13:21:44.282074 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pod701369d5-716b-4e3d-b3ad-c896ffba2f49] : unable to destroy cgroup paths for cgroup [kubepods besteffort pod701369d5-716b-4e3d-b3ad-c896ffba2f49] : Timed out while waiting for systemd to remove kubepods-besteffort-pod701369d5_716b_4e3d_b3ad_c896ffba2f49.slice" pod="openstack/dnsmasq-dns-5bf47b49b7-snn92" podUID="701369d5-716b-4e3d-b3ad-c896ffba2f49" Feb 17 13:21:44 crc kubenswrapper[4955]: I0217 13:21:44.536999 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-snn92" Feb 17 13:21:44 crc kubenswrapper[4955]: I0217 13:21:44.592424 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-snn92"] Feb 17 13:21:44 crc kubenswrapper[4955]: I0217 13:21:44.602356 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-snn92"] Feb 17 13:21:44 crc kubenswrapper[4955]: I0217 13:21:44.897598 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-f8cjz-config-z9pqp" Feb 17 13:21:45 crc kubenswrapper[4955]: I0217 13:21:45.001653 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/797dd93e-1f77-4edf-a02e-398795307f04-var-run-ovn\") pod \"797dd93e-1f77-4edf-a02e-398795307f04\" (UID: \"797dd93e-1f77-4edf-a02e-398795307f04\") " Feb 17 13:21:45 crc kubenswrapper[4955]: I0217 13:21:45.001741 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mlwht\" (UniqueName: \"kubernetes.io/projected/797dd93e-1f77-4edf-a02e-398795307f04-kube-api-access-mlwht\") pod \"797dd93e-1f77-4edf-a02e-398795307f04\" (UID: \"797dd93e-1f77-4edf-a02e-398795307f04\") " Feb 17 13:21:45 crc kubenswrapper[4955]: I0217 13:21:45.001749 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/797dd93e-1f77-4edf-a02e-398795307f04-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "797dd93e-1f77-4edf-a02e-398795307f04" (UID: "797dd93e-1f77-4edf-a02e-398795307f04"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 13:21:45 crc kubenswrapper[4955]: I0217 13:21:45.001837 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/797dd93e-1f77-4edf-a02e-398795307f04-scripts\") pod \"797dd93e-1f77-4edf-a02e-398795307f04\" (UID: \"797dd93e-1f77-4edf-a02e-398795307f04\") " Feb 17 13:21:45 crc kubenswrapper[4955]: I0217 13:21:45.001853 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/797dd93e-1f77-4edf-a02e-398795307f04-var-log-ovn\") pod \"797dd93e-1f77-4edf-a02e-398795307f04\" (UID: \"797dd93e-1f77-4edf-a02e-398795307f04\") " Feb 17 13:21:45 crc kubenswrapper[4955]: I0217 13:21:45.001927 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/797dd93e-1f77-4edf-a02e-398795307f04-additional-scripts\") pod \"797dd93e-1f77-4edf-a02e-398795307f04\" (UID: \"797dd93e-1f77-4edf-a02e-398795307f04\") " Feb 17 13:21:45 crc kubenswrapper[4955]: I0217 13:21:45.001957 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/797dd93e-1f77-4edf-a02e-398795307f04-var-run\") pod \"797dd93e-1f77-4edf-a02e-398795307f04\" (UID: \"797dd93e-1f77-4edf-a02e-398795307f04\") " Feb 17 13:21:45 crc kubenswrapper[4955]: I0217 13:21:45.002052 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/797dd93e-1f77-4edf-a02e-398795307f04-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "797dd93e-1f77-4edf-a02e-398795307f04" (UID: "797dd93e-1f77-4edf-a02e-398795307f04"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 13:21:45 crc kubenswrapper[4955]: I0217 13:21:45.002352 4955 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/797dd93e-1f77-4edf-a02e-398795307f04-var-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:45 crc kubenswrapper[4955]: I0217 13:21:45.002371 4955 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/797dd93e-1f77-4edf-a02e-398795307f04-var-log-ovn\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:45 crc kubenswrapper[4955]: I0217 13:21:45.002370 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/797dd93e-1f77-4edf-a02e-398795307f04-var-run" (OuterVolumeSpecName: "var-run") pod "797dd93e-1f77-4edf-a02e-398795307f04" (UID: "797dd93e-1f77-4edf-a02e-398795307f04"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 13:21:45 crc kubenswrapper[4955]: I0217 13:21:45.002696 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/797dd93e-1f77-4edf-a02e-398795307f04-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "797dd93e-1f77-4edf-a02e-398795307f04" (UID: "797dd93e-1f77-4edf-a02e-398795307f04"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:21:45 crc kubenswrapper[4955]: I0217 13:21:45.003592 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/797dd93e-1f77-4edf-a02e-398795307f04-scripts" (OuterVolumeSpecName: "scripts") pod "797dd93e-1f77-4edf-a02e-398795307f04" (UID: "797dd93e-1f77-4edf-a02e-398795307f04"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:21:45 crc kubenswrapper[4955]: I0217 13:21:45.017208 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/797dd93e-1f77-4edf-a02e-398795307f04-kube-api-access-mlwht" (OuterVolumeSpecName: "kube-api-access-mlwht") pod "797dd93e-1f77-4edf-a02e-398795307f04" (UID: "797dd93e-1f77-4edf-a02e-398795307f04"). InnerVolumeSpecName "kube-api-access-mlwht". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:21:45 crc kubenswrapper[4955]: I0217 13:21:45.103855 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mlwht\" (UniqueName: \"kubernetes.io/projected/797dd93e-1f77-4edf-a02e-398795307f04-kube-api-access-mlwht\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:45 crc kubenswrapper[4955]: I0217 13:21:45.103892 4955 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/797dd93e-1f77-4edf-a02e-398795307f04-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:45 crc kubenswrapper[4955]: I0217 13:21:45.103904 4955 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/797dd93e-1f77-4edf-a02e-398795307f04-additional-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:45 crc kubenswrapper[4955]: I0217 13:21:45.103915 4955 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/797dd93e-1f77-4edf-a02e-398795307f04-var-run\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:45 crc kubenswrapper[4955]: I0217 13:21:45.524988 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-f8cjz" Feb 17 13:21:45 crc kubenswrapper[4955]: I0217 13:21:45.544563 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-f8cjz-config-z9pqp" event={"ID":"797dd93e-1f77-4edf-a02e-398795307f04","Type":"ContainerDied","Data":"1277a30060293d6c24091a2eba6db55289c2cfbb165d20777f0c6b196b9ce008"} Feb 17 13:21:45 crc kubenswrapper[4955]: I0217 13:21:45.544609 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1277a30060293d6c24091a2eba6db55289c2cfbb165d20777f0c6b196b9ce008" Feb 17 13:21:45 crc kubenswrapper[4955]: I0217 13:21:45.544672 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-f8cjz-config-z9pqp" Feb 17 13:21:46 crc kubenswrapper[4955]: I0217 13:21:46.011575 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-f8cjz-config-z9pqp"] Feb 17 13:21:46 crc kubenswrapper[4955]: I0217 13:21:46.019342 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-f8cjz-config-z9pqp"] Feb 17 13:21:46 crc kubenswrapper[4955]: I0217 13:21:46.020060 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/fb6d9434-4476-438b-90e4-4781bd85b0bb-etc-swift\") pod \"swift-storage-0\" (UID: \"fb6d9434-4476-438b-90e4-4781bd85b0bb\") " pod="openstack/swift-storage-0" Feb 17 13:21:46 crc kubenswrapper[4955]: I0217 13:21:46.034580 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/fb6d9434-4476-438b-90e4-4781bd85b0bb-etc-swift\") pod \"swift-storage-0\" (UID: \"fb6d9434-4476-438b-90e4-4781bd85b0bb\") " pod="openstack/swift-storage-0" Feb 17 13:21:46 crc kubenswrapper[4955]: I0217 13:21:46.068878 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Feb 17 13:21:46 crc kubenswrapper[4955]: I0217 13:21:46.289048 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="701369d5-716b-4e3d-b3ad-c896ffba2f49" path="/var/lib/kubelet/pods/701369d5-716b-4e3d-b3ad-c896ffba2f49/volumes" Feb 17 13:21:46 crc kubenswrapper[4955]: I0217 13:21:46.291208 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="797dd93e-1f77-4edf-a02e-398795307f04" path="/var/lib/kubelet/pods/797dd93e-1f77-4edf-a02e-398795307f04/volumes" Feb 17 13:21:46 crc kubenswrapper[4955]: I0217 13:21:46.596809 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Feb 17 13:21:47 crc kubenswrapper[4955]: I0217 13:21:47.559721 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"fb6d9434-4476-438b-90e4-4781bd85b0bb","Type":"ContainerStarted","Data":"4125bc804f2294934a165109a38948ce7e9d9eee750f8ccf487bf43e08c8599b"} Feb 17 13:21:48 crc kubenswrapper[4955]: I0217 13:21:48.568173 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"fb6d9434-4476-438b-90e4-4781bd85b0bb","Type":"ContainerStarted","Data":"a06c2a23f3a0b36087406231206d33efc38c85831112f0fd3e7f8f6c547dcc45"} Feb 17 13:21:48 crc kubenswrapper[4955]: I0217 13:21:48.568504 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"fb6d9434-4476-438b-90e4-4781bd85b0bb","Type":"ContainerStarted","Data":"d1d5ce61577b815393291cb072f725323e6837ae0ddbadbba4d9ba8fa7d01c84"} Feb 17 13:21:48 crc kubenswrapper[4955]: I0217 13:21:48.568517 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"fb6d9434-4476-438b-90e4-4781bd85b0bb","Type":"ContainerStarted","Data":"92b8790d76ce0e4351ad47f06f085b4971874dfdc4626e3991542b7b1bfbecf8"} Feb 17 13:21:48 crc kubenswrapper[4955]: I0217 13:21:48.568527 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"fb6d9434-4476-438b-90e4-4781bd85b0bb","Type":"ContainerStarted","Data":"c766f6b9b76c2070b31e732c5b832832d80f9497873a3a4f610e141d9e48633f"} Feb 17 13:21:49 crc kubenswrapper[4955]: I0217 13:21:49.577017 4955 generic.go:334] "Generic (PLEG): container finished" podID="aed5c07b-3d7a-43c1-a11f-4a4220289bef" containerID="b171b99fad766be9125f2dc416c5ba69a27c46624d56bc5d39d482c4c9459ea1" exitCode=0 Feb 17 13:21:49 crc kubenswrapper[4955]: I0217 13:21:49.577153 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-jwz6b" event={"ID":"aed5c07b-3d7a-43c1-a11f-4a4220289bef","Type":"ContainerDied","Data":"b171b99fad766be9125f2dc416c5ba69a27c46624d56bc5d39d482c4c9459ea1"} Feb 17 13:21:50 crc kubenswrapper[4955]: I0217 13:21:50.594922 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"fb6d9434-4476-438b-90e4-4781bd85b0bb","Type":"ContainerStarted","Data":"9d0f92b1aaa317eddf55a3e5ad17c47a2b31ab90cf3490fc8a894309ac5ef6f7"} Feb 17 13:21:50 crc kubenswrapper[4955]: I0217 13:21:50.595499 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"fb6d9434-4476-438b-90e4-4781bd85b0bb","Type":"ContainerStarted","Data":"6d8a4ee8ed50b0640b2cdaae0f460eaaf63a9bdc6bdc0e215d6579e2a38a87d6"} Feb 17 13:21:50 crc kubenswrapper[4955]: I0217 13:21:50.595512 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"fb6d9434-4476-438b-90e4-4781bd85b0bb","Type":"ContainerStarted","Data":"786c4bb47cd9de21bf96387b125e7c405a730ed076207ac3747e94d6c295897e"} Feb 17 13:21:50 crc kubenswrapper[4955]: I0217 13:21:50.595521 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"fb6d9434-4476-438b-90e4-4781bd85b0bb","Type":"ContainerStarted","Data":"d6badc02ddaf4a1fe1b7d0d9576acea540943ab37de29d1ecdc651ba34d5f50d"} Feb 17 13:21:50 crc kubenswrapper[4955]: I0217 13:21:50.979109 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-jwz6b" Feb 17 13:21:51 crc kubenswrapper[4955]: I0217 13:21:51.143477 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aed5c07b-3d7a-43c1-a11f-4a4220289bef-combined-ca-bundle\") pod \"aed5c07b-3d7a-43c1-a11f-4a4220289bef\" (UID: \"aed5c07b-3d7a-43c1-a11f-4a4220289bef\") " Feb 17 13:21:51 crc kubenswrapper[4955]: I0217 13:21:51.144023 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/aed5c07b-3d7a-43c1-a11f-4a4220289bef-db-sync-config-data\") pod \"aed5c07b-3d7a-43c1-a11f-4a4220289bef\" (UID: \"aed5c07b-3d7a-43c1-a11f-4a4220289bef\") " Feb 17 13:21:51 crc kubenswrapper[4955]: I0217 13:21:51.144067 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aed5c07b-3d7a-43c1-a11f-4a4220289bef-config-data\") pod \"aed5c07b-3d7a-43c1-a11f-4a4220289bef\" (UID: \"aed5c07b-3d7a-43c1-a11f-4a4220289bef\") " Feb 17 13:21:51 crc kubenswrapper[4955]: I0217 13:21:51.144144 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xpsh\" (UniqueName: \"kubernetes.io/projected/aed5c07b-3d7a-43c1-a11f-4a4220289bef-kube-api-access-2xpsh\") pod \"aed5c07b-3d7a-43c1-a11f-4a4220289bef\" (UID: \"aed5c07b-3d7a-43c1-a11f-4a4220289bef\") " Feb 17 13:21:51 crc kubenswrapper[4955]: I0217 13:21:51.149000 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aed5c07b-3d7a-43c1-a11f-4a4220289bef-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "aed5c07b-3d7a-43c1-a11f-4a4220289bef" (UID: "aed5c07b-3d7a-43c1-a11f-4a4220289bef"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:21:51 crc kubenswrapper[4955]: I0217 13:21:51.159124 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aed5c07b-3d7a-43c1-a11f-4a4220289bef-kube-api-access-2xpsh" (OuterVolumeSpecName: "kube-api-access-2xpsh") pod "aed5c07b-3d7a-43c1-a11f-4a4220289bef" (UID: "aed5c07b-3d7a-43c1-a11f-4a4220289bef"). InnerVolumeSpecName "kube-api-access-2xpsh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:21:51 crc kubenswrapper[4955]: I0217 13:21:51.165968 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aed5c07b-3d7a-43c1-a11f-4a4220289bef-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aed5c07b-3d7a-43c1-a11f-4a4220289bef" (UID: "aed5c07b-3d7a-43c1-a11f-4a4220289bef"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:21:51 crc kubenswrapper[4955]: I0217 13:21:51.195034 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aed5c07b-3d7a-43c1-a11f-4a4220289bef-config-data" (OuterVolumeSpecName: "config-data") pod "aed5c07b-3d7a-43c1-a11f-4a4220289bef" (UID: "aed5c07b-3d7a-43c1-a11f-4a4220289bef"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:21:51 crc kubenswrapper[4955]: I0217 13:21:51.252399 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aed5c07b-3d7a-43c1-a11f-4a4220289bef-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:51 crc kubenswrapper[4955]: I0217 13:21:51.252444 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xpsh\" (UniqueName: \"kubernetes.io/projected/aed5c07b-3d7a-43c1-a11f-4a4220289bef-kube-api-access-2xpsh\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:51 crc kubenswrapper[4955]: I0217 13:21:51.252466 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aed5c07b-3d7a-43c1-a11f-4a4220289bef-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:51 crc kubenswrapper[4955]: I0217 13:21:51.252478 4955 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/aed5c07b-3d7a-43c1-a11f-4a4220289bef-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:51 crc kubenswrapper[4955]: I0217 13:21:51.611244 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"fb6d9434-4476-438b-90e4-4781bd85b0bb","Type":"ContainerStarted","Data":"e56e6926e9a5abc99143cf7952b4b86a574190d6b8a41b9ca87bf548aafa4106"} Feb 17 13:21:51 crc kubenswrapper[4955]: I0217 13:21:51.613345 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-jwz6b" event={"ID":"aed5c07b-3d7a-43c1-a11f-4a4220289bef","Type":"ContainerDied","Data":"6b7e8abcbf0d6f84ebbc3b884eafa2b0b4be79cc92dca2ec4808cb08a9d447d8"} Feb 17 13:21:51 crc kubenswrapper[4955]: I0217 13:21:51.613378 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6b7e8abcbf0d6f84ebbc3b884eafa2b0b4be79cc92dca2ec4808cb08a9d447d8" Feb 17 13:21:51 crc kubenswrapper[4955]: I0217 13:21:51.613456 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-jwz6b" Feb 17 13:21:52 crc kubenswrapper[4955]: I0217 13:21:52.071327 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74dc88fc-bt2xw"] Feb 17 13:21:52 crc kubenswrapper[4955]: E0217 13:21:52.071868 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="797dd93e-1f77-4edf-a02e-398795307f04" containerName="ovn-config" Feb 17 13:21:52 crc kubenswrapper[4955]: I0217 13:21:52.071880 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="797dd93e-1f77-4edf-a02e-398795307f04" containerName="ovn-config" Feb 17 13:21:52 crc kubenswrapper[4955]: E0217 13:21:52.071899 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aed5c07b-3d7a-43c1-a11f-4a4220289bef" containerName="glance-db-sync" Feb 17 13:21:52 crc kubenswrapper[4955]: I0217 13:21:52.071904 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="aed5c07b-3d7a-43c1-a11f-4a4220289bef" containerName="glance-db-sync" Feb 17 13:21:52 crc kubenswrapper[4955]: I0217 13:21:52.072053 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="aed5c07b-3d7a-43c1-a11f-4a4220289bef" containerName="glance-db-sync" Feb 17 13:21:52 crc kubenswrapper[4955]: I0217 13:21:52.072066 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="797dd93e-1f77-4edf-a02e-398795307f04" containerName="ovn-config" Feb 17 13:21:52 crc kubenswrapper[4955]: I0217 13:21:52.080999 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74dc88fc-bt2xw" Feb 17 13:21:52 crc kubenswrapper[4955]: I0217 13:21:52.099436 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74dc88fc-bt2xw"] Feb 17 13:21:52 crc kubenswrapper[4955]: I0217 13:21:52.171861 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7dfb2627-eb1d-44f0-b4c2-eea0b404b170-ovsdbserver-sb\") pod \"dnsmasq-dns-74dc88fc-bt2xw\" (UID: \"7dfb2627-eb1d-44f0-b4c2-eea0b404b170\") " pod="openstack/dnsmasq-dns-74dc88fc-bt2xw" Feb 17 13:21:52 crc kubenswrapper[4955]: I0217 13:21:52.171919 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bh6px\" (UniqueName: \"kubernetes.io/projected/7dfb2627-eb1d-44f0-b4c2-eea0b404b170-kube-api-access-bh6px\") pod \"dnsmasq-dns-74dc88fc-bt2xw\" (UID: \"7dfb2627-eb1d-44f0-b4c2-eea0b404b170\") " pod="openstack/dnsmasq-dns-74dc88fc-bt2xw" Feb 17 13:21:52 crc kubenswrapper[4955]: I0217 13:21:52.171990 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7dfb2627-eb1d-44f0-b4c2-eea0b404b170-dns-svc\") pod \"dnsmasq-dns-74dc88fc-bt2xw\" (UID: \"7dfb2627-eb1d-44f0-b4c2-eea0b404b170\") " pod="openstack/dnsmasq-dns-74dc88fc-bt2xw" Feb 17 13:21:52 crc kubenswrapper[4955]: I0217 13:21:52.172047 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7dfb2627-eb1d-44f0-b4c2-eea0b404b170-ovsdbserver-nb\") pod \"dnsmasq-dns-74dc88fc-bt2xw\" (UID: \"7dfb2627-eb1d-44f0-b4c2-eea0b404b170\") " pod="openstack/dnsmasq-dns-74dc88fc-bt2xw" Feb 17 13:21:52 crc kubenswrapper[4955]: I0217 13:21:52.172092 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7dfb2627-eb1d-44f0-b4c2-eea0b404b170-config\") pod \"dnsmasq-dns-74dc88fc-bt2xw\" (UID: \"7dfb2627-eb1d-44f0-b4c2-eea0b404b170\") " pod="openstack/dnsmasq-dns-74dc88fc-bt2xw" Feb 17 13:21:52 crc kubenswrapper[4955]: I0217 13:21:52.273324 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7dfb2627-eb1d-44f0-b4c2-eea0b404b170-ovsdbserver-sb\") pod \"dnsmasq-dns-74dc88fc-bt2xw\" (UID: \"7dfb2627-eb1d-44f0-b4c2-eea0b404b170\") " pod="openstack/dnsmasq-dns-74dc88fc-bt2xw" Feb 17 13:21:52 crc kubenswrapper[4955]: I0217 13:21:52.273394 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bh6px\" (UniqueName: \"kubernetes.io/projected/7dfb2627-eb1d-44f0-b4c2-eea0b404b170-kube-api-access-bh6px\") pod \"dnsmasq-dns-74dc88fc-bt2xw\" (UID: \"7dfb2627-eb1d-44f0-b4c2-eea0b404b170\") " pod="openstack/dnsmasq-dns-74dc88fc-bt2xw" Feb 17 13:21:52 crc kubenswrapper[4955]: I0217 13:21:52.273473 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7dfb2627-eb1d-44f0-b4c2-eea0b404b170-dns-svc\") pod \"dnsmasq-dns-74dc88fc-bt2xw\" (UID: \"7dfb2627-eb1d-44f0-b4c2-eea0b404b170\") " pod="openstack/dnsmasq-dns-74dc88fc-bt2xw" Feb 17 13:21:52 crc kubenswrapper[4955]: I0217 13:21:52.273529 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7dfb2627-eb1d-44f0-b4c2-eea0b404b170-ovsdbserver-nb\") pod \"dnsmasq-dns-74dc88fc-bt2xw\" (UID: \"7dfb2627-eb1d-44f0-b4c2-eea0b404b170\") " pod="openstack/dnsmasq-dns-74dc88fc-bt2xw" Feb 17 13:21:52 crc kubenswrapper[4955]: I0217 13:21:52.273571 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7dfb2627-eb1d-44f0-b4c2-eea0b404b170-config\") pod \"dnsmasq-dns-74dc88fc-bt2xw\" (UID: \"7dfb2627-eb1d-44f0-b4c2-eea0b404b170\") " pod="openstack/dnsmasq-dns-74dc88fc-bt2xw" Feb 17 13:21:52 crc kubenswrapper[4955]: I0217 13:21:52.274432 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7dfb2627-eb1d-44f0-b4c2-eea0b404b170-ovsdbserver-sb\") pod \"dnsmasq-dns-74dc88fc-bt2xw\" (UID: \"7dfb2627-eb1d-44f0-b4c2-eea0b404b170\") " pod="openstack/dnsmasq-dns-74dc88fc-bt2xw" Feb 17 13:21:52 crc kubenswrapper[4955]: I0217 13:21:52.274531 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7dfb2627-eb1d-44f0-b4c2-eea0b404b170-dns-svc\") pod \"dnsmasq-dns-74dc88fc-bt2xw\" (UID: \"7dfb2627-eb1d-44f0-b4c2-eea0b404b170\") " pod="openstack/dnsmasq-dns-74dc88fc-bt2xw" Feb 17 13:21:52 crc kubenswrapper[4955]: I0217 13:21:52.274691 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7dfb2627-eb1d-44f0-b4c2-eea0b404b170-config\") pod \"dnsmasq-dns-74dc88fc-bt2xw\" (UID: \"7dfb2627-eb1d-44f0-b4c2-eea0b404b170\") " pod="openstack/dnsmasq-dns-74dc88fc-bt2xw" Feb 17 13:21:52 crc kubenswrapper[4955]: I0217 13:21:52.274821 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7dfb2627-eb1d-44f0-b4c2-eea0b404b170-ovsdbserver-nb\") pod \"dnsmasq-dns-74dc88fc-bt2xw\" (UID: \"7dfb2627-eb1d-44f0-b4c2-eea0b404b170\") " pod="openstack/dnsmasq-dns-74dc88fc-bt2xw" Feb 17 13:21:52 crc kubenswrapper[4955]: I0217 13:21:52.309590 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bh6px\" (UniqueName: \"kubernetes.io/projected/7dfb2627-eb1d-44f0-b4c2-eea0b404b170-kube-api-access-bh6px\") pod \"dnsmasq-dns-74dc88fc-bt2xw\" (UID: \"7dfb2627-eb1d-44f0-b4c2-eea0b404b170\") " pod="openstack/dnsmasq-dns-74dc88fc-bt2xw" Feb 17 13:21:52 crc kubenswrapper[4955]: I0217 13:21:52.420294 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74dc88fc-bt2xw" Feb 17 13:21:52 crc kubenswrapper[4955]: I0217 13:21:52.646997 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"fb6d9434-4476-438b-90e4-4781bd85b0bb","Type":"ContainerStarted","Data":"1b9437656e3502cfa911418e4c15fc054586564edd5ccd2581820035552071f9"} Feb 17 13:21:52 crc kubenswrapper[4955]: I0217 13:21:52.647289 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"fb6d9434-4476-438b-90e4-4781bd85b0bb","Type":"ContainerStarted","Data":"b621f8351be80804d207e7204f9f351a8d7eaf27cefc93f0e139823a39ba1eb8"} Feb 17 13:21:52 crc kubenswrapper[4955]: I0217 13:21:52.647302 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"fb6d9434-4476-438b-90e4-4781bd85b0bb","Type":"ContainerStarted","Data":"38bd5fc2b324288b2ba89a4205c5aa3b4ab5bbff0780f411f15c79d05c68f976"} Feb 17 13:21:52 crc kubenswrapper[4955]: I0217 13:21:52.874333 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74dc88fc-bt2xw"] Feb 17 13:21:53 crc kubenswrapper[4955]: I0217 13:21:53.667449 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"fb6d9434-4476-438b-90e4-4781bd85b0bb","Type":"ContainerStarted","Data":"5a55d029d9efc0a6d784f8a5c75e576dcddf86914473fc0821323343e93a2339"} Feb 17 13:21:53 crc kubenswrapper[4955]: I0217 13:21:53.667954 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"fb6d9434-4476-438b-90e4-4781bd85b0bb","Type":"ContainerStarted","Data":"479ceec35ba014af56adbeba7fdc4b731dd56a34b25a085f67caf94d57948cee"} Feb 17 13:21:53 crc kubenswrapper[4955]: I0217 13:21:53.667986 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"fb6d9434-4476-438b-90e4-4781bd85b0bb","Type":"ContainerStarted","Data":"87ffb13f4517ab5f1c5cf0c8ee355fc3b3c2ceb7204dc5cfaa1520db0055bbe4"} Feb 17 13:21:53 crc kubenswrapper[4955]: I0217 13:21:53.673453 4955 generic.go:334] "Generic (PLEG): container finished" podID="7dfb2627-eb1d-44f0-b4c2-eea0b404b170" containerID="cb5afab742c34169ec27e429061b294a63fd6fcbb3ffe8316270f25b74ad99c3" exitCode=0 Feb 17 13:21:53 crc kubenswrapper[4955]: I0217 13:21:53.673643 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74dc88fc-bt2xw" event={"ID":"7dfb2627-eb1d-44f0-b4c2-eea0b404b170","Type":"ContainerDied","Data":"cb5afab742c34169ec27e429061b294a63fd6fcbb3ffe8316270f25b74ad99c3"} Feb 17 13:21:53 crc kubenswrapper[4955]: I0217 13:21:53.674456 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74dc88fc-bt2xw" event={"ID":"7dfb2627-eb1d-44f0-b4c2-eea0b404b170","Type":"ContainerStarted","Data":"11952759f507b9214f2ff285de21a12e6040b25e33a5b08bd74f0079a17abb6e"} Feb 17 13:21:53 crc kubenswrapper[4955]: I0217 13:21:53.719738 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=36.908672952 podStartE2EDuration="41.719709934s" podCreationTimestamp="2026-02-17 13:21:12 +0000 UTC" firstStartedPulling="2026-02-17 13:21:46.610647312 +0000 UTC m=+1045.133376855" lastFinishedPulling="2026-02-17 13:21:51.421684294 +0000 UTC m=+1049.944413837" observedRunningTime="2026-02-17 13:21:53.709698999 +0000 UTC m=+1052.232428572" watchObservedRunningTime="2026-02-17 13:21:53.719709934 +0000 UTC m=+1052.242439487" Feb 17 13:21:53 crc kubenswrapper[4955]: I0217 13:21:53.986717 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74dc88fc-bt2xw"] Feb 17 13:21:54 crc kubenswrapper[4955]: I0217 13:21:54.014925 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-xxclj"] Feb 17 13:21:54 crc kubenswrapper[4955]: I0217 13:21:54.016140 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-xxclj" Feb 17 13:21:54 crc kubenswrapper[4955]: I0217 13:21:54.020180 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Feb 17 13:21:54 crc kubenswrapper[4955]: I0217 13:21:54.033649 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-xxclj"] Feb 17 13:21:54 crc kubenswrapper[4955]: I0217 13:21:54.111110 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-xxclj\" (UID: \"7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a\") " pod="openstack/dnsmasq-dns-5f59b8f679-xxclj" Feb 17 13:21:54 crc kubenswrapper[4955]: I0217 13:21:54.111152 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-xxclj\" (UID: \"7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a\") " pod="openstack/dnsmasq-dns-5f59b8f679-xxclj" Feb 17 13:21:54 crc kubenswrapper[4955]: I0217 13:21:54.111293 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlkwb\" (UniqueName: \"kubernetes.io/projected/7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a-kube-api-access-vlkwb\") pod \"dnsmasq-dns-5f59b8f679-xxclj\" (UID: \"7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a\") " pod="openstack/dnsmasq-dns-5f59b8f679-xxclj" Feb 17 13:21:54 crc kubenswrapper[4955]: I0217 13:21:54.111356 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-xxclj\" (UID: \"7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a\") " pod="openstack/dnsmasq-dns-5f59b8f679-xxclj" Feb 17 13:21:54 crc kubenswrapper[4955]: I0217 13:21:54.111399 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a-config\") pod \"dnsmasq-dns-5f59b8f679-xxclj\" (UID: \"7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a\") " pod="openstack/dnsmasq-dns-5f59b8f679-xxclj" Feb 17 13:21:54 crc kubenswrapper[4955]: I0217 13:21:54.111446 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-xxclj\" (UID: \"7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a\") " pod="openstack/dnsmasq-dns-5f59b8f679-xxclj" Feb 17 13:21:54 crc kubenswrapper[4955]: I0217 13:21:54.212943 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-xxclj\" (UID: \"7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a\") " pod="openstack/dnsmasq-dns-5f59b8f679-xxclj" Feb 17 13:21:54 crc kubenswrapper[4955]: I0217 13:21:54.213005 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlkwb\" (UniqueName: \"kubernetes.io/projected/7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a-kube-api-access-vlkwb\") pod \"dnsmasq-dns-5f59b8f679-xxclj\" (UID: \"7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a\") " pod="openstack/dnsmasq-dns-5f59b8f679-xxclj" Feb 17 13:21:54 crc kubenswrapper[4955]: I0217 13:21:54.213050 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-xxclj\" (UID: \"7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a\") " pod="openstack/dnsmasq-dns-5f59b8f679-xxclj" Feb 17 13:21:54 crc kubenswrapper[4955]: I0217 13:21:54.213093 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a-config\") pod \"dnsmasq-dns-5f59b8f679-xxclj\" (UID: \"7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a\") " pod="openstack/dnsmasq-dns-5f59b8f679-xxclj" Feb 17 13:21:54 crc kubenswrapper[4955]: I0217 13:21:54.213140 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-xxclj\" (UID: \"7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a\") " pod="openstack/dnsmasq-dns-5f59b8f679-xxclj" Feb 17 13:21:54 crc kubenswrapper[4955]: I0217 13:21:54.213319 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-xxclj\" (UID: \"7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a\") " pod="openstack/dnsmasq-dns-5f59b8f679-xxclj" Feb 17 13:21:54 crc kubenswrapper[4955]: I0217 13:21:54.214017 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a-config\") pod \"dnsmasq-dns-5f59b8f679-xxclj\" (UID: \"7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a\") " pod="openstack/dnsmasq-dns-5f59b8f679-xxclj" Feb 17 13:21:54 crc kubenswrapper[4955]: I0217 13:21:54.214285 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-xxclj\" (UID: \"7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a\") " pod="openstack/dnsmasq-dns-5f59b8f679-xxclj" Feb 17 13:21:54 crc kubenswrapper[4955]: I0217 13:21:54.214313 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-xxclj\" (UID: \"7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a\") " pod="openstack/dnsmasq-dns-5f59b8f679-xxclj" Feb 17 13:21:54 crc kubenswrapper[4955]: I0217 13:21:54.214526 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-xxclj\" (UID: \"7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a\") " pod="openstack/dnsmasq-dns-5f59b8f679-xxclj" Feb 17 13:21:54 crc kubenswrapper[4955]: I0217 13:21:54.214869 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-xxclj\" (UID: \"7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a\") " pod="openstack/dnsmasq-dns-5f59b8f679-xxclj" Feb 17 13:21:54 crc kubenswrapper[4955]: I0217 13:21:54.233968 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlkwb\" (UniqueName: \"kubernetes.io/projected/7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a-kube-api-access-vlkwb\") pod \"dnsmasq-dns-5f59b8f679-xxclj\" (UID: \"7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a\") " pod="openstack/dnsmasq-dns-5f59b8f679-xxclj" Feb 17 13:21:54 crc kubenswrapper[4955]: I0217 13:21:54.330164 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-xxclj" Feb 17 13:21:54 crc kubenswrapper[4955]: I0217 13:21:54.479932 4955 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","poddd0771e6-147c-4069-88e1-c9ae122dc55d"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort poddd0771e6-147c-4069-88e1-c9ae122dc55d] : Timed out while waiting for systemd to remove kubepods-besteffort-poddd0771e6_147c_4069_88e1_c9ae122dc55d.slice" Feb 17 13:21:54 crc kubenswrapper[4955]: E0217 13:21:54.480294 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort poddd0771e6-147c-4069-88e1-c9ae122dc55d] : unable to destroy cgroup paths for cgroup [kubepods besteffort poddd0771e6-147c-4069-88e1-c9ae122dc55d] : Timed out while waiting for systemd to remove kubepods-besteffort-poddd0771e6_147c_4069_88e1_c9ae122dc55d.slice" pod="openstack/dnsmasq-dns-8554648995-jpznm" podUID="dd0771e6-147c-4069-88e1-c9ae122dc55d" Feb 17 13:21:54 crc kubenswrapper[4955]: I0217 13:21:54.684715 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-jpznm" Feb 17 13:21:54 crc kubenswrapper[4955]: I0217 13:21:54.685847 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74dc88fc-bt2xw" event={"ID":"7dfb2627-eb1d-44f0-b4c2-eea0b404b170","Type":"ContainerStarted","Data":"6ffb95f28fc27e09fce8df244355a1223f65f9ffbf7112e8554af603fc41431f"} Feb 17 13:21:54 crc kubenswrapper[4955]: I0217 13:21:54.685887 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74dc88fc-bt2xw" Feb 17 13:21:54 crc kubenswrapper[4955]: I0217 13:21:54.708205 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74dc88fc-bt2xw" podStartSLOduration=2.708186339 podStartE2EDuration="2.708186339s" podCreationTimestamp="2026-02-17 13:21:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:21:54.705703428 +0000 UTC m=+1053.228432971" watchObservedRunningTime="2026-02-17 13:21:54.708186339 +0000 UTC m=+1053.230915882" Feb 17 13:21:54 crc kubenswrapper[4955]: I0217 13:21:54.728050 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-jpznm"] Feb 17 13:21:54 crc kubenswrapper[4955]: I0217 13:21:54.736671 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8554648995-jpznm"] Feb 17 13:21:54 crc kubenswrapper[4955]: I0217 13:21:54.762613 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-xxclj"] Feb 17 13:21:54 crc kubenswrapper[4955]: W0217 13:21:54.765904 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7bcb9f4f_14cb_4e80_b7e6_b2e0e84dff3a.slice/crio-4e0642b790a638c92b3b5b870c8743a91d4bac05432f296cf349cd9141d42f23 WatchSource:0}: Error finding container 4e0642b790a638c92b3b5b870c8743a91d4bac05432f296cf349cd9141d42f23: Status 404 returned error can't find the container with id 4e0642b790a638c92b3b5b870c8743a91d4bac05432f296cf349cd9141d42f23 Feb 17 13:21:55 crc kubenswrapper[4955]: I0217 13:21:55.695656 4955 generic.go:334] "Generic (PLEG): container finished" podID="7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a" containerID="3adb0f9fd4de50ae4f78609af2df357877052a387dab551eb76a7960ec86083e" exitCode=0 Feb 17 13:21:55 crc kubenswrapper[4955]: I0217 13:21:55.695715 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-xxclj" event={"ID":"7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a","Type":"ContainerDied","Data":"3adb0f9fd4de50ae4f78609af2df357877052a387dab551eb76a7960ec86083e"} Feb 17 13:21:55 crc kubenswrapper[4955]: I0217 13:21:55.696079 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-xxclj" event={"ID":"7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a","Type":"ContainerStarted","Data":"4e0642b790a638c92b3b5b870c8743a91d4bac05432f296cf349cd9141d42f23"} Feb 17 13:21:55 crc kubenswrapper[4955]: I0217 13:21:55.696289 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74dc88fc-bt2xw" podUID="7dfb2627-eb1d-44f0-b4c2-eea0b404b170" containerName="dnsmasq-dns" containerID="cri-o://6ffb95f28fc27e09fce8df244355a1223f65f9ffbf7112e8554af603fc41431f" gracePeriod=10 Feb 17 13:21:56 crc kubenswrapper[4955]: I0217 13:21:56.174046 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74dc88fc-bt2xw" Feb 17 13:21:56 crc kubenswrapper[4955]: I0217 13:21:56.233737 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd0771e6-147c-4069-88e1-c9ae122dc55d" path="/var/lib/kubelet/pods/dd0771e6-147c-4069-88e1-c9ae122dc55d/volumes" Feb 17 13:21:56 crc kubenswrapper[4955]: I0217 13:21:56.353299 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7dfb2627-eb1d-44f0-b4c2-eea0b404b170-config\") pod \"7dfb2627-eb1d-44f0-b4c2-eea0b404b170\" (UID: \"7dfb2627-eb1d-44f0-b4c2-eea0b404b170\") " Feb 17 13:21:56 crc kubenswrapper[4955]: I0217 13:21:56.354360 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7dfb2627-eb1d-44f0-b4c2-eea0b404b170-ovsdbserver-sb\") pod \"7dfb2627-eb1d-44f0-b4c2-eea0b404b170\" (UID: \"7dfb2627-eb1d-44f0-b4c2-eea0b404b170\") " Feb 17 13:21:56 crc kubenswrapper[4955]: I0217 13:21:56.354415 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7dfb2627-eb1d-44f0-b4c2-eea0b404b170-dns-svc\") pod \"7dfb2627-eb1d-44f0-b4c2-eea0b404b170\" (UID: \"7dfb2627-eb1d-44f0-b4c2-eea0b404b170\") " Feb 17 13:21:56 crc kubenswrapper[4955]: I0217 13:21:56.354449 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bh6px\" (UniqueName: \"kubernetes.io/projected/7dfb2627-eb1d-44f0-b4c2-eea0b404b170-kube-api-access-bh6px\") pod \"7dfb2627-eb1d-44f0-b4c2-eea0b404b170\" (UID: \"7dfb2627-eb1d-44f0-b4c2-eea0b404b170\") " Feb 17 13:21:56 crc kubenswrapper[4955]: I0217 13:21:56.354537 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7dfb2627-eb1d-44f0-b4c2-eea0b404b170-ovsdbserver-nb\") pod \"7dfb2627-eb1d-44f0-b4c2-eea0b404b170\" (UID: \"7dfb2627-eb1d-44f0-b4c2-eea0b404b170\") " Feb 17 13:21:56 crc kubenswrapper[4955]: I0217 13:21:56.372701 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7dfb2627-eb1d-44f0-b4c2-eea0b404b170-kube-api-access-bh6px" (OuterVolumeSpecName: "kube-api-access-bh6px") pod "7dfb2627-eb1d-44f0-b4c2-eea0b404b170" (UID: "7dfb2627-eb1d-44f0-b4c2-eea0b404b170"). InnerVolumeSpecName "kube-api-access-bh6px". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:21:56 crc kubenswrapper[4955]: I0217 13:21:56.401370 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7dfb2627-eb1d-44f0-b4c2-eea0b404b170-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7dfb2627-eb1d-44f0-b4c2-eea0b404b170" (UID: "7dfb2627-eb1d-44f0-b4c2-eea0b404b170"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:21:56 crc kubenswrapper[4955]: I0217 13:21:56.405774 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7dfb2627-eb1d-44f0-b4c2-eea0b404b170-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7dfb2627-eb1d-44f0-b4c2-eea0b404b170" (UID: "7dfb2627-eb1d-44f0-b4c2-eea0b404b170"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:21:56 crc kubenswrapper[4955]: I0217 13:21:56.406037 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7dfb2627-eb1d-44f0-b4c2-eea0b404b170-config" (OuterVolumeSpecName: "config") pod "7dfb2627-eb1d-44f0-b4c2-eea0b404b170" (UID: "7dfb2627-eb1d-44f0-b4c2-eea0b404b170"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:21:56 crc kubenswrapper[4955]: I0217 13:21:56.423099 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7dfb2627-eb1d-44f0-b4c2-eea0b404b170-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7dfb2627-eb1d-44f0-b4c2-eea0b404b170" (UID: "7dfb2627-eb1d-44f0-b4c2-eea0b404b170"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:21:56 crc kubenswrapper[4955]: I0217 13:21:56.456661 4955 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7dfb2627-eb1d-44f0-b4c2-eea0b404b170-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:56 crc kubenswrapper[4955]: I0217 13:21:56.456697 4955 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7dfb2627-eb1d-44f0-b4c2-eea0b404b170-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:56 crc kubenswrapper[4955]: I0217 13:21:56.456708 4955 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7dfb2627-eb1d-44f0-b4c2-eea0b404b170-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:56 crc kubenswrapper[4955]: I0217 13:21:56.456716 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bh6px\" (UniqueName: \"kubernetes.io/projected/7dfb2627-eb1d-44f0-b4c2-eea0b404b170-kube-api-access-bh6px\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:56 crc kubenswrapper[4955]: I0217 13:21:56.456726 4955 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7dfb2627-eb1d-44f0-b4c2-eea0b404b170-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 17 13:21:56 crc kubenswrapper[4955]: I0217 13:21:56.704105 4955 generic.go:334] "Generic (PLEG): container finished" podID="7dfb2627-eb1d-44f0-b4c2-eea0b404b170" containerID="6ffb95f28fc27e09fce8df244355a1223f65f9ffbf7112e8554af603fc41431f" exitCode=0 Feb 17 13:21:56 crc kubenswrapper[4955]: I0217 13:21:56.704212 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74dc88fc-bt2xw" event={"ID":"7dfb2627-eb1d-44f0-b4c2-eea0b404b170","Type":"ContainerDied","Data":"6ffb95f28fc27e09fce8df244355a1223f65f9ffbf7112e8554af603fc41431f"} Feb 17 13:21:56 crc kubenswrapper[4955]: I0217 13:21:56.704250 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74dc88fc-bt2xw" event={"ID":"7dfb2627-eb1d-44f0-b4c2-eea0b404b170","Type":"ContainerDied","Data":"11952759f507b9214f2ff285de21a12e6040b25e33a5b08bd74f0079a17abb6e"} Feb 17 13:21:56 crc kubenswrapper[4955]: I0217 13:21:56.704200 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74dc88fc-bt2xw" Feb 17 13:21:56 crc kubenswrapper[4955]: I0217 13:21:56.704277 4955 scope.go:117] "RemoveContainer" containerID="6ffb95f28fc27e09fce8df244355a1223f65f9ffbf7112e8554af603fc41431f" Feb 17 13:21:56 crc kubenswrapper[4955]: I0217 13:21:56.706536 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-xxclj" event={"ID":"7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a","Type":"ContainerStarted","Data":"4fcb9bb9cd8261f98d8dc33fd2041813f105d13915a97619409dd56f1dd69a52"} Feb 17 13:21:56 crc kubenswrapper[4955]: I0217 13:21:56.707044 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5f59b8f679-xxclj" Feb 17 13:21:56 crc kubenswrapper[4955]: I0217 13:21:56.723955 4955 scope.go:117] "RemoveContainer" containerID="cb5afab742c34169ec27e429061b294a63fd6fcbb3ffe8316270f25b74ad99c3" Feb 17 13:21:56 crc kubenswrapper[4955]: I0217 13:21:56.741913 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5f59b8f679-xxclj" podStartSLOduration=3.74188968 podStartE2EDuration="3.74188968s" podCreationTimestamp="2026-02-17 13:21:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:21:56.728955761 +0000 UTC m=+1055.251685374" watchObservedRunningTime="2026-02-17 13:21:56.74188968 +0000 UTC m=+1055.264619263" Feb 17 13:21:56 crc kubenswrapper[4955]: E0217 13:21:56.744129 4955 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7dfb2627_eb1d_44f0_b4c2_eea0b404b170.slice\": RecentStats: unable to find data in memory cache]" Feb 17 13:21:56 crc kubenswrapper[4955]: I0217 13:21:56.752216 4955 scope.go:117] "RemoveContainer" containerID="6ffb95f28fc27e09fce8df244355a1223f65f9ffbf7112e8554af603fc41431f" Feb 17 13:21:56 crc kubenswrapper[4955]: E0217 13:21:56.753004 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ffb95f28fc27e09fce8df244355a1223f65f9ffbf7112e8554af603fc41431f\": container with ID starting with 6ffb95f28fc27e09fce8df244355a1223f65f9ffbf7112e8554af603fc41431f not found: ID does not exist" containerID="6ffb95f28fc27e09fce8df244355a1223f65f9ffbf7112e8554af603fc41431f" Feb 17 13:21:56 crc kubenswrapper[4955]: I0217 13:21:56.753049 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ffb95f28fc27e09fce8df244355a1223f65f9ffbf7112e8554af603fc41431f"} err="failed to get container status \"6ffb95f28fc27e09fce8df244355a1223f65f9ffbf7112e8554af603fc41431f\": rpc error: code = NotFound desc = could not find container \"6ffb95f28fc27e09fce8df244355a1223f65f9ffbf7112e8554af603fc41431f\": container with ID starting with 6ffb95f28fc27e09fce8df244355a1223f65f9ffbf7112e8554af603fc41431f not found: ID does not exist" Feb 17 13:21:56 crc kubenswrapper[4955]: I0217 13:21:56.753076 4955 scope.go:117] "RemoveContainer" containerID="cb5afab742c34169ec27e429061b294a63fd6fcbb3ffe8316270f25b74ad99c3" Feb 17 13:21:56 crc kubenswrapper[4955]: E0217 13:21:56.754051 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb5afab742c34169ec27e429061b294a63fd6fcbb3ffe8316270f25b74ad99c3\": container with ID starting with cb5afab742c34169ec27e429061b294a63fd6fcbb3ffe8316270f25b74ad99c3 not found: ID does not exist" containerID="cb5afab742c34169ec27e429061b294a63fd6fcbb3ffe8316270f25b74ad99c3" Feb 17 13:21:56 crc kubenswrapper[4955]: I0217 13:21:56.754090 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb5afab742c34169ec27e429061b294a63fd6fcbb3ffe8316270f25b74ad99c3"} err="failed to get container status \"cb5afab742c34169ec27e429061b294a63fd6fcbb3ffe8316270f25b74ad99c3\": rpc error: code = NotFound desc = could not find container \"cb5afab742c34169ec27e429061b294a63fd6fcbb3ffe8316270f25b74ad99c3\": container with ID starting with cb5afab742c34169ec27e429061b294a63fd6fcbb3ffe8316270f25b74ad99c3 not found: ID does not exist" Feb 17 13:21:56 crc kubenswrapper[4955]: I0217 13:21:56.758440 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74dc88fc-bt2xw"] Feb 17 13:21:56 crc kubenswrapper[4955]: I0217 13:21:56.766319 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74dc88fc-bt2xw"] Feb 17 13:21:57 crc kubenswrapper[4955]: I0217 13:21:57.192992 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Feb 17 13:21:57 crc kubenswrapper[4955]: I0217 13:21:57.502178 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:21:57 crc kubenswrapper[4955]: I0217 13:21:57.636490 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-tqmd4"] Feb 17 13:21:57 crc kubenswrapper[4955]: E0217 13:21:57.636898 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dfb2627-eb1d-44f0-b4c2-eea0b404b170" containerName="dnsmasq-dns" Feb 17 13:21:57 crc kubenswrapper[4955]: I0217 13:21:57.636919 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dfb2627-eb1d-44f0-b4c2-eea0b404b170" containerName="dnsmasq-dns" Feb 17 13:21:57 crc kubenswrapper[4955]: E0217 13:21:57.636938 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dfb2627-eb1d-44f0-b4c2-eea0b404b170" containerName="init" Feb 17 13:21:57 crc kubenswrapper[4955]: I0217 13:21:57.636946 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dfb2627-eb1d-44f0-b4c2-eea0b404b170" containerName="init" Feb 17 13:21:57 crc kubenswrapper[4955]: I0217 13:21:57.637155 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dfb2627-eb1d-44f0-b4c2-eea0b404b170" containerName="dnsmasq-dns" Feb 17 13:21:57 crc kubenswrapper[4955]: I0217 13:21:57.637767 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-tqmd4" Feb 17 13:21:57 crc kubenswrapper[4955]: I0217 13:21:57.649837 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-tqmd4"] Feb 17 13:21:57 crc kubenswrapper[4955]: I0217 13:21:57.783725 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a61deb1d-ae88-49f8-8954-2059c469e84f-operator-scripts\") pod \"cinder-db-create-tqmd4\" (UID: \"a61deb1d-ae88-49f8-8954-2059c469e84f\") " pod="openstack/cinder-db-create-tqmd4" Feb 17 13:21:57 crc kubenswrapper[4955]: I0217 13:21:57.783817 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pffl\" (UniqueName: \"kubernetes.io/projected/a61deb1d-ae88-49f8-8954-2059c469e84f-kube-api-access-7pffl\") pod \"cinder-db-create-tqmd4\" (UID: \"a61deb1d-ae88-49f8-8954-2059c469e84f\") " pod="openstack/cinder-db-create-tqmd4" Feb 17 13:21:57 crc kubenswrapper[4955]: I0217 13:21:57.793366 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-4mktx"] Feb 17 13:21:57 crc kubenswrapper[4955]: I0217 13:21:57.794302 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-4mktx" Feb 17 13:21:57 crc kubenswrapper[4955]: I0217 13:21:57.826686 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-a0dc-account-create-update-rbbn6"] Feb 17 13:21:57 crc kubenswrapper[4955]: I0217 13:21:57.835750 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-a0dc-account-create-update-rbbn6" Feb 17 13:21:57 crc kubenswrapper[4955]: I0217 13:21:57.841123 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Feb 17 13:21:57 crc kubenswrapper[4955]: I0217 13:21:57.847530 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-4mktx"] Feb 17 13:21:57 crc kubenswrapper[4955]: I0217 13:21:57.894257 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4fe8b1c0-db08-4b58-b1cb-218631c424ab-operator-scripts\") pod \"barbican-db-create-4mktx\" (UID: \"4fe8b1c0-db08-4b58-b1cb-218631c424ab\") " pod="openstack/barbican-db-create-4mktx" Feb 17 13:21:57 crc kubenswrapper[4955]: I0217 13:21:57.894323 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a61deb1d-ae88-49f8-8954-2059c469e84f-operator-scripts\") pod \"cinder-db-create-tqmd4\" (UID: \"a61deb1d-ae88-49f8-8954-2059c469e84f\") " pod="openstack/cinder-db-create-tqmd4" Feb 17 13:21:57 crc kubenswrapper[4955]: I0217 13:21:57.894387 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pffl\" (UniqueName: \"kubernetes.io/projected/a61deb1d-ae88-49f8-8954-2059c469e84f-kube-api-access-7pffl\") pod \"cinder-db-create-tqmd4\" (UID: \"a61deb1d-ae88-49f8-8954-2059c469e84f\") " pod="openstack/cinder-db-create-tqmd4" Feb 17 13:21:57 crc kubenswrapper[4955]: I0217 13:21:57.894482 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkk47\" (UniqueName: \"kubernetes.io/projected/4fe8b1c0-db08-4b58-b1cb-218631c424ab-kube-api-access-pkk47\") pod \"barbican-db-create-4mktx\" (UID: \"4fe8b1c0-db08-4b58-b1cb-218631c424ab\") " pod="openstack/barbican-db-create-4mktx" Feb 17 13:21:57 crc kubenswrapper[4955]: I0217 13:21:57.895480 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a61deb1d-ae88-49f8-8954-2059c469e84f-operator-scripts\") pod \"cinder-db-create-tqmd4\" (UID: \"a61deb1d-ae88-49f8-8954-2059c469e84f\") " pod="openstack/cinder-db-create-tqmd4" Feb 17 13:21:57 crc kubenswrapper[4955]: I0217 13:21:57.947724 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-a0dc-account-create-update-rbbn6"] Feb 17 13:21:57 crc kubenswrapper[4955]: I0217 13:21:57.971032 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pffl\" (UniqueName: \"kubernetes.io/projected/a61deb1d-ae88-49f8-8954-2059c469e84f-kube-api-access-7pffl\") pod \"cinder-db-create-tqmd4\" (UID: \"a61deb1d-ae88-49f8-8954-2059c469e84f\") " pod="openstack/cinder-db-create-tqmd4" Feb 17 13:21:57 crc kubenswrapper[4955]: I0217 13:21:57.979773 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-6ss66"] Feb 17 13:21:57 crc kubenswrapper[4955]: I0217 13:21:57.980823 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-6ss66" Feb 17 13:21:57 crc kubenswrapper[4955]: I0217 13:21:57.988875 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-6ss66"] Feb 17 13:21:57 crc kubenswrapper[4955]: I0217 13:21:57.995466 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vr296\" (UniqueName: \"kubernetes.io/projected/4da596f3-7aff-4cfd-984e-0fd58d82a25c-kube-api-access-vr296\") pod \"cinder-a0dc-account-create-update-rbbn6\" (UID: \"4da596f3-7aff-4cfd-984e-0fd58d82a25c\") " pod="openstack/cinder-a0dc-account-create-update-rbbn6" Feb 17 13:21:57 crc kubenswrapper[4955]: I0217 13:21:57.995534 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkk47\" (UniqueName: \"kubernetes.io/projected/4fe8b1c0-db08-4b58-b1cb-218631c424ab-kube-api-access-pkk47\") pod \"barbican-db-create-4mktx\" (UID: \"4fe8b1c0-db08-4b58-b1cb-218631c424ab\") " pod="openstack/barbican-db-create-4mktx" Feb 17 13:21:57 crc kubenswrapper[4955]: I0217 13:21:57.995565 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4da596f3-7aff-4cfd-984e-0fd58d82a25c-operator-scripts\") pod \"cinder-a0dc-account-create-update-rbbn6\" (UID: \"4da596f3-7aff-4cfd-984e-0fd58d82a25c\") " pod="openstack/cinder-a0dc-account-create-update-rbbn6" Feb 17 13:21:57 crc kubenswrapper[4955]: I0217 13:21:57.995609 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4fe8b1c0-db08-4b58-b1cb-218631c424ab-operator-scripts\") pod \"barbican-db-create-4mktx\" (UID: \"4fe8b1c0-db08-4b58-b1cb-218631c424ab\") " pod="openstack/barbican-db-create-4mktx" Feb 17 13:21:57 crc kubenswrapper[4955]: I0217 13:21:57.996337 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4fe8b1c0-db08-4b58-b1cb-218631c424ab-operator-scripts\") pod \"barbican-db-create-4mktx\" (UID: \"4fe8b1c0-db08-4b58-b1cb-218631c424ab\") " pod="openstack/barbican-db-create-4mktx" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.024827 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkk47\" (UniqueName: \"kubernetes.io/projected/4fe8b1c0-db08-4b58-b1cb-218631c424ab-kube-api-access-pkk47\") pod \"barbican-db-create-4mktx\" (UID: \"4fe8b1c0-db08-4b58-b1cb-218631c424ab\") " pod="openstack/barbican-db-create-4mktx" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.034165 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-hcggw"] Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.035624 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-hcggw" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.057900 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.058132 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.058230 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-qh96v" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.058364 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.088914 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-hcggw"] Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.097209 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vr296\" (UniqueName: \"kubernetes.io/projected/4da596f3-7aff-4cfd-984e-0fd58d82a25c-kube-api-access-vr296\") pod \"cinder-a0dc-account-create-update-rbbn6\" (UID: \"4da596f3-7aff-4cfd-984e-0fd58d82a25c\") " pod="openstack/cinder-a0dc-account-create-update-rbbn6" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.097255 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clsnh\" (UniqueName: \"kubernetes.io/projected/dd1f70a3-aaee-464a-aa89-621279758849-kube-api-access-clsnh\") pod \"neutron-db-create-6ss66\" (UID: \"dd1f70a3-aaee-464a-aa89-621279758849\") " pod="openstack/neutron-db-create-6ss66" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.097295 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4da596f3-7aff-4cfd-984e-0fd58d82a25c-operator-scripts\") pod \"cinder-a0dc-account-create-update-rbbn6\" (UID: \"4da596f3-7aff-4cfd-984e-0fd58d82a25c\") " pod="openstack/cinder-a0dc-account-create-update-rbbn6" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.097346 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd1f70a3-aaee-464a-aa89-621279758849-operator-scripts\") pod \"neutron-db-create-6ss66\" (UID: \"dd1f70a3-aaee-464a-aa89-621279758849\") " pod="openstack/neutron-db-create-6ss66" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.098253 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4da596f3-7aff-4cfd-984e-0fd58d82a25c-operator-scripts\") pod \"cinder-a0dc-account-create-update-rbbn6\" (UID: \"4da596f3-7aff-4cfd-984e-0fd58d82a25c\") " pod="openstack/cinder-a0dc-account-create-update-rbbn6" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.106696 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-4mktx" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.127031 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vr296\" (UniqueName: \"kubernetes.io/projected/4da596f3-7aff-4cfd-984e-0fd58d82a25c-kube-api-access-vr296\") pod \"cinder-a0dc-account-create-update-rbbn6\" (UID: \"4da596f3-7aff-4cfd-984e-0fd58d82a25c\") " pod="openstack/cinder-a0dc-account-create-update-rbbn6" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.165495 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-ab8c-account-create-update-49rjx"] Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.166412 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ab8c-account-create-update-49rjx" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.170440 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.186916 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-a0dc-account-create-update-rbbn6" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.200916 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd1f70a3-aaee-464a-aa89-621279758849-operator-scripts\") pod \"neutron-db-create-6ss66\" (UID: \"dd1f70a3-aaee-464a-aa89-621279758849\") " pod="openstack/neutron-db-create-6ss66" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.200957 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a-config-data\") pod \"keystone-db-sync-hcggw\" (UID: \"25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a\") " pod="openstack/keystone-db-sync-hcggw" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.200995 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbngc\" (UniqueName: \"kubernetes.io/projected/25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a-kube-api-access-zbngc\") pod \"keystone-db-sync-hcggw\" (UID: \"25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a\") " pod="openstack/keystone-db-sync-hcggw" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.201023 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a-combined-ca-bundle\") pod \"keystone-db-sync-hcggw\" (UID: \"25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a\") " pod="openstack/keystone-db-sync-hcggw" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.201065 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clsnh\" (UniqueName: \"kubernetes.io/projected/dd1f70a3-aaee-464a-aa89-621279758849-kube-api-access-clsnh\") pod \"neutron-db-create-6ss66\" (UID: \"dd1f70a3-aaee-464a-aa89-621279758849\") " pod="openstack/neutron-db-create-6ss66" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.201957 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd1f70a3-aaee-464a-aa89-621279758849-operator-scripts\") pod \"neutron-db-create-6ss66\" (UID: \"dd1f70a3-aaee-464a-aa89-621279758849\") " pod="openstack/neutron-db-create-6ss66" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.217376 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-ab8c-account-create-update-49rjx"] Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.237326 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clsnh\" (UniqueName: \"kubernetes.io/projected/dd1f70a3-aaee-464a-aa89-621279758849-kube-api-access-clsnh\") pod \"neutron-db-create-6ss66\" (UID: \"dd1f70a3-aaee-464a-aa89-621279758849\") " pod="openstack/neutron-db-create-6ss66" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.251214 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7dfb2627-eb1d-44f0-b4c2-eea0b404b170" path="/var/lib/kubelet/pods/7dfb2627-eb1d-44f0-b4c2-eea0b404b170/volumes" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.253532 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-tqmd4" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.297677 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-3d6f-account-create-update-zsh9p"] Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.302182 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a-combined-ca-bundle\") pod \"keystone-db-sync-hcggw\" (UID: \"25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a\") " pod="openstack/keystone-db-sync-hcggw" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.302267 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2fa427d9-5f2e-4cb9-934b-edd1086e0f42-operator-scripts\") pod \"neutron-ab8c-account-create-update-49rjx\" (UID: \"2fa427d9-5f2e-4cb9-934b-edd1086e0f42\") " pod="openstack/neutron-ab8c-account-create-update-49rjx" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.302381 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a-config-data\") pod \"keystone-db-sync-hcggw\" (UID: \"25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a\") " pod="openstack/keystone-db-sync-hcggw" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.302445 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbngc\" (UniqueName: \"kubernetes.io/projected/25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a-kube-api-access-zbngc\") pod \"keystone-db-sync-hcggw\" (UID: \"25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a\") " pod="openstack/keystone-db-sync-hcggw" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.302477 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwrtz\" (UniqueName: \"kubernetes.io/projected/2fa427d9-5f2e-4cb9-934b-edd1086e0f42-kube-api-access-fwrtz\") pod \"neutron-ab8c-account-create-update-49rjx\" (UID: \"2fa427d9-5f2e-4cb9-934b-edd1086e0f42\") " pod="openstack/neutron-ab8c-account-create-update-49rjx" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.306635 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a-combined-ca-bundle\") pod \"keystone-db-sync-hcggw\" (UID: \"25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a\") " pod="openstack/keystone-db-sync-hcggw" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.307300 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a-config-data\") pod \"keystone-db-sync-hcggw\" (UID: \"25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a\") " pod="openstack/keystone-db-sync-hcggw" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.310249 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-3d6f-account-create-update-zsh9p" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.314667 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.319846 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-3d6f-account-create-update-zsh9p"] Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.337624 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbngc\" (UniqueName: \"kubernetes.io/projected/25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a-kube-api-access-zbngc\") pod \"keystone-db-sync-hcggw\" (UID: \"25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a\") " pod="openstack/keystone-db-sync-hcggw" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.372058 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-6ss66" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.383151 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-hcggw" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.404970 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2fa427d9-5f2e-4cb9-934b-edd1086e0f42-operator-scripts\") pod \"neutron-ab8c-account-create-update-49rjx\" (UID: \"2fa427d9-5f2e-4cb9-934b-edd1086e0f42\") " pod="openstack/neutron-ab8c-account-create-update-49rjx" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.405101 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7cb66d5d-60a5-486d-b9ff-35e2ebeb4d8b-operator-scripts\") pod \"barbican-3d6f-account-create-update-zsh9p\" (UID: \"7cb66d5d-60a5-486d-b9ff-35e2ebeb4d8b\") " pod="openstack/barbican-3d6f-account-create-update-zsh9p" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.405151 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jnsp\" (UniqueName: \"kubernetes.io/projected/7cb66d5d-60a5-486d-b9ff-35e2ebeb4d8b-kube-api-access-8jnsp\") pod \"barbican-3d6f-account-create-update-zsh9p\" (UID: \"7cb66d5d-60a5-486d-b9ff-35e2ebeb4d8b\") " pod="openstack/barbican-3d6f-account-create-update-zsh9p" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.405198 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwrtz\" (UniqueName: \"kubernetes.io/projected/2fa427d9-5f2e-4cb9-934b-edd1086e0f42-kube-api-access-fwrtz\") pod \"neutron-ab8c-account-create-update-49rjx\" (UID: \"2fa427d9-5f2e-4cb9-934b-edd1086e0f42\") " pod="openstack/neutron-ab8c-account-create-update-49rjx" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.405873 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2fa427d9-5f2e-4cb9-934b-edd1086e0f42-operator-scripts\") pod \"neutron-ab8c-account-create-update-49rjx\" (UID: \"2fa427d9-5f2e-4cb9-934b-edd1086e0f42\") " pod="openstack/neutron-ab8c-account-create-update-49rjx" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.423896 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwrtz\" (UniqueName: \"kubernetes.io/projected/2fa427d9-5f2e-4cb9-934b-edd1086e0f42-kube-api-access-fwrtz\") pod \"neutron-ab8c-account-create-update-49rjx\" (UID: \"2fa427d9-5f2e-4cb9-934b-edd1086e0f42\") " pod="openstack/neutron-ab8c-account-create-update-49rjx" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.507372 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jnsp\" (UniqueName: \"kubernetes.io/projected/7cb66d5d-60a5-486d-b9ff-35e2ebeb4d8b-kube-api-access-8jnsp\") pod \"barbican-3d6f-account-create-update-zsh9p\" (UID: \"7cb66d5d-60a5-486d-b9ff-35e2ebeb4d8b\") " pod="openstack/barbican-3d6f-account-create-update-zsh9p" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.507517 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7cb66d5d-60a5-486d-b9ff-35e2ebeb4d8b-operator-scripts\") pod \"barbican-3d6f-account-create-update-zsh9p\" (UID: \"7cb66d5d-60a5-486d-b9ff-35e2ebeb4d8b\") " pod="openstack/barbican-3d6f-account-create-update-zsh9p" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.508261 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7cb66d5d-60a5-486d-b9ff-35e2ebeb4d8b-operator-scripts\") pod \"barbican-3d6f-account-create-update-zsh9p\" (UID: \"7cb66d5d-60a5-486d-b9ff-35e2ebeb4d8b\") " pod="openstack/barbican-3d6f-account-create-update-zsh9p" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.535307 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jnsp\" (UniqueName: \"kubernetes.io/projected/7cb66d5d-60a5-486d-b9ff-35e2ebeb4d8b-kube-api-access-8jnsp\") pod \"barbican-3d6f-account-create-update-zsh9p\" (UID: \"7cb66d5d-60a5-486d-b9ff-35e2ebeb4d8b\") " pod="openstack/barbican-3d6f-account-create-update-zsh9p" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.567957 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ab8c-account-create-update-49rjx" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.648336 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-3d6f-account-create-update-zsh9p" Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.664724 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-4mktx"] Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.725831 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-4mktx" event={"ID":"4fe8b1c0-db08-4b58-b1cb-218631c424ab","Type":"ContainerStarted","Data":"20f7823a30494bc345da282f4205ce36402f554efe0012253a2802005b0e2988"} Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.851036 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-a0dc-account-create-update-rbbn6"] Feb 17 13:21:58 crc kubenswrapper[4955]: I0217 13:21:58.923894 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-tqmd4"] Feb 17 13:21:59 crc kubenswrapper[4955]: I0217 13:21:59.063832 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-6ss66"] Feb 17 13:21:59 crc kubenswrapper[4955]: I0217 13:21:59.073110 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-hcggw"] Feb 17 13:21:59 crc kubenswrapper[4955]: I0217 13:21:59.157012 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-ab8c-account-create-update-49rjx"] Feb 17 13:21:59 crc kubenswrapper[4955]: W0217 13:21:59.173709 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2fa427d9_5f2e_4cb9_934b_edd1086e0f42.slice/crio-73e39cfc63e27d4acd1274d16ddb162ad7985c648cc7997e05c1e4047e0986f0 WatchSource:0}: Error finding container 73e39cfc63e27d4acd1274d16ddb162ad7985c648cc7997e05c1e4047e0986f0: Status 404 returned error can't find the container with id 73e39cfc63e27d4acd1274d16ddb162ad7985c648cc7997e05c1e4047e0986f0 Feb 17 13:21:59 crc kubenswrapper[4955]: I0217 13:21:59.279296 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-3d6f-account-create-update-zsh9p"] Feb 17 13:21:59 crc kubenswrapper[4955]: W0217 13:21:59.310636 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7cb66d5d_60a5_486d_b9ff_35e2ebeb4d8b.slice/crio-5c54198ed25e7a4d53de61cd3061ffbff1fcd3410dacaca9e71c03a327b392e0 WatchSource:0}: Error finding container 5c54198ed25e7a4d53de61cd3061ffbff1fcd3410dacaca9e71c03a327b392e0: Status 404 returned error can't find the container with id 5c54198ed25e7a4d53de61cd3061ffbff1fcd3410dacaca9e71c03a327b392e0 Feb 17 13:21:59 crc kubenswrapper[4955]: I0217 13:21:59.738307 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-hcggw" event={"ID":"25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a","Type":"ContainerStarted","Data":"95af02a5119e291c0333fd97ecd5f746454fddc015b9c3f7fce849b8b111d445"} Feb 17 13:21:59 crc kubenswrapper[4955]: I0217 13:21:59.741120 4955 generic.go:334] "Generic (PLEG): container finished" podID="4da596f3-7aff-4cfd-984e-0fd58d82a25c" containerID="c0f874e3765ae3cf7397baf8468592643150f43238602ea03aaaffc680bb61f5" exitCode=0 Feb 17 13:21:59 crc kubenswrapper[4955]: I0217 13:21:59.741220 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-a0dc-account-create-update-rbbn6" event={"ID":"4da596f3-7aff-4cfd-984e-0fd58d82a25c","Type":"ContainerDied","Data":"c0f874e3765ae3cf7397baf8468592643150f43238602ea03aaaffc680bb61f5"} Feb 17 13:21:59 crc kubenswrapper[4955]: I0217 13:21:59.741310 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-a0dc-account-create-update-rbbn6" event={"ID":"4da596f3-7aff-4cfd-984e-0fd58d82a25c","Type":"ContainerStarted","Data":"e1234988ed9b1db48881db75126d8babe91b2113d0689e2ca9867da4887eabae"} Feb 17 13:21:59 crc kubenswrapper[4955]: I0217 13:21:59.743177 4955 generic.go:334] "Generic (PLEG): container finished" podID="4fe8b1c0-db08-4b58-b1cb-218631c424ab" containerID="95d036ca02fb8de95169f573c8c35bc0d7f7a2b7c41b806464700bc89b2cd855" exitCode=0 Feb 17 13:21:59 crc kubenswrapper[4955]: I0217 13:21:59.743238 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-4mktx" event={"ID":"4fe8b1c0-db08-4b58-b1cb-218631c424ab","Type":"ContainerDied","Data":"95d036ca02fb8de95169f573c8c35bc0d7f7a2b7c41b806464700bc89b2cd855"} Feb 17 13:21:59 crc kubenswrapper[4955]: I0217 13:21:59.745296 4955 generic.go:334] "Generic (PLEG): container finished" podID="a61deb1d-ae88-49f8-8954-2059c469e84f" containerID="bb55ac7b0c2ff9916def101839e6a50422bafb9dc82d108b0de43548d2d99d62" exitCode=0 Feb 17 13:21:59 crc kubenswrapper[4955]: I0217 13:21:59.745353 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-tqmd4" event={"ID":"a61deb1d-ae88-49f8-8954-2059c469e84f","Type":"ContainerDied","Data":"bb55ac7b0c2ff9916def101839e6a50422bafb9dc82d108b0de43548d2d99d62"} Feb 17 13:21:59 crc kubenswrapper[4955]: I0217 13:21:59.745372 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-tqmd4" event={"ID":"a61deb1d-ae88-49f8-8954-2059c469e84f","Type":"ContainerStarted","Data":"a115081be771adfcbb27ab756e993e17aaedcd7dcb44ff2b04089e31c4a4d422"} Feb 17 13:21:59 crc kubenswrapper[4955]: I0217 13:21:59.748048 4955 generic.go:334] "Generic (PLEG): container finished" podID="2fa427d9-5f2e-4cb9-934b-edd1086e0f42" containerID="ea54a9d6751ef4e8c620fe8828d31a2851d55b7afa7e29803b3bd5f40fcd3ddb" exitCode=0 Feb 17 13:21:59 crc kubenswrapper[4955]: I0217 13:21:59.748109 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ab8c-account-create-update-49rjx" event={"ID":"2fa427d9-5f2e-4cb9-934b-edd1086e0f42","Type":"ContainerDied","Data":"ea54a9d6751ef4e8c620fe8828d31a2851d55b7afa7e29803b3bd5f40fcd3ddb"} Feb 17 13:21:59 crc kubenswrapper[4955]: I0217 13:21:59.748130 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ab8c-account-create-update-49rjx" event={"ID":"2fa427d9-5f2e-4cb9-934b-edd1086e0f42","Type":"ContainerStarted","Data":"73e39cfc63e27d4acd1274d16ddb162ad7985c648cc7997e05c1e4047e0986f0"} Feb 17 13:21:59 crc kubenswrapper[4955]: I0217 13:21:59.753554 4955 generic.go:334] "Generic (PLEG): container finished" podID="dd1f70a3-aaee-464a-aa89-621279758849" containerID="f4e8cff5f113bfbc77a2f69000f3d32f76c5c84fd69217f12e593f37eb3f93f1" exitCode=0 Feb 17 13:21:59 crc kubenswrapper[4955]: I0217 13:21:59.753607 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-6ss66" event={"ID":"dd1f70a3-aaee-464a-aa89-621279758849","Type":"ContainerDied","Data":"f4e8cff5f113bfbc77a2f69000f3d32f76c5c84fd69217f12e593f37eb3f93f1"} Feb 17 13:21:59 crc kubenswrapper[4955]: I0217 13:21:59.753631 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-6ss66" event={"ID":"dd1f70a3-aaee-464a-aa89-621279758849","Type":"ContainerStarted","Data":"1ecb54d8110f542c643895c3b665030d9a44943d1d180d6c4f5a144c6e11d7e1"} Feb 17 13:21:59 crc kubenswrapper[4955]: I0217 13:21:59.757094 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-3d6f-account-create-update-zsh9p" event={"ID":"7cb66d5d-60a5-486d-b9ff-35e2ebeb4d8b","Type":"ContainerStarted","Data":"9b4356ed08e500a83ad864d58ae6b2c7ec3a5ce1ae99e89074f50655811b46dd"} Feb 17 13:21:59 crc kubenswrapper[4955]: I0217 13:21:59.757120 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-3d6f-account-create-update-zsh9p" event={"ID":"7cb66d5d-60a5-486d-b9ff-35e2ebeb4d8b","Type":"ContainerStarted","Data":"5c54198ed25e7a4d53de61cd3061ffbff1fcd3410dacaca9e71c03a327b392e0"} Feb 17 13:22:00 crc kubenswrapper[4955]: I0217 13:22:00.766905 4955 generic.go:334] "Generic (PLEG): container finished" podID="7cb66d5d-60a5-486d-b9ff-35e2ebeb4d8b" containerID="9b4356ed08e500a83ad864d58ae6b2c7ec3a5ce1ae99e89074f50655811b46dd" exitCode=0 Feb 17 13:22:00 crc kubenswrapper[4955]: I0217 13:22:00.768061 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-3d6f-account-create-update-zsh9p" event={"ID":"7cb66d5d-60a5-486d-b9ff-35e2ebeb4d8b","Type":"ContainerDied","Data":"9b4356ed08e500a83ad864d58ae6b2c7ec3a5ce1ae99e89074f50655811b46dd"} Feb 17 13:22:01 crc kubenswrapper[4955]: I0217 13:22:01.129414 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-3d6f-account-create-update-zsh9p" Feb 17 13:22:01 crc kubenswrapper[4955]: I0217 13:22:01.260248 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jnsp\" (UniqueName: \"kubernetes.io/projected/7cb66d5d-60a5-486d-b9ff-35e2ebeb4d8b-kube-api-access-8jnsp\") pod \"7cb66d5d-60a5-486d-b9ff-35e2ebeb4d8b\" (UID: \"7cb66d5d-60a5-486d-b9ff-35e2ebeb4d8b\") " Feb 17 13:22:01 crc kubenswrapper[4955]: I0217 13:22:01.260326 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7cb66d5d-60a5-486d-b9ff-35e2ebeb4d8b-operator-scripts\") pod \"7cb66d5d-60a5-486d-b9ff-35e2ebeb4d8b\" (UID: \"7cb66d5d-60a5-486d-b9ff-35e2ebeb4d8b\") " Feb 17 13:22:01 crc kubenswrapper[4955]: I0217 13:22:01.261366 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cb66d5d-60a5-486d-b9ff-35e2ebeb4d8b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7cb66d5d-60a5-486d-b9ff-35e2ebeb4d8b" (UID: "7cb66d5d-60a5-486d-b9ff-35e2ebeb4d8b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:22:01 crc kubenswrapper[4955]: I0217 13:22:01.265841 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cb66d5d-60a5-486d-b9ff-35e2ebeb4d8b-kube-api-access-8jnsp" (OuterVolumeSpecName: "kube-api-access-8jnsp") pod "7cb66d5d-60a5-486d-b9ff-35e2ebeb4d8b" (UID: "7cb66d5d-60a5-486d-b9ff-35e2ebeb4d8b"). InnerVolumeSpecName "kube-api-access-8jnsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:22:01 crc kubenswrapper[4955]: I0217 13:22:01.362685 4955 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7cb66d5d-60a5-486d-b9ff-35e2ebeb4d8b-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:01 crc kubenswrapper[4955]: I0217 13:22:01.362728 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jnsp\" (UniqueName: \"kubernetes.io/projected/7cb66d5d-60a5-486d-b9ff-35e2ebeb4d8b-kube-api-access-8jnsp\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:01 crc kubenswrapper[4955]: I0217 13:22:01.776449 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-3d6f-account-create-update-zsh9p" event={"ID":"7cb66d5d-60a5-486d-b9ff-35e2ebeb4d8b","Type":"ContainerDied","Data":"5c54198ed25e7a4d53de61cd3061ffbff1fcd3410dacaca9e71c03a327b392e0"} Feb 17 13:22:01 crc kubenswrapper[4955]: I0217 13:22:01.776711 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c54198ed25e7a4d53de61cd3061ffbff1fcd3410dacaca9e71c03a327b392e0" Feb 17 13:22:01 crc kubenswrapper[4955]: I0217 13:22:01.776552 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-3d6f-account-create-update-zsh9p" Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.405117 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-4mktx" Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.411978 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-a0dc-account-create-update-rbbn6" Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.420434 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-6ss66" Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.427267 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-tqmd4" Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.434216 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ab8c-account-create-update-49rjx" Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.495650 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwrtz\" (UniqueName: \"kubernetes.io/projected/2fa427d9-5f2e-4cb9-934b-edd1086e0f42-kube-api-access-fwrtz\") pod \"2fa427d9-5f2e-4cb9-934b-edd1086e0f42\" (UID: \"2fa427d9-5f2e-4cb9-934b-edd1086e0f42\") " Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.495708 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pkk47\" (UniqueName: \"kubernetes.io/projected/4fe8b1c0-db08-4b58-b1cb-218631c424ab-kube-api-access-pkk47\") pod \"4fe8b1c0-db08-4b58-b1cb-218631c424ab\" (UID: \"4fe8b1c0-db08-4b58-b1cb-218631c424ab\") " Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.495743 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vr296\" (UniqueName: \"kubernetes.io/projected/4da596f3-7aff-4cfd-984e-0fd58d82a25c-kube-api-access-vr296\") pod \"4da596f3-7aff-4cfd-984e-0fd58d82a25c\" (UID: \"4da596f3-7aff-4cfd-984e-0fd58d82a25c\") " Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.495773 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2fa427d9-5f2e-4cb9-934b-edd1086e0f42-operator-scripts\") pod \"2fa427d9-5f2e-4cb9-934b-edd1086e0f42\" (UID: \"2fa427d9-5f2e-4cb9-934b-edd1086e0f42\") " Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.495912 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4fe8b1c0-db08-4b58-b1cb-218631c424ab-operator-scripts\") pod \"4fe8b1c0-db08-4b58-b1cb-218631c424ab\" (UID: \"4fe8b1c0-db08-4b58-b1cb-218631c424ab\") " Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.495964 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a61deb1d-ae88-49f8-8954-2059c469e84f-operator-scripts\") pod \"a61deb1d-ae88-49f8-8954-2059c469e84f\" (UID: \"a61deb1d-ae88-49f8-8954-2059c469e84f\") " Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.496009 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd1f70a3-aaee-464a-aa89-621279758849-operator-scripts\") pod \"dd1f70a3-aaee-464a-aa89-621279758849\" (UID: \"dd1f70a3-aaee-464a-aa89-621279758849\") " Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.496062 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-clsnh\" (UniqueName: \"kubernetes.io/projected/dd1f70a3-aaee-464a-aa89-621279758849-kube-api-access-clsnh\") pod \"dd1f70a3-aaee-464a-aa89-621279758849\" (UID: \"dd1f70a3-aaee-464a-aa89-621279758849\") " Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.496086 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7pffl\" (UniqueName: \"kubernetes.io/projected/a61deb1d-ae88-49f8-8954-2059c469e84f-kube-api-access-7pffl\") pod \"a61deb1d-ae88-49f8-8954-2059c469e84f\" (UID: \"a61deb1d-ae88-49f8-8954-2059c469e84f\") " Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.496122 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4da596f3-7aff-4cfd-984e-0fd58d82a25c-operator-scripts\") pod \"4da596f3-7aff-4cfd-984e-0fd58d82a25c\" (UID: \"4da596f3-7aff-4cfd-984e-0fd58d82a25c\") " Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.496795 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4da596f3-7aff-4cfd-984e-0fd58d82a25c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4da596f3-7aff-4cfd-984e-0fd58d82a25c" (UID: "4da596f3-7aff-4cfd-984e-0fd58d82a25c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.497129 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fa427d9-5f2e-4cb9-934b-edd1086e0f42-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2fa427d9-5f2e-4cb9-934b-edd1086e0f42" (UID: "2fa427d9-5f2e-4cb9-934b-edd1086e0f42"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.497458 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fe8b1c0-db08-4b58-b1cb-218631c424ab-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4fe8b1c0-db08-4b58-b1cb-218631c424ab" (UID: "4fe8b1c0-db08-4b58-b1cb-218631c424ab"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.497825 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd1f70a3-aaee-464a-aa89-621279758849-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dd1f70a3-aaee-464a-aa89-621279758849" (UID: "dd1f70a3-aaee-464a-aa89-621279758849"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.498172 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a61deb1d-ae88-49f8-8954-2059c469e84f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a61deb1d-ae88-49f8-8954-2059c469e84f" (UID: "a61deb1d-ae88-49f8-8954-2059c469e84f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.504596 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd1f70a3-aaee-464a-aa89-621279758849-kube-api-access-clsnh" (OuterVolumeSpecName: "kube-api-access-clsnh") pod "dd1f70a3-aaee-464a-aa89-621279758849" (UID: "dd1f70a3-aaee-464a-aa89-621279758849"). InnerVolumeSpecName "kube-api-access-clsnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.514391 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a61deb1d-ae88-49f8-8954-2059c469e84f-kube-api-access-7pffl" (OuterVolumeSpecName: "kube-api-access-7pffl") pod "a61deb1d-ae88-49f8-8954-2059c469e84f" (UID: "a61deb1d-ae88-49f8-8954-2059c469e84f"). InnerVolumeSpecName "kube-api-access-7pffl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.514933 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4da596f3-7aff-4cfd-984e-0fd58d82a25c-kube-api-access-vr296" (OuterVolumeSpecName: "kube-api-access-vr296") pod "4da596f3-7aff-4cfd-984e-0fd58d82a25c" (UID: "4da596f3-7aff-4cfd-984e-0fd58d82a25c"). InnerVolumeSpecName "kube-api-access-vr296". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.515058 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fa427d9-5f2e-4cb9-934b-edd1086e0f42-kube-api-access-fwrtz" (OuterVolumeSpecName: "kube-api-access-fwrtz") pod "2fa427d9-5f2e-4cb9-934b-edd1086e0f42" (UID: "2fa427d9-5f2e-4cb9-934b-edd1086e0f42"). InnerVolumeSpecName "kube-api-access-fwrtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.516680 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fe8b1c0-db08-4b58-b1cb-218631c424ab-kube-api-access-pkk47" (OuterVolumeSpecName: "kube-api-access-pkk47") pod "4fe8b1c0-db08-4b58-b1cb-218631c424ab" (UID: "4fe8b1c0-db08-4b58-b1cb-218631c424ab"). InnerVolumeSpecName "kube-api-access-pkk47". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.597885 4955 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4fe8b1c0-db08-4b58-b1cb-218631c424ab-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.597926 4955 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a61deb1d-ae88-49f8-8954-2059c469e84f-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.597936 4955 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd1f70a3-aaee-464a-aa89-621279758849-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.597946 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-clsnh\" (UniqueName: \"kubernetes.io/projected/dd1f70a3-aaee-464a-aa89-621279758849-kube-api-access-clsnh\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.597957 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7pffl\" (UniqueName: \"kubernetes.io/projected/a61deb1d-ae88-49f8-8954-2059c469e84f-kube-api-access-7pffl\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.597966 4955 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4da596f3-7aff-4cfd-984e-0fd58d82a25c-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.597975 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwrtz\" (UniqueName: \"kubernetes.io/projected/2fa427d9-5f2e-4cb9-934b-edd1086e0f42-kube-api-access-fwrtz\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.598000 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pkk47\" (UniqueName: \"kubernetes.io/projected/4fe8b1c0-db08-4b58-b1cb-218631c424ab-kube-api-access-pkk47\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.598010 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vr296\" (UniqueName: \"kubernetes.io/projected/4da596f3-7aff-4cfd-984e-0fd58d82a25c-kube-api-access-vr296\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.598018 4955 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2fa427d9-5f2e-4cb9-934b-edd1086e0f42-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.792349 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-6ss66" event={"ID":"dd1f70a3-aaee-464a-aa89-621279758849","Type":"ContainerDied","Data":"1ecb54d8110f542c643895c3b665030d9a44943d1d180d6c4f5a144c6e11d7e1"} Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.792668 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ecb54d8110f542c643895c3b665030d9a44943d1d180d6c4f5a144c6e11d7e1" Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.792393 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-6ss66" Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.794849 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-a0dc-account-create-update-rbbn6" event={"ID":"4da596f3-7aff-4cfd-984e-0fd58d82a25c","Type":"ContainerDied","Data":"e1234988ed9b1db48881db75126d8babe91b2113d0689e2ca9867da4887eabae"} Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.794889 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e1234988ed9b1db48881db75126d8babe91b2113d0689e2ca9867da4887eabae" Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.794957 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-a0dc-account-create-update-rbbn6" Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.799904 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-4mktx" event={"ID":"4fe8b1c0-db08-4b58-b1cb-218631c424ab","Type":"ContainerDied","Data":"20f7823a30494bc345da282f4205ce36402f554efe0012253a2802005b0e2988"} Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.799947 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="20f7823a30494bc345da282f4205ce36402f554efe0012253a2802005b0e2988" Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.800016 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-4mktx" Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.801807 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-tqmd4" event={"ID":"a61deb1d-ae88-49f8-8954-2059c469e84f","Type":"ContainerDied","Data":"a115081be771adfcbb27ab756e993e17aaedcd7dcb44ff2b04089e31c4a4d422"} Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.801839 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a115081be771adfcbb27ab756e993e17aaedcd7dcb44ff2b04089e31c4a4d422" Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.801892 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-tqmd4" Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.803607 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ab8c-account-create-update-49rjx" event={"ID":"2fa427d9-5f2e-4cb9-934b-edd1086e0f42","Type":"ContainerDied","Data":"73e39cfc63e27d4acd1274d16ddb162ad7985c648cc7997e05c1e4047e0986f0"} Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.803808 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73e39cfc63e27d4acd1274d16ddb162ad7985c648cc7997e05c1e4047e0986f0" Feb 17 13:22:03 crc kubenswrapper[4955]: I0217 13:22:03.803927 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ab8c-account-create-update-49rjx" Feb 17 13:22:04 crc kubenswrapper[4955]: I0217 13:22:04.332002 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5f59b8f679-xxclj" Feb 17 13:22:04 crc kubenswrapper[4955]: I0217 13:22:04.396823 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-dwwb9"] Feb 17 13:22:04 crc kubenswrapper[4955]: I0217 13:22:04.397049 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-dwwb9" podUID="61f387f2-8694-4f14-a7a7-f996a7ed8efe" containerName="dnsmasq-dns" containerID="cri-o://aee79c390be2fc0b558e8b284c1bbef59caa978e0391cb988dc7b7af84f05fe1" gracePeriod=10 Feb 17 13:22:04 crc kubenswrapper[4955]: I0217 13:22:04.813397 4955 generic.go:334] "Generic (PLEG): container finished" podID="61f387f2-8694-4f14-a7a7-f996a7ed8efe" containerID="aee79c390be2fc0b558e8b284c1bbef59caa978e0391cb988dc7b7af84f05fe1" exitCode=0 Feb 17 13:22:04 crc kubenswrapper[4955]: I0217 13:22:04.813472 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-dwwb9" event={"ID":"61f387f2-8694-4f14-a7a7-f996a7ed8efe","Type":"ContainerDied","Data":"aee79c390be2fc0b558e8b284c1bbef59caa978e0391cb988dc7b7af84f05fe1"} Feb 17 13:22:08 crc kubenswrapper[4955]: I0217 13:22:08.400559 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-dwwb9" Feb 17 13:22:08 crc kubenswrapper[4955]: I0217 13:22:08.486275 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-px2ht\" (UniqueName: \"kubernetes.io/projected/61f387f2-8694-4f14-a7a7-f996a7ed8efe-kube-api-access-px2ht\") pod \"61f387f2-8694-4f14-a7a7-f996a7ed8efe\" (UID: \"61f387f2-8694-4f14-a7a7-f996a7ed8efe\") " Feb 17 13:22:08 crc kubenswrapper[4955]: I0217 13:22:08.486332 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61f387f2-8694-4f14-a7a7-f996a7ed8efe-dns-svc\") pod \"61f387f2-8694-4f14-a7a7-f996a7ed8efe\" (UID: \"61f387f2-8694-4f14-a7a7-f996a7ed8efe\") " Feb 17 13:22:08 crc kubenswrapper[4955]: I0217 13:22:08.486383 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/61f387f2-8694-4f14-a7a7-f996a7ed8efe-ovsdbserver-nb\") pod \"61f387f2-8694-4f14-a7a7-f996a7ed8efe\" (UID: \"61f387f2-8694-4f14-a7a7-f996a7ed8efe\") " Feb 17 13:22:08 crc kubenswrapper[4955]: I0217 13:22:08.486498 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/61f387f2-8694-4f14-a7a7-f996a7ed8efe-ovsdbserver-sb\") pod \"61f387f2-8694-4f14-a7a7-f996a7ed8efe\" (UID: \"61f387f2-8694-4f14-a7a7-f996a7ed8efe\") " Feb 17 13:22:08 crc kubenswrapper[4955]: I0217 13:22:08.486525 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61f387f2-8694-4f14-a7a7-f996a7ed8efe-config\") pod \"61f387f2-8694-4f14-a7a7-f996a7ed8efe\" (UID: \"61f387f2-8694-4f14-a7a7-f996a7ed8efe\") " Feb 17 13:22:08 crc kubenswrapper[4955]: I0217 13:22:08.490623 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61f387f2-8694-4f14-a7a7-f996a7ed8efe-kube-api-access-px2ht" (OuterVolumeSpecName: "kube-api-access-px2ht") pod "61f387f2-8694-4f14-a7a7-f996a7ed8efe" (UID: "61f387f2-8694-4f14-a7a7-f996a7ed8efe"). InnerVolumeSpecName "kube-api-access-px2ht". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:22:08 crc kubenswrapper[4955]: I0217 13:22:08.525260 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61f387f2-8694-4f14-a7a7-f996a7ed8efe-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "61f387f2-8694-4f14-a7a7-f996a7ed8efe" (UID: "61f387f2-8694-4f14-a7a7-f996a7ed8efe"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:22:08 crc kubenswrapper[4955]: I0217 13:22:08.525274 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61f387f2-8694-4f14-a7a7-f996a7ed8efe-config" (OuterVolumeSpecName: "config") pod "61f387f2-8694-4f14-a7a7-f996a7ed8efe" (UID: "61f387f2-8694-4f14-a7a7-f996a7ed8efe"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:22:08 crc kubenswrapper[4955]: I0217 13:22:08.530905 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61f387f2-8694-4f14-a7a7-f996a7ed8efe-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "61f387f2-8694-4f14-a7a7-f996a7ed8efe" (UID: "61f387f2-8694-4f14-a7a7-f996a7ed8efe"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:22:08 crc kubenswrapper[4955]: I0217 13:22:08.532799 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61f387f2-8694-4f14-a7a7-f996a7ed8efe-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "61f387f2-8694-4f14-a7a7-f996a7ed8efe" (UID: "61f387f2-8694-4f14-a7a7-f996a7ed8efe"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:22:08 crc kubenswrapper[4955]: I0217 13:22:08.588453 4955 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/61f387f2-8694-4f14-a7a7-f996a7ed8efe-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:08 crc kubenswrapper[4955]: I0217 13:22:08.588491 4955 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61f387f2-8694-4f14-a7a7-f996a7ed8efe-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:08 crc kubenswrapper[4955]: I0217 13:22:08.588502 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-px2ht\" (UniqueName: \"kubernetes.io/projected/61f387f2-8694-4f14-a7a7-f996a7ed8efe-kube-api-access-px2ht\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:08 crc kubenswrapper[4955]: I0217 13:22:08.588512 4955 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61f387f2-8694-4f14-a7a7-f996a7ed8efe-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:08 crc kubenswrapper[4955]: I0217 13:22:08.588520 4955 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/61f387f2-8694-4f14-a7a7-f996a7ed8efe-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:08 crc kubenswrapper[4955]: I0217 13:22:08.864915 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-dwwb9" event={"ID":"61f387f2-8694-4f14-a7a7-f996a7ed8efe","Type":"ContainerDied","Data":"8c46dd48f25cc39cb6b50f19f918a7aa1bfd09bcb04e9e36799ff7039cf0bf84"} Feb 17 13:22:08 crc kubenswrapper[4955]: I0217 13:22:08.864977 4955 scope.go:117] "RemoveContainer" containerID="aee79c390be2fc0b558e8b284c1bbef59caa978e0391cb988dc7b7af84f05fe1" Feb 17 13:22:08 crc kubenswrapper[4955]: I0217 13:22:08.865031 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-dwwb9" Feb 17 13:22:08 crc kubenswrapper[4955]: I0217 13:22:08.874487 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-hcggw" event={"ID":"25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a","Type":"ContainerStarted","Data":"498a728e83f06ac0535fc1f340791bb545f474bf48353cf58ae5b6346f916cb8"} Feb 17 13:22:08 crc kubenswrapper[4955]: I0217 13:22:08.900977 4955 scope.go:117] "RemoveContainer" containerID="f2b4cbe4ffcb0fdc840eb2b1cbf94fb31dd007c9ce03477751ed41ca9969e12c" Feb 17 13:22:08 crc kubenswrapper[4955]: I0217 13:22:08.905369 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-hcggw" podStartSLOduration=2.762204972 podStartE2EDuration="11.905348805s" podCreationTimestamp="2026-02-17 13:21:57 +0000 UTC" firstStartedPulling="2026-02-17 13:21:59.109775342 +0000 UTC m=+1057.632504885" lastFinishedPulling="2026-02-17 13:22:08.252919175 +0000 UTC m=+1066.775648718" observedRunningTime="2026-02-17 13:22:08.901350301 +0000 UTC m=+1067.424079864" watchObservedRunningTime="2026-02-17 13:22:08.905348805 +0000 UTC m=+1067.428078348" Feb 17 13:22:08 crc kubenswrapper[4955]: I0217 13:22:08.930001 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-dwwb9"] Feb 17 13:22:08 crc kubenswrapper[4955]: I0217 13:22:08.936183 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-dwwb9"] Feb 17 13:22:10 crc kubenswrapper[4955]: I0217 13:22:10.237145 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61f387f2-8694-4f14-a7a7-f996a7ed8efe" path="/var/lib/kubelet/pods/61f387f2-8694-4f14-a7a7-f996a7ed8efe/volumes" Feb 17 13:22:11 crc kubenswrapper[4955]: I0217 13:22:11.916393 4955 generic.go:334] "Generic (PLEG): container finished" podID="25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a" containerID="498a728e83f06ac0535fc1f340791bb545f474bf48353cf58ae5b6346f916cb8" exitCode=0 Feb 17 13:22:11 crc kubenswrapper[4955]: I0217 13:22:11.916472 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-hcggw" event={"ID":"25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a","Type":"ContainerDied","Data":"498a728e83f06ac0535fc1f340791bb545f474bf48353cf58ae5b6346f916cb8"} Feb 17 13:22:13 crc kubenswrapper[4955]: I0217 13:22:13.208263 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-dwwb9" podUID="61f387f2-8694-4f14-a7a7-f996a7ed8efe" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.112:5353: i/o timeout" Feb 17 13:22:13 crc kubenswrapper[4955]: I0217 13:22:13.242996 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-hcggw" Feb 17 13:22:13 crc kubenswrapper[4955]: I0217 13:22:13.368959 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a-config-data\") pod \"25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a\" (UID: \"25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a\") " Feb 17 13:22:13 crc kubenswrapper[4955]: I0217 13:22:13.369038 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a-combined-ca-bundle\") pod \"25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a\" (UID: \"25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a\") " Feb 17 13:22:13 crc kubenswrapper[4955]: I0217 13:22:13.369171 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zbngc\" (UniqueName: \"kubernetes.io/projected/25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a-kube-api-access-zbngc\") pod \"25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a\" (UID: \"25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a\") " Feb 17 13:22:13 crc kubenswrapper[4955]: I0217 13:22:13.381004 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a-kube-api-access-zbngc" (OuterVolumeSpecName: "kube-api-access-zbngc") pod "25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a" (UID: "25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a"). InnerVolumeSpecName "kube-api-access-zbngc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:22:13 crc kubenswrapper[4955]: I0217 13:22:13.405091 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a" (UID: "25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:22:13 crc kubenswrapper[4955]: I0217 13:22:13.435718 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a-config-data" (OuterVolumeSpecName: "config-data") pod "25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a" (UID: "25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:22:13 crc kubenswrapper[4955]: I0217 13:22:13.470761 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zbngc\" (UniqueName: \"kubernetes.io/projected/25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a-kube-api-access-zbngc\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:13 crc kubenswrapper[4955]: I0217 13:22:13.470822 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:13 crc kubenswrapper[4955]: I0217 13:22:13.470837 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:13 crc kubenswrapper[4955]: I0217 13:22:13.935545 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-hcggw" event={"ID":"25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a","Type":"ContainerDied","Data":"95af02a5119e291c0333fd97ecd5f746454fddc015b9c3f7fce849b8b111d445"} Feb 17 13:22:13 crc kubenswrapper[4955]: I0217 13:22:13.935896 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="95af02a5119e291c0333fd97ecd5f746454fddc015b9c3f7fce849b8b111d445" Feb 17 13:22:13 crc kubenswrapper[4955]: I0217 13:22:13.935636 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-hcggw" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.245422 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-4pqz5"] Feb 17 13:22:14 crc kubenswrapper[4955]: E0217 13:22:14.245723 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4da596f3-7aff-4cfd-984e-0fd58d82a25c" containerName="mariadb-account-create-update" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.245735 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="4da596f3-7aff-4cfd-984e-0fd58d82a25c" containerName="mariadb-account-create-update" Feb 17 13:22:14 crc kubenswrapper[4955]: E0217 13:22:14.245752 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61f387f2-8694-4f14-a7a7-f996a7ed8efe" containerName="init" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.245758 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="61f387f2-8694-4f14-a7a7-f996a7ed8efe" containerName="init" Feb 17 13:22:14 crc kubenswrapper[4955]: E0217 13:22:14.245765 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cb66d5d-60a5-486d-b9ff-35e2ebeb4d8b" containerName="mariadb-account-create-update" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.245771 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cb66d5d-60a5-486d-b9ff-35e2ebeb4d8b" containerName="mariadb-account-create-update" Feb 17 13:22:14 crc kubenswrapper[4955]: E0217 13:22:14.245796 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fa427d9-5f2e-4cb9-934b-edd1086e0f42" containerName="mariadb-account-create-update" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.245804 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fa427d9-5f2e-4cb9-934b-edd1086e0f42" containerName="mariadb-account-create-update" Feb 17 13:22:14 crc kubenswrapper[4955]: E0217 13:22:14.245814 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fe8b1c0-db08-4b58-b1cb-218631c424ab" containerName="mariadb-database-create" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.245820 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fe8b1c0-db08-4b58-b1cb-218631c424ab" containerName="mariadb-database-create" Feb 17 13:22:14 crc kubenswrapper[4955]: E0217 13:22:14.245832 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a" containerName="keystone-db-sync" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.245838 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a" containerName="keystone-db-sync" Feb 17 13:22:14 crc kubenswrapper[4955]: E0217 13:22:14.245860 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61f387f2-8694-4f14-a7a7-f996a7ed8efe" containerName="dnsmasq-dns" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.245866 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="61f387f2-8694-4f14-a7a7-f996a7ed8efe" containerName="dnsmasq-dns" Feb 17 13:22:14 crc kubenswrapper[4955]: E0217 13:22:14.245877 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a61deb1d-ae88-49f8-8954-2059c469e84f" containerName="mariadb-database-create" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.245882 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="a61deb1d-ae88-49f8-8954-2059c469e84f" containerName="mariadb-database-create" Feb 17 13:22:14 crc kubenswrapper[4955]: E0217 13:22:14.245892 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd1f70a3-aaee-464a-aa89-621279758849" containerName="mariadb-database-create" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.245899 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd1f70a3-aaee-464a-aa89-621279758849" containerName="mariadb-database-create" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.246031 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a" containerName="keystone-db-sync" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.246044 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="4fe8b1c0-db08-4b58-b1cb-218631c424ab" containerName="mariadb-database-create" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.246050 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd1f70a3-aaee-464a-aa89-621279758849" containerName="mariadb-database-create" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.246060 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fa427d9-5f2e-4cb9-934b-edd1086e0f42" containerName="mariadb-account-create-update" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.246070 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cb66d5d-60a5-486d-b9ff-35e2ebeb4d8b" containerName="mariadb-account-create-update" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.246079 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="61f387f2-8694-4f14-a7a7-f996a7ed8efe" containerName="dnsmasq-dns" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.246086 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="a61deb1d-ae88-49f8-8954-2059c469e84f" containerName="mariadb-database-create" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.246097 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="4da596f3-7aff-4cfd-984e-0fd58d82a25c" containerName="mariadb-account-create-update" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.246579 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4pqz5" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.257700 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.257968 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.258043 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.258212 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.258405 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-qh96v" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.260069 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-wr4x8"] Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.268324 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-wr4x8" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.289347 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-wr4x8"] Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.301383 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-4pqz5"] Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.388230 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e-credential-keys\") pod \"keystone-bootstrap-4pqz5\" (UID: \"fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e\") " pod="openstack/keystone-bootstrap-4pqz5" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.388283 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e-scripts\") pod \"keystone-bootstrap-4pqz5\" (UID: \"fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e\") " pod="openstack/keystone-bootstrap-4pqz5" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.388315 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-wr4x8\" (UID: \"e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2\") " pod="openstack/dnsmasq-dns-bbf5cc879-wr4x8" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.388338 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e-config-data\") pod \"keystone-bootstrap-4pqz5\" (UID: \"fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e\") " pod="openstack/keystone-bootstrap-4pqz5" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.388362 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-wr4x8\" (UID: \"e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2\") " pod="openstack/dnsmasq-dns-bbf5cc879-wr4x8" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.388399 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-wr4x8\" (UID: \"e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2\") " pod="openstack/dnsmasq-dns-bbf5cc879-wr4x8" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.388483 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2-config\") pod \"dnsmasq-dns-bbf5cc879-wr4x8\" (UID: \"e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2\") " pod="openstack/dnsmasq-dns-bbf5cc879-wr4x8" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.388504 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-wr4x8\" (UID: \"e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2\") " pod="openstack/dnsmasq-dns-bbf5cc879-wr4x8" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.388537 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fj7rl\" (UniqueName: \"kubernetes.io/projected/fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e-kube-api-access-fj7rl\") pod \"keystone-bootstrap-4pqz5\" (UID: \"fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e\") " pod="openstack/keystone-bootstrap-4pqz5" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.388560 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tx2q4\" (UniqueName: \"kubernetes.io/projected/e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2-kube-api-access-tx2q4\") pod \"dnsmasq-dns-bbf5cc879-wr4x8\" (UID: \"e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2\") " pod="openstack/dnsmasq-dns-bbf5cc879-wr4x8" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.388594 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e-fernet-keys\") pod \"keystone-bootstrap-4pqz5\" (UID: \"fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e\") " pod="openstack/keystone-bootstrap-4pqz5" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.388624 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e-combined-ca-bundle\") pod \"keystone-bootstrap-4pqz5\" (UID: \"fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e\") " pod="openstack/keystone-bootstrap-4pqz5" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.490546 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2-config\") pod \"dnsmasq-dns-bbf5cc879-wr4x8\" (UID: \"e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2\") " pod="openstack/dnsmasq-dns-bbf5cc879-wr4x8" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.490597 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-wr4x8\" (UID: \"e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2\") " pod="openstack/dnsmasq-dns-bbf5cc879-wr4x8" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.490637 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fj7rl\" (UniqueName: \"kubernetes.io/projected/fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e-kube-api-access-fj7rl\") pod \"keystone-bootstrap-4pqz5\" (UID: \"fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e\") " pod="openstack/keystone-bootstrap-4pqz5" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.490658 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tx2q4\" (UniqueName: \"kubernetes.io/projected/e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2-kube-api-access-tx2q4\") pod \"dnsmasq-dns-bbf5cc879-wr4x8\" (UID: \"e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2\") " pod="openstack/dnsmasq-dns-bbf5cc879-wr4x8" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.490693 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e-fernet-keys\") pod \"keystone-bootstrap-4pqz5\" (UID: \"fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e\") " pod="openstack/keystone-bootstrap-4pqz5" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.490720 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e-combined-ca-bundle\") pod \"keystone-bootstrap-4pqz5\" (UID: \"fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e\") " pod="openstack/keystone-bootstrap-4pqz5" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.490752 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e-credential-keys\") pod \"keystone-bootstrap-4pqz5\" (UID: \"fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e\") " pod="openstack/keystone-bootstrap-4pqz5" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.490792 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e-scripts\") pod \"keystone-bootstrap-4pqz5\" (UID: \"fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e\") " pod="openstack/keystone-bootstrap-4pqz5" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.490822 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-wr4x8\" (UID: \"e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2\") " pod="openstack/dnsmasq-dns-bbf5cc879-wr4x8" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.490850 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e-config-data\") pod \"keystone-bootstrap-4pqz5\" (UID: \"fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e\") " pod="openstack/keystone-bootstrap-4pqz5" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.490883 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-wr4x8\" (UID: \"e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2\") " pod="openstack/dnsmasq-dns-bbf5cc879-wr4x8" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.490918 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-wr4x8\" (UID: \"e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2\") " pod="openstack/dnsmasq-dns-bbf5cc879-wr4x8" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.492097 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-wr4x8\" (UID: \"e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2\") " pod="openstack/dnsmasq-dns-bbf5cc879-wr4x8" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.492651 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2-config\") pod \"dnsmasq-dns-bbf5cc879-wr4x8\" (UID: \"e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2\") " pod="openstack/dnsmasq-dns-bbf5cc879-wr4x8" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.495084 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-wr4x8\" (UID: \"e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2\") " pod="openstack/dnsmasq-dns-bbf5cc879-wr4x8" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.504744 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-wr4x8\" (UID: \"e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2\") " pod="openstack/dnsmasq-dns-bbf5cc879-wr4x8" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.505386 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-wr4x8\" (UID: \"e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2\") " pod="openstack/dnsmasq-dns-bbf5cc879-wr4x8" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.513997 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e-config-data\") pod \"keystone-bootstrap-4pqz5\" (UID: \"fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e\") " pod="openstack/keystone-bootstrap-4pqz5" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.515179 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e-credential-keys\") pod \"keystone-bootstrap-4pqz5\" (UID: \"fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e\") " pod="openstack/keystone-bootstrap-4pqz5" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.522892 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e-fernet-keys\") pod \"keystone-bootstrap-4pqz5\" (UID: \"fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e\") " pod="openstack/keystone-bootstrap-4pqz5" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.523288 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e-combined-ca-bundle\") pod \"keystone-bootstrap-4pqz5\" (UID: \"fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e\") " pod="openstack/keystone-bootstrap-4pqz5" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.560829 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fj7rl\" (UniqueName: \"kubernetes.io/projected/fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e-kube-api-access-fj7rl\") pod \"keystone-bootstrap-4pqz5\" (UID: \"fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e\") " pod="openstack/keystone-bootstrap-4pqz5" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.570368 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-76cdc74c45-lss64"] Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.575223 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-76cdc74c45-lss64" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.577092 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e-scripts\") pod \"keystone-bootstrap-4pqz5\" (UID: \"fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e\") " pod="openstack/keystone-bootstrap-4pqz5" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.580177 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tx2q4\" (UniqueName: \"kubernetes.io/projected/e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2-kube-api-access-tx2q4\") pod \"dnsmasq-dns-bbf5cc879-wr4x8\" (UID: \"e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2\") " pod="openstack/dnsmasq-dns-bbf5cc879-wr4x8" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.603288 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4pqz5" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.603984 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-wr4x8" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.604401 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.604475 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.604645 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.604425 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-8qqrs" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.703603 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-76cdc74c45-lss64"] Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.708145 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/dd349c9d-f6b6-48b5-a22f-e2c1d81f6711-horizon-secret-key\") pod \"horizon-76cdc74c45-lss64\" (UID: \"dd349c9d-f6b6-48b5-a22f-e2c1d81f6711\") " pod="openstack/horizon-76cdc74c45-lss64" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.708238 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd349c9d-f6b6-48b5-a22f-e2c1d81f6711-scripts\") pod \"horizon-76cdc74c45-lss64\" (UID: \"dd349c9d-f6b6-48b5-a22f-e2c1d81f6711\") " pod="openstack/horizon-76cdc74c45-lss64" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.708268 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2k65n\" (UniqueName: \"kubernetes.io/projected/dd349c9d-f6b6-48b5-a22f-e2c1d81f6711-kube-api-access-2k65n\") pod \"horizon-76cdc74c45-lss64\" (UID: \"dd349c9d-f6b6-48b5-a22f-e2c1d81f6711\") " pod="openstack/horizon-76cdc74c45-lss64" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.708303 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dd349c9d-f6b6-48b5-a22f-e2c1d81f6711-config-data\") pod \"horizon-76cdc74c45-lss64\" (UID: \"dd349c9d-f6b6-48b5-a22f-e2c1d81f6711\") " pod="openstack/horizon-76cdc74c45-lss64" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.708337 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd349c9d-f6b6-48b5-a22f-e2c1d81f6711-logs\") pod \"horizon-76cdc74c45-lss64\" (UID: \"dd349c9d-f6b6-48b5-a22f-e2c1d81f6711\") " pod="openstack/horizon-76cdc74c45-lss64" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.733844 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-w6s4s"] Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.735194 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-w6s4s" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.747454 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.747678 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.748535 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-zw4xp" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.810289 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd349c9d-f6b6-48b5-a22f-e2c1d81f6711-scripts\") pod \"horizon-76cdc74c45-lss64\" (UID: \"dd349c9d-f6b6-48b5-a22f-e2c1d81f6711\") " pod="openstack/horizon-76cdc74c45-lss64" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.811981 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2k65n\" (UniqueName: \"kubernetes.io/projected/dd349c9d-f6b6-48b5-a22f-e2c1d81f6711-kube-api-access-2k65n\") pod \"horizon-76cdc74c45-lss64\" (UID: \"dd349c9d-f6b6-48b5-a22f-e2c1d81f6711\") " pod="openstack/horizon-76cdc74c45-lss64" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.812042 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dd349c9d-f6b6-48b5-a22f-e2c1d81f6711-config-data\") pod \"horizon-76cdc74c45-lss64\" (UID: \"dd349c9d-f6b6-48b5-a22f-e2c1d81f6711\") " pod="openstack/horizon-76cdc74c45-lss64" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.812080 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd349c9d-f6b6-48b5-a22f-e2c1d81f6711-logs\") pod \"horizon-76cdc74c45-lss64\" (UID: \"dd349c9d-f6b6-48b5-a22f-e2c1d81f6711\") " pod="openstack/horizon-76cdc74c45-lss64" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.812106 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1c2d0c5-7647-447f-a1a6-b37d8c09310c-config-data\") pod \"cinder-db-sync-w6s4s\" (UID: \"e1c2d0c5-7647-447f-a1a6-b37d8c09310c\") " pod="openstack/cinder-db-sync-w6s4s" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.812162 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e1c2d0c5-7647-447f-a1a6-b37d8c09310c-etc-machine-id\") pod \"cinder-db-sync-w6s4s\" (UID: \"e1c2d0c5-7647-447f-a1a6-b37d8c09310c\") " pod="openstack/cinder-db-sync-w6s4s" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.812209 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1c2d0c5-7647-447f-a1a6-b37d8c09310c-scripts\") pod \"cinder-db-sync-w6s4s\" (UID: \"e1c2d0c5-7647-447f-a1a6-b37d8c09310c\") " pod="openstack/cinder-db-sync-w6s4s" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.812276 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e1c2d0c5-7647-447f-a1a6-b37d8c09310c-db-sync-config-data\") pod \"cinder-db-sync-w6s4s\" (UID: \"e1c2d0c5-7647-447f-a1a6-b37d8c09310c\") " pod="openstack/cinder-db-sync-w6s4s" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.812304 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1c2d0c5-7647-447f-a1a6-b37d8c09310c-combined-ca-bundle\") pod \"cinder-db-sync-w6s4s\" (UID: \"e1c2d0c5-7647-447f-a1a6-b37d8c09310c\") " pod="openstack/cinder-db-sync-w6s4s" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.812331 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/dd349c9d-f6b6-48b5-a22f-e2c1d81f6711-horizon-secret-key\") pod \"horizon-76cdc74c45-lss64\" (UID: \"dd349c9d-f6b6-48b5-a22f-e2c1d81f6711\") " pod="openstack/horizon-76cdc74c45-lss64" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.812358 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6w2b6\" (UniqueName: \"kubernetes.io/projected/e1c2d0c5-7647-447f-a1a6-b37d8c09310c-kube-api-access-6w2b6\") pod \"cinder-db-sync-w6s4s\" (UID: \"e1c2d0c5-7647-447f-a1a6-b37d8c09310c\") " pod="openstack/cinder-db-sync-w6s4s" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.814135 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dd349c9d-f6b6-48b5-a22f-e2c1d81f6711-config-data\") pod \"horizon-76cdc74c45-lss64\" (UID: \"dd349c9d-f6b6-48b5-a22f-e2c1d81f6711\") " pod="openstack/horizon-76cdc74c45-lss64" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.814397 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd349c9d-f6b6-48b5-a22f-e2c1d81f6711-logs\") pod \"horizon-76cdc74c45-lss64\" (UID: \"dd349c9d-f6b6-48b5-a22f-e2c1d81f6711\") " pod="openstack/horizon-76cdc74c45-lss64" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.820423 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd349c9d-f6b6-48b5-a22f-e2c1d81f6711-scripts\") pod \"horizon-76cdc74c45-lss64\" (UID: \"dd349c9d-f6b6-48b5-a22f-e2c1d81f6711\") " pod="openstack/horizon-76cdc74c45-lss64" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.835622 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/dd349c9d-f6b6-48b5-a22f-e2c1d81f6711-horizon-secret-key\") pod \"horizon-76cdc74c45-lss64\" (UID: \"dd349c9d-f6b6-48b5-a22f-e2c1d81f6711\") " pod="openstack/horizon-76cdc74c45-lss64" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.836077 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-w6s4s"] Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.845556 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2k65n\" (UniqueName: \"kubernetes.io/projected/dd349c9d-f6b6-48b5-a22f-e2c1d81f6711-kube-api-access-2k65n\") pod \"horizon-76cdc74c45-lss64\" (UID: \"dd349c9d-f6b6-48b5-a22f-e2c1d81f6711\") " pod="openstack/horizon-76cdc74c45-lss64" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.872171 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-tr2q9"] Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.873342 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6cf59496f7-6ntrd"] Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.874450 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6cf59496f7-6ntrd" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.875295 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-tr2q9" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.882435 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.882661 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-tttcl" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.905866 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-48hsq"] Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.907186 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-48hsq" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.914315 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1d695519-3143-4f53-ab46-5823bc04dc0d-db-sync-config-data\") pod \"barbican-db-sync-tr2q9\" (UID: \"1d695519-3143-4f53-ab46-5823bc04dc0d\") " pod="openstack/barbican-db-sync-tr2q9" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.914480 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d1b9c8f8-6829-4eaf-b5bf-48121eaaf818-logs\") pod \"horizon-6cf59496f7-6ntrd\" (UID: \"d1b9c8f8-6829-4eaf-b5bf-48121eaaf818\") " pod="openstack/horizon-6cf59496f7-6ntrd" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.914530 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d1b9c8f8-6829-4eaf-b5bf-48121eaaf818-config-data\") pod \"horizon-6cf59496f7-6ntrd\" (UID: \"d1b9c8f8-6829-4eaf-b5bf-48121eaaf818\") " pod="openstack/horizon-6cf59496f7-6ntrd" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.914546 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d1b9c8f8-6829-4eaf-b5bf-48121eaaf818-scripts\") pod \"horizon-6cf59496f7-6ntrd\" (UID: \"d1b9c8f8-6829-4eaf-b5bf-48121eaaf818\") " pod="openstack/horizon-6cf59496f7-6ntrd" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.914569 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1c2d0c5-7647-447f-a1a6-b37d8c09310c-config-data\") pod \"cinder-db-sync-w6s4s\" (UID: \"e1c2d0c5-7647-447f-a1a6-b37d8c09310c\") " pod="openstack/cinder-db-sync-w6s4s" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.914604 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e1c2d0c5-7647-447f-a1a6-b37d8c09310c-etc-machine-id\") pod \"cinder-db-sync-w6s4s\" (UID: \"e1c2d0c5-7647-447f-a1a6-b37d8c09310c\") " pod="openstack/cinder-db-sync-w6s4s" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.914634 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1c2d0c5-7647-447f-a1a6-b37d8c09310c-scripts\") pod \"cinder-db-sync-w6s4s\" (UID: \"e1c2d0c5-7647-447f-a1a6-b37d8c09310c\") " pod="openstack/cinder-db-sync-w6s4s" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.914653 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bv2g2\" (UniqueName: \"kubernetes.io/projected/d1b9c8f8-6829-4eaf-b5bf-48121eaaf818-kube-api-access-bv2g2\") pod \"horizon-6cf59496f7-6ntrd\" (UID: \"d1b9c8f8-6829-4eaf-b5bf-48121eaaf818\") " pod="openstack/horizon-6cf59496f7-6ntrd" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.914688 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d695519-3143-4f53-ab46-5823bc04dc0d-combined-ca-bundle\") pod \"barbican-db-sync-tr2q9\" (UID: \"1d695519-3143-4f53-ab46-5823bc04dc0d\") " pod="openstack/barbican-db-sync-tr2q9" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.914713 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e1c2d0c5-7647-447f-a1a6-b37d8c09310c-db-sync-config-data\") pod \"cinder-db-sync-w6s4s\" (UID: \"e1c2d0c5-7647-447f-a1a6-b37d8c09310c\") " pod="openstack/cinder-db-sync-w6s4s" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.914733 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1c2d0c5-7647-447f-a1a6-b37d8c09310c-combined-ca-bundle\") pod \"cinder-db-sync-w6s4s\" (UID: \"e1c2d0c5-7647-447f-a1a6-b37d8c09310c\") " pod="openstack/cinder-db-sync-w6s4s" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.914754 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rl7sq\" (UniqueName: \"kubernetes.io/projected/1d695519-3143-4f53-ab46-5823bc04dc0d-kube-api-access-rl7sq\") pod \"barbican-db-sync-tr2q9\" (UID: \"1d695519-3143-4f53-ab46-5823bc04dc0d\") " pod="openstack/barbican-db-sync-tr2q9" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.915681 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6w2b6\" (UniqueName: \"kubernetes.io/projected/e1c2d0c5-7647-447f-a1a6-b37d8c09310c-kube-api-access-6w2b6\") pod \"cinder-db-sync-w6s4s\" (UID: \"e1c2d0c5-7647-447f-a1a6-b37d8c09310c\") " pod="openstack/cinder-db-sync-w6s4s" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.915716 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d1b9c8f8-6829-4eaf-b5bf-48121eaaf818-horizon-secret-key\") pod \"horizon-6cf59496f7-6ntrd\" (UID: \"d1b9c8f8-6829-4eaf-b5bf-48121eaaf818\") " pod="openstack/horizon-6cf59496f7-6ntrd" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.921193 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.921412 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-kcgqq" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.921567 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.926401 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e1c2d0c5-7647-447f-a1a6-b37d8c09310c-etc-machine-id\") pod \"cinder-db-sync-w6s4s\" (UID: \"e1c2d0c5-7647-447f-a1a6-b37d8c09310c\") " pod="openstack/cinder-db-sync-w6s4s" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.932296 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1c2d0c5-7647-447f-a1a6-b37d8c09310c-combined-ca-bundle\") pod \"cinder-db-sync-w6s4s\" (UID: \"e1c2d0c5-7647-447f-a1a6-b37d8c09310c\") " pod="openstack/cinder-db-sync-w6s4s" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.932320 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1c2d0c5-7647-447f-a1a6-b37d8c09310c-config-data\") pod \"cinder-db-sync-w6s4s\" (UID: \"e1c2d0c5-7647-447f-a1a6-b37d8c09310c\") " pod="openstack/cinder-db-sync-w6s4s" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.933200 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1c2d0c5-7647-447f-a1a6-b37d8c09310c-scripts\") pod \"cinder-db-sync-w6s4s\" (UID: \"e1c2d0c5-7647-447f-a1a6-b37d8c09310c\") " pod="openstack/cinder-db-sync-w6s4s" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.933890 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e1c2d0c5-7647-447f-a1a6-b37d8c09310c-db-sync-config-data\") pod \"cinder-db-sync-w6s4s\" (UID: \"e1c2d0c5-7647-447f-a1a6-b37d8c09310c\") " pod="openstack/cinder-db-sync-w6s4s" Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.960885 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-tr2q9"] Feb 17 13:22:14 crc kubenswrapper[4955]: I0217 13:22:14.997535 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6cf59496f7-6ntrd"] Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.015412 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-48hsq"] Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.017466 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d1b9c8f8-6829-4eaf-b5bf-48121eaaf818-horizon-secret-key\") pod \"horizon-6cf59496f7-6ntrd\" (UID: \"d1b9c8f8-6829-4eaf-b5bf-48121eaaf818\") " pod="openstack/horizon-6cf59496f7-6ntrd" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.017507 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/be85ffb8-5128-4903-9f7d-cdd462a44b95-config\") pod \"neutron-db-sync-48hsq\" (UID: \"be85ffb8-5128-4903-9f7d-cdd462a44b95\") " pod="openstack/neutron-db-sync-48hsq" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.017540 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1d695519-3143-4f53-ab46-5823bc04dc0d-db-sync-config-data\") pod \"barbican-db-sync-tr2q9\" (UID: \"1d695519-3143-4f53-ab46-5823bc04dc0d\") " pod="openstack/barbican-db-sync-tr2q9" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.017575 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d1b9c8f8-6829-4eaf-b5bf-48121eaaf818-logs\") pod \"horizon-6cf59496f7-6ntrd\" (UID: \"d1b9c8f8-6829-4eaf-b5bf-48121eaaf818\") " pod="openstack/horizon-6cf59496f7-6ntrd" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.017613 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d1b9c8f8-6829-4eaf-b5bf-48121eaaf818-config-data\") pod \"horizon-6cf59496f7-6ntrd\" (UID: \"d1b9c8f8-6829-4eaf-b5bf-48121eaaf818\") " pod="openstack/horizon-6cf59496f7-6ntrd" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.017634 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d1b9c8f8-6829-4eaf-b5bf-48121eaaf818-scripts\") pod \"horizon-6cf59496f7-6ntrd\" (UID: \"d1b9c8f8-6829-4eaf-b5bf-48121eaaf818\") " pod="openstack/horizon-6cf59496f7-6ntrd" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.017672 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qhbv\" (UniqueName: \"kubernetes.io/projected/be85ffb8-5128-4903-9f7d-cdd462a44b95-kube-api-access-7qhbv\") pod \"neutron-db-sync-48hsq\" (UID: \"be85ffb8-5128-4903-9f7d-cdd462a44b95\") " pod="openstack/neutron-db-sync-48hsq" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.017727 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be85ffb8-5128-4903-9f7d-cdd462a44b95-combined-ca-bundle\") pod \"neutron-db-sync-48hsq\" (UID: \"be85ffb8-5128-4903-9f7d-cdd462a44b95\") " pod="openstack/neutron-db-sync-48hsq" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.017752 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bv2g2\" (UniqueName: \"kubernetes.io/projected/d1b9c8f8-6829-4eaf-b5bf-48121eaaf818-kube-api-access-bv2g2\") pod \"horizon-6cf59496f7-6ntrd\" (UID: \"d1b9c8f8-6829-4eaf-b5bf-48121eaaf818\") " pod="openstack/horizon-6cf59496f7-6ntrd" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.017824 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d695519-3143-4f53-ab46-5823bc04dc0d-combined-ca-bundle\") pod \"barbican-db-sync-tr2q9\" (UID: \"1d695519-3143-4f53-ab46-5823bc04dc0d\") " pod="openstack/barbican-db-sync-tr2q9" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.017862 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rl7sq\" (UniqueName: \"kubernetes.io/projected/1d695519-3143-4f53-ab46-5823bc04dc0d-kube-api-access-rl7sq\") pod \"barbican-db-sync-tr2q9\" (UID: \"1d695519-3143-4f53-ab46-5823bc04dc0d\") " pod="openstack/barbican-db-sync-tr2q9" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.018707 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6w2b6\" (UniqueName: \"kubernetes.io/projected/e1c2d0c5-7647-447f-a1a6-b37d8c09310c-kube-api-access-6w2b6\") pod \"cinder-db-sync-w6s4s\" (UID: \"e1c2d0c5-7647-447f-a1a6-b37d8c09310c\") " pod="openstack/cinder-db-sync-w6s4s" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.019270 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d1b9c8f8-6829-4eaf-b5bf-48121eaaf818-scripts\") pod \"horizon-6cf59496f7-6ntrd\" (UID: \"d1b9c8f8-6829-4eaf-b5bf-48121eaaf818\") " pod="openstack/horizon-6cf59496f7-6ntrd" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.019837 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d1b9c8f8-6829-4eaf-b5bf-48121eaaf818-logs\") pod \"horizon-6cf59496f7-6ntrd\" (UID: \"d1b9c8f8-6829-4eaf-b5bf-48121eaaf818\") " pod="openstack/horizon-6cf59496f7-6ntrd" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.029296 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.030486 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d695519-3143-4f53-ab46-5823bc04dc0d-combined-ca-bundle\") pod \"barbican-db-sync-tr2q9\" (UID: \"1d695519-3143-4f53-ab46-5823bc04dc0d\") " pod="openstack/barbican-db-sync-tr2q9" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.033191 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.036493 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.036711 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.044943 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d1b9c8f8-6829-4eaf-b5bf-48121eaaf818-config-data\") pod \"horizon-6cf59496f7-6ntrd\" (UID: \"d1b9c8f8-6829-4eaf-b5bf-48121eaaf818\") " pod="openstack/horizon-6cf59496f7-6ntrd" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.057828 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rl7sq\" (UniqueName: \"kubernetes.io/projected/1d695519-3143-4f53-ab46-5823bc04dc0d-kube-api-access-rl7sq\") pod \"barbican-db-sync-tr2q9\" (UID: \"1d695519-3143-4f53-ab46-5823bc04dc0d\") " pod="openstack/barbican-db-sync-tr2q9" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.061221 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d1b9c8f8-6829-4eaf-b5bf-48121eaaf818-horizon-secret-key\") pod \"horizon-6cf59496f7-6ntrd\" (UID: \"d1b9c8f8-6829-4eaf-b5bf-48121eaaf818\") " pod="openstack/horizon-6cf59496f7-6ntrd" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.061228 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1d695519-3143-4f53-ab46-5823bc04dc0d-db-sync-config-data\") pod \"barbican-db-sync-tr2q9\" (UID: \"1d695519-3143-4f53-ab46-5823bc04dc0d\") " pod="openstack/barbican-db-sync-tr2q9" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.078695 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.084278 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-76cdc74c45-lss64" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.089831 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-wr4x8"] Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.097927 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-nfdxn"] Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.103889 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-w6s4s" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.104061 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bv2g2\" (UniqueName: \"kubernetes.io/projected/d1b9c8f8-6829-4eaf-b5bf-48121eaaf818-kube-api-access-bv2g2\") pod \"horizon-6cf59496f7-6ntrd\" (UID: \"d1b9c8f8-6829-4eaf-b5bf-48121eaaf818\") " pod="openstack/horizon-6cf59496f7-6ntrd" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.120175 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/be85ffb8-5128-4903-9f7d-cdd462a44b95-config\") pod \"neutron-db-sync-48hsq\" (UID: \"be85ffb8-5128-4903-9f7d-cdd462a44b95\") " pod="openstack/neutron-db-sync-48hsq" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.120282 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qhbv\" (UniqueName: \"kubernetes.io/projected/be85ffb8-5128-4903-9f7d-cdd462a44b95-kube-api-access-7qhbv\") pod \"neutron-db-sync-48hsq\" (UID: \"be85ffb8-5128-4903-9f7d-cdd462a44b95\") " pod="openstack/neutron-db-sync-48hsq" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.120346 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be85ffb8-5128-4903-9f7d-cdd462a44b95-combined-ca-bundle\") pod \"neutron-db-sync-48hsq\" (UID: \"be85ffb8-5128-4903-9f7d-cdd462a44b95\") " pod="openstack/neutron-db-sync-48hsq" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.123457 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nfdxn" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.124124 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-nfdxn"] Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.126666 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.126843 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.126965 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-llchf" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.137049 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be85ffb8-5128-4903-9f7d-cdd462a44b95-combined-ca-bundle\") pod \"neutron-db-sync-48hsq\" (UID: \"be85ffb8-5128-4903-9f7d-cdd462a44b95\") " pod="openstack/neutron-db-sync-48hsq" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.141431 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/be85ffb8-5128-4903-9f7d-cdd462a44b95-config\") pod \"neutron-db-sync-48hsq\" (UID: \"be85ffb8-5128-4903-9f7d-cdd462a44b95\") " pod="openstack/neutron-db-sync-48hsq" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.185710 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-k2p8k"] Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.187831 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-k2p8k" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.197047 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-k2p8k"] Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.218302 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qhbv\" (UniqueName: \"kubernetes.io/projected/be85ffb8-5128-4903-9f7d-cdd462a44b95-kube-api-access-7qhbv\") pod \"neutron-db-sync-48hsq\" (UID: \"be85ffb8-5128-4903-9f7d-cdd462a44b95\") " pod="openstack/neutron-db-sync-48hsq" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.222389 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01547c5a-5ca7-435b-ba75-52e075800cf3-log-httpd\") pod \"ceilometer-0\" (UID: \"01547c5a-5ca7-435b-ba75-52e075800cf3\") " pod="openstack/ceilometer-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.222462 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01547c5a-5ca7-435b-ba75-52e075800cf3-config-data\") pod \"ceilometer-0\" (UID: \"01547c5a-5ca7-435b-ba75-52e075800cf3\") " pod="openstack/ceilometer-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.222515 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01547c5a-5ca7-435b-ba75-52e075800cf3-run-httpd\") pod \"ceilometer-0\" (UID: \"01547c5a-5ca7-435b-ba75-52e075800cf3\") " pod="openstack/ceilometer-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.222541 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01547c5a-5ca7-435b-ba75-52e075800cf3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"01547c5a-5ca7-435b-ba75-52e075800cf3\") " pod="openstack/ceilometer-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.222575 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/700f73a1-f56d-4d4e-9822-438472980124-scripts\") pod \"placement-db-sync-nfdxn\" (UID: \"700f73a1-f56d-4d4e-9822-438472980124\") " pod="openstack/placement-db-sync-nfdxn" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.222596 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4ck2\" (UniqueName: \"kubernetes.io/projected/700f73a1-f56d-4d4e-9822-438472980124-kube-api-access-s4ck2\") pod \"placement-db-sync-nfdxn\" (UID: \"700f73a1-f56d-4d4e-9822-438472980124\") " pod="openstack/placement-db-sync-nfdxn" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.222637 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/01547c5a-5ca7-435b-ba75-52e075800cf3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"01547c5a-5ca7-435b-ba75-52e075800cf3\") " pod="openstack/ceilometer-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.222671 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xcff\" (UniqueName: \"kubernetes.io/projected/01547c5a-5ca7-435b-ba75-52e075800cf3-kube-api-access-4xcff\") pod \"ceilometer-0\" (UID: \"01547c5a-5ca7-435b-ba75-52e075800cf3\") " pod="openstack/ceilometer-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.222701 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/700f73a1-f56d-4d4e-9822-438472980124-combined-ca-bundle\") pod \"placement-db-sync-nfdxn\" (UID: \"700f73a1-f56d-4d4e-9822-438472980124\") " pod="openstack/placement-db-sync-nfdxn" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.222771 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/700f73a1-f56d-4d4e-9822-438472980124-config-data\") pod \"placement-db-sync-nfdxn\" (UID: \"700f73a1-f56d-4d4e-9822-438472980124\") " pod="openstack/placement-db-sync-nfdxn" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.231115 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/700f73a1-f56d-4d4e-9822-438472980124-logs\") pod \"placement-db-sync-nfdxn\" (UID: \"700f73a1-f56d-4d4e-9822-438472980124\") " pod="openstack/placement-db-sync-nfdxn" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.231251 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01547c5a-5ca7-435b-ba75-52e075800cf3-scripts\") pod \"ceilometer-0\" (UID: \"01547c5a-5ca7-435b-ba75-52e075800cf3\") " pod="openstack/ceilometer-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.245311 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.245722 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6cf59496f7-6ntrd" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.265291 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.265397 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.269892 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.270100 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.270213 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-d6l2j" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.270345 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.293862 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-tr2q9" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.316418 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-48hsq" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.334335 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/01547c5a-5ca7-435b-ba75-52e075800cf3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"01547c5a-5ca7-435b-ba75-52e075800cf3\") " pod="openstack/ceilometer-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.338423 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xcff\" (UniqueName: \"kubernetes.io/projected/01547c5a-5ca7-435b-ba75-52e075800cf3-kube-api-access-4xcff\") pod \"ceilometer-0\" (UID: \"01547c5a-5ca7-435b-ba75-52e075800cf3\") " pod="openstack/ceilometer-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.338492 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/700f73a1-f56d-4d4e-9822-438472980124-combined-ca-bundle\") pod \"placement-db-sync-nfdxn\" (UID: \"700f73a1-f56d-4d4e-9822-438472980124\") " pod="openstack/placement-db-sync-nfdxn" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.338601 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/700f73a1-f56d-4d4e-9822-438472980124-config-data\") pod \"placement-db-sync-nfdxn\" (UID: \"700f73a1-f56d-4d4e-9822-438472980124\") " pod="openstack/placement-db-sync-nfdxn" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.338622 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mp9k\" (UniqueName: \"kubernetes.io/projected/04855a3a-d072-45f0-95d1-fb2f93264a52-kube-api-access-8mp9k\") pod \"dnsmasq-dns-56df8fb6b7-k2p8k\" (UID: \"04855a3a-d072-45f0-95d1-fb2f93264a52\") " pod="openstack/dnsmasq-dns-56df8fb6b7-k2p8k" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.338646 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/700f73a1-f56d-4d4e-9822-438472980124-logs\") pod \"placement-db-sync-nfdxn\" (UID: \"700f73a1-f56d-4d4e-9822-438472980124\") " pod="openstack/placement-db-sync-nfdxn" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.338682 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/04855a3a-d072-45f0-95d1-fb2f93264a52-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-k2p8k\" (UID: \"04855a3a-d072-45f0-95d1-fb2f93264a52\") " pod="openstack/dnsmasq-dns-56df8fb6b7-k2p8k" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.338726 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01547c5a-5ca7-435b-ba75-52e075800cf3-scripts\") pod \"ceilometer-0\" (UID: \"01547c5a-5ca7-435b-ba75-52e075800cf3\") " pod="openstack/ceilometer-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.338747 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01547c5a-5ca7-435b-ba75-52e075800cf3-log-httpd\") pod \"ceilometer-0\" (UID: \"01547c5a-5ca7-435b-ba75-52e075800cf3\") " pod="openstack/ceilometer-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.338770 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/04855a3a-d072-45f0-95d1-fb2f93264a52-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-k2p8k\" (UID: \"04855a3a-d072-45f0-95d1-fb2f93264a52\") " pod="openstack/dnsmasq-dns-56df8fb6b7-k2p8k" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.338819 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01547c5a-5ca7-435b-ba75-52e075800cf3-config-data\") pod \"ceilometer-0\" (UID: \"01547c5a-5ca7-435b-ba75-52e075800cf3\") " pod="openstack/ceilometer-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.338844 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/04855a3a-d072-45f0-95d1-fb2f93264a52-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-k2p8k\" (UID: \"04855a3a-d072-45f0-95d1-fb2f93264a52\") " pod="openstack/dnsmasq-dns-56df8fb6b7-k2p8k" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.338871 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01547c5a-5ca7-435b-ba75-52e075800cf3-run-httpd\") pod \"ceilometer-0\" (UID: \"01547c5a-5ca7-435b-ba75-52e075800cf3\") " pod="openstack/ceilometer-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.338888 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04855a3a-d072-45f0-95d1-fb2f93264a52-config\") pod \"dnsmasq-dns-56df8fb6b7-k2p8k\" (UID: \"04855a3a-d072-45f0-95d1-fb2f93264a52\") " pod="openstack/dnsmasq-dns-56df8fb6b7-k2p8k" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.338911 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01547c5a-5ca7-435b-ba75-52e075800cf3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"01547c5a-5ca7-435b-ba75-52e075800cf3\") " pod="openstack/ceilometer-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.338938 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4ck2\" (UniqueName: \"kubernetes.io/projected/700f73a1-f56d-4d4e-9822-438472980124-kube-api-access-s4ck2\") pod \"placement-db-sync-nfdxn\" (UID: \"700f73a1-f56d-4d4e-9822-438472980124\") " pod="openstack/placement-db-sync-nfdxn" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.338955 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/700f73a1-f56d-4d4e-9822-438472980124-scripts\") pod \"placement-db-sync-nfdxn\" (UID: \"700f73a1-f56d-4d4e-9822-438472980124\") " pod="openstack/placement-db-sync-nfdxn" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.338995 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/04855a3a-d072-45f0-95d1-fb2f93264a52-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-k2p8k\" (UID: \"04855a3a-d072-45f0-95d1-fb2f93264a52\") " pod="openstack/dnsmasq-dns-56df8fb6b7-k2p8k" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.342470 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/700f73a1-f56d-4d4e-9822-438472980124-logs\") pod \"placement-db-sync-nfdxn\" (UID: \"700f73a1-f56d-4d4e-9822-438472980124\") " pod="openstack/placement-db-sync-nfdxn" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.342939 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01547c5a-5ca7-435b-ba75-52e075800cf3-run-httpd\") pod \"ceilometer-0\" (UID: \"01547c5a-5ca7-435b-ba75-52e075800cf3\") " pod="openstack/ceilometer-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.342958 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01547c5a-5ca7-435b-ba75-52e075800cf3-log-httpd\") pod \"ceilometer-0\" (UID: \"01547c5a-5ca7-435b-ba75-52e075800cf3\") " pod="openstack/ceilometer-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.344254 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/700f73a1-f56d-4d4e-9822-438472980124-scripts\") pod \"placement-db-sync-nfdxn\" (UID: \"700f73a1-f56d-4d4e-9822-438472980124\") " pod="openstack/placement-db-sync-nfdxn" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.348464 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01547c5a-5ca7-435b-ba75-52e075800cf3-config-data\") pod \"ceilometer-0\" (UID: \"01547c5a-5ca7-435b-ba75-52e075800cf3\") " pod="openstack/ceilometer-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.348908 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/01547c5a-5ca7-435b-ba75-52e075800cf3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"01547c5a-5ca7-435b-ba75-52e075800cf3\") " pod="openstack/ceilometer-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.351732 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.361900 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01547c5a-5ca7-435b-ba75-52e075800cf3-scripts\") pod \"ceilometer-0\" (UID: \"01547c5a-5ca7-435b-ba75-52e075800cf3\") " pod="openstack/ceilometer-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.362489 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/700f73a1-f56d-4d4e-9822-438472980124-combined-ca-bundle\") pod \"placement-db-sync-nfdxn\" (UID: \"700f73a1-f56d-4d4e-9822-438472980124\") " pod="openstack/placement-db-sync-nfdxn" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.363927 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/700f73a1-f56d-4d4e-9822-438472980124-config-data\") pod \"placement-db-sync-nfdxn\" (UID: \"700f73a1-f56d-4d4e-9822-438472980124\") " pod="openstack/placement-db-sync-nfdxn" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.367371 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.371270 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01547c5a-5ca7-435b-ba75-52e075800cf3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"01547c5a-5ca7-435b-ba75-52e075800cf3\") " pod="openstack/ceilometer-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.378897 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.379248 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.399020 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4ck2\" (UniqueName: \"kubernetes.io/projected/700f73a1-f56d-4d4e-9822-438472980124-kube-api-access-s4ck2\") pod \"placement-db-sync-nfdxn\" (UID: \"700f73a1-f56d-4d4e-9822-438472980124\") " pod="openstack/placement-db-sync-nfdxn" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.404624 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xcff\" (UniqueName: \"kubernetes.io/projected/01547c5a-5ca7-435b-ba75-52e075800cf3-kube-api-access-4xcff\") pod \"ceilometer-0\" (UID: \"01547c5a-5ca7-435b-ba75-52e075800cf3\") " pod="openstack/ceilometer-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.413116 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.446433 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9911635b-d76a-43da-910d-2c5c82bf2900-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9911635b-d76a-43da-910d-2c5c82bf2900\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.446494 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/04855a3a-d072-45f0-95d1-fb2f93264a52-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-k2p8k\" (UID: \"04855a3a-d072-45f0-95d1-fb2f93264a52\") " pod="openstack/dnsmasq-dns-56df8fb6b7-k2p8k" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.446537 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9911635b-d76a-43da-910d-2c5c82bf2900-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9911635b-d76a-43da-910d-2c5c82bf2900\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.451975 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/04855a3a-d072-45f0-95d1-fb2f93264a52-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-k2p8k\" (UID: \"04855a3a-d072-45f0-95d1-fb2f93264a52\") " pod="openstack/dnsmasq-dns-56df8fb6b7-k2p8k" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.452025 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/da645eab-8b87-4ef3-b10a-2401382e5f19-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"da645eab-8b87-4ef3-b10a-2401382e5f19\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.452053 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9911635b-d76a-43da-910d-2c5c82bf2900-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9911635b-d76a-43da-910d-2c5c82bf2900\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.452070 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da645eab-8b87-4ef3-b10a-2401382e5f19-config-data\") pod \"glance-default-internal-api-0\" (UID: \"da645eab-8b87-4ef3-b10a-2401382e5f19\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.452101 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/04855a3a-d072-45f0-95d1-fb2f93264a52-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-k2p8k\" (UID: \"04855a3a-d072-45f0-95d1-fb2f93264a52\") " pod="openstack/dnsmasq-dns-56df8fb6b7-k2p8k" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.452135 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04855a3a-d072-45f0-95d1-fb2f93264a52-config\") pod \"dnsmasq-dns-56df8fb6b7-k2p8k\" (UID: \"04855a3a-d072-45f0-95d1-fb2f93264a52\") " pod="openstack/dnsmasq-dns-56df8fb6b7-k2p8k" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.452179 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gcxk\" (UniqueName: \"kubernetes.io/projected/da645eab-8b87-4ef3-b10a-2401382e5f19-kube-api-access-2gcxk\") pod \"glance-default-internal-api-0\" (UID: \"da645eab-8b87-4ef3-b10a-2401382e5f19\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.452220 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"9911635b-d76a-43da-910d-2c5c82bf2900\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.452260 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/04855a3a-d072-45f0-95d1-fb2f93264a52-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-k2p8k\" (UID: \"04855a3a-d072-45f0-95d1-fb2f93264a52\") " pod="openstack/dnsmasq-dns-56df8fb6b7-k2p8k" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.452280 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/da645eab-8b87-4ef3-b10a-2401382e5f19-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"da645eab-8b87-4ef3-b10a-2401382e5f19\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.452297 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snqv6\" (UniqueName: \"kubernetes.io/projected/9911635b-d76a-43da-910d-2c5c82bf2900-kube-api-access-snqv6\") pod \"glance-default-external-api-0\" (UID: \"9911635b-d76a-43da-910d-2c5c82bf2900\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.452328 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da645eab-8b87-4ef3-b10a-2401382e5f19-scripts\") pod \"glance-default-internal-api-0\" (UID: \"da645eab-8b87-4ef3-b10a-2401382e5f19\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.452348 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9911635b-d76a-43da-910d-2c5c82bf2900-logs\") pod \"glance-default-external-api-0\" (UID: \"9911635b-d76a-43da-910d-2c5c82bf2900\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.452367 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da645eab-8b87-4ef3-b10a-2401382e5f19-logs\") pod \"glance-default-internal-api-0\" (UID: \"da645eab-8b87-4ef3-b10a-2401382e5f19\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.452422 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"da645eab-8b87-4ef3-b10a-2401382e5f19\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.452443 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9911635b-d76a-43da-910d-2c5c82bf2900-scripts\") pod \"glance-default-external-api-0\" (UID: \"9911635b-d76a-43da-910d-2c5c82bf2900\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.452527 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9911635b-d76a-43da-910d-2c5c82bf2900-config-data\") pod \"glance-default-external-api-0\" (UID: \"9911635b-d76a-43da-910d-2c5c82bf2900\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.452558 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mp9k\" (UniqueName: \"kubernetes.io/projected/04855a3a-d072-45f0-95d1-fb2f93264a52-kube-api-access-8mp9k\") pod \"dnsmasq-dns-56df8fb6b7-k2p8k\" (UID: \"04855a3a-d072-45f0-95d1-fb2f93264a52\") " pod="openstack/dnsmasq-dns-56df8fb6b7-k2p8k" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.452572 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da645eab-8b87-4ef3-b10a-2401382e5f19-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"da645eab-8b87-4ef3-b10a-2401382e5f19\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.457059 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/04855a3a-d072-45f0-95d1-fb2f93264a52-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-k2p8k\" (UID: \"04855a3a-d072-45f0-95d1-fb2f93264a52\") " pod="openstack/dnsmasq-dns-56df8fb6b7-k2p8k" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.463703 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/04855a3a-d072-45f0-95d1-fb2f93264a52-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-k2p8k\" (UID: \"04855a3a-d072-45f0-95d1-fb2f93264a52\") " pod="openstack/dnsmasq-dns-56df8fb6b7-k2p8k" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.464390 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04855a3a-d072-45f0-95d1-fb2f93264a52-config\") pod \"dnsmasq-dns-56df8fb6b7-k2p8k\" (UID: \"04855a3a-d072-45f0-95d1-fb2f93264a52\") " pod="openstack/dnsmasq-dns-56df8fb6b7-k2p8k" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.471441 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/04855a3a-d072-45f0-95d1-fb2f93264a52-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-k2p8k\" (UID: \"04855a3a-d072-45f0-95d1-fb2f93264a52\") " pod="openstack/dnsmasq-dns-56df8fb6b7-k2p8k" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.473928 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/04855a3a-d072-45f0-95d1-fb2f93264a52-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-k2p8k\" (UID: \"04855a3a-d072-45f0-95d1-fb2f93264a52\") " pod="openstack/dnsmasq-dns-56df8fb6b7-k2p8k" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.485602 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mp9k\" (UniqueName: \"kubernetes.io/projected/04855a3a-d072-45f0-95d1-fb2f93264a52-kube-api-access-8mp9k\") pod \"dnsmasq-dns-56df8fb6b7-k2p8k\" (UID: \"04855a3a-d072-45f0-95d1-fb2f93264a52\") " pod="openstack/dnsmasq-dns-56df8fb6b7-k2p8k" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.487760 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nfdxn" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.525233 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-k2p8k" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.559610 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9911635b-d76a-43da-910d-2c5c82bf2900-logs\") pod \"glance-default-external-api-0\" (UID: \"9911635b-d76a-43da-910d-2c5c82bf2900\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.559658 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da645eab-8b87-4ef3-b10a-2401382e5f19-logs\") pod \"glance-default-internal-api-0\" (UID: \"da645eab-8b87-4ef3-b10a-2401382e5f19\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.559689 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"da645eab-8b87-4ef3-b10a-2401382e5f19\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.559705 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9911635b-d76a-43da-910d-2c5c82bf2900-scripts\") pod \"glance-default-external-api-0\" (UID: \"9911635b-d76a-43da-910d-2c5c82bf2900\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.559748 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9911635b-d76a-43da-910d-2c5c82bf2900-config-data\") pod \"glance-default-external-api-0\" (UID: \"9911635b-d76a-43da-910d-2c5c82bf2900\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.559769 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da645eab-8b87-4ef3-b10a-2401382e5f19-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"da645eab-8b87-4ef3-b10a-2401382e5f19\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.559804 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9911635b-d76a-43da-910d-2c5c82bf2900-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9911635b-d76a-43da-910d-2c5c82bf2900\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.559828 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9911635b-d76a-43da-910d-2c5c82bf2900-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9911635b-d76a-43da-910d-2c5c82bf2900\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.559882 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/da645eab-8b87-4ef3-b10a-2401382e5f19-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"da645eab-8b87-4ef3-b10a-2401382e5f19\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.559898 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9911635b-d76a-43da-910d-2c5c82bf2900-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9911635b-d76a-43da-910d-2c5c82bf2900\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.559911 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da645eab-8b87-4ef3-b10a-2401382e5f19-config-data\") pod \"glance-default-internal-api-0\" (UID: \"da645eab-8b87-4ef3-b10a-2401382e5f19\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.559945 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gcxk\" (UniqueName: \"kubernetes.io/projected/da645eab-8b87-4ef3-b10a-2401382e5f19-kube-api-access-2gcxk\") pod \"glance-default-internal-api-0\" (UID: \"da645eab-8b87-4ef3-b10a-2401382e5f19\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.561395 4955 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"da645eab-8b87-4ef3-b10a-2401382e5f19\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.563691 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"9911635b-d76a-43da-910d-2c5c82bf2900\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.563827 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/da645eab-8b87-4ef3-b10a-2401382e5f19-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"da645eab-8b87-4ef3-b10a-2401382e5f19\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.563845 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snqv6\" (UniqueName: \"kubernetes.io/projected/9911635b-d76a-43da-910d-2c5c82bf2900-kube-api-access-snqv6\") pod \"glance-default-external-api-0\" (UID: \"9911635b-d76a-43da-910d-2c5c82bf2900\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.563875 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da645eab-8b87-4ef3-b10a-2401382e5f19-scripts\") pod \"glance-default-internal-api-0\" (UID: \"da645eab-8b87-4ef3-b10a-2401382e5f19\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.567234 4955 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"9911635b-d76a-43da-910d-2c5c82bf2900\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-external-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.568682 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/da645eab-8b87-4ef3-b10a-2401382e5f19-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"da645eab-8b87-4ef3-b10a-2401382e5f19\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.569006 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9911635b-d76a-43da-910d-2c5c82bf2900-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9911635b-d76a-43da-910d-2c5c82bf2900\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.569238 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9911635b-d76a-43da-910d-2c5c82bf2900-logs\") pod \"glance-default-external-api-0\" (UID: \"9911635b-d76a-43da-910d-2c5c82bf2900\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.569449 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da645eab-8b87-4ef3-b10a-2401382e5f19-logs\") pod \"glance-default-internal-api-0\" (UID: \"da645eab-8b87-4ef3-b10a-2401382e5f19\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.572193 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da645eab-8b87-4ef3-b10a-2401382e5f19-scripts\") pod \"glance-default-internal-api-0\" (UID: \"da645eab-8b87-4ef3-b10a-2401382e5f19\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.573777 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/da645eab-8b87-4ef3-b10a-2401382e5f19-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"da645eab-8b87-4ef3-b10a-2401382e5f19\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.580055 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da645eab-8b87-4ef3-b10a-2401382e5f19-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"da645eab-8b87-4ef3-b10a-2401382e5f19\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.580254 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da645eab-8b87-4ef3-b10a-2401382e5f19-config-data\") pod \"glance-default-internal-api-0\" (UID: \"da645eab-8b87-4ef3-b10a-2401382e5f19\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.594816 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9911635b-d76a-43da-910d-2c5c82bf2900-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9911635b-d76a-43da-910d-2c5c82bf2900\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.595096 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9911635b-d76a-43da-910d-2c5c82bf2900-scripts\") pod \"glance-default-external-api-0\" (UID: \"9911635b-d76a-43da-910d-2c5c82bf2900\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.596263 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gcxk\" (UniqueName: \"kubernetes.io/projected/da645eab-8b87-4ef3-b10a-2401382e5f19-kube-api-access-2gcxk\") pod \"glance-default-internal-api-0\" (UID: \"da645eab-8b87-4ef3-b10a-2401382e5f19\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.600413 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snqv6\" (UniqueName: \"kubernetes.io/projected/9911635b-d76a-43da-910d-2c5c82bf2900-kube-api-access-snqv6\") pod \"glance-default-external-api-0\" (UID: \"9911635b-d76a-43da-910d-2c5c82bf2900\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.609677 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9911635b-d76a-43da-910d-2c5c82bf2900-config-data\") pod \"glance-default-external-api-0\" (UID: \"9911635b-d76a-43da-910d-2c5c82bf2900\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.614587 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9911635b-d76a-43da-910d-2c5c82bf2900-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9911635b-d76a-43da-910d-2c5c82bf2900\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.665805 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"da645eab-8b87-4ef3-b10a-2401382e5f19\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.684106 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.686910 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"9911635b-d76a-43da-910d-2c5c82bf2900\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.686938 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-4pqz5"] Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.747177 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.845489 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-wr4x8"] Feb 17 13:22:15 crc kubenswrapper[4955]: W0217 13:22:15.855888 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8bcff51_1ddc_4dfe_b45e_33aa814c5ef2.slice/crio-7b04bb2991822ab6b30890c24ae21f5ba300b29b93fe9874af0f493aeaeb4b0a WatchSource:0}: Error finding container 7b04bb2991822ab6b30890c24ae21f5ba300b29b93fe9874af0f493aeaeb4b0a: Status 404 returned error can't find the container with id 7b04bb2991822ab6b30890c24ae21f5ba300b29b93fe9874af0f493aeaeb4b0a Feb 17 13:22:15 crc kubenswrapper[4955]: I0217 13:22:15.894004 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 17 13:22:16 crc kubenswrapper[4955]: I0217 13:22:16.040664 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-wr4x8" event={"ID":"e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2","Type":"ContainerStarted","Data":"7b04bb2991822ab6b30890c24ae21f5ba300b29b93fe9874af0f493aeaeb4b0a"} Feb 17 13:22:16 crc kubenswrapper[4955]: I0217 13:22:16.043533 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4pqz5" event={"ID":"fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e","Type":"ContainerStarted","Data":"fc805665a5a34766d800b2d06f2a7ef6aa2e74c1684cae4a273c47ffcc73940a"} Feb 17 13:22:16 crc kubenswrapper[4955]: I0217 13:22:16.267012 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-w6s4s"] Feb 17 13:22:16 crc kubenswrapper[4955]: I0217 13:22:16.287824 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-tr2q9"] Feb 17 13:22:16 crc kubenswrapper[4955]: I0217 13:22:16.312225 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-76cdc74c45-lss64"] Feb 17 13:22:16 crc kubenswrapper[4955]: W0217 13:22:16.324998 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd1b9c8f8_6829_4eaf_b5bf_48121eaaf818.slice/crio-53baeb1e2441390fb1afebdf726ab4dd0cf54f830b271eb797ad7942b592192d WatchSource:0}: Error finding container 53baeb1e2441390fb1afebdf726ab4dd0cf54f830b271eb797ad7942b592192d: Status 404 returned error can't find the container with id 53baeb1e2441390fb1afebdf726ab4dd0cf54f830b271eb797ad7942b592192d Feb 17 13:22:16 crc kubenswrapper[4955]: I0217 13:22:16.344299 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6cf59496f7-6ntrd"] Feb 17 13:22:16 crc kubenswrapper[4955]: W0217 13:22:16.398586 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01547c5a_5ca7_435b_ba75_52e075800cf3.slice/crio-c1577238904bde655336171e1f902b973ec8981862e0b663c6e6dacfeb580d1d WatchSource:0}: Error finding container c1577238904bde655336171e1f902b973ec8981862e0b663c6e6dacfeb580d1d: Status 404 returned error can't find the container with id c1577238904bde655336171e1f902b973ec8981862e0b663c6e6dacfeb580d1d Feb 17 13:22:16 crc kubenswrapper[4955]: W0217 13:22:16.400206 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod700f73a1_f56d_4d4e_9822_438472980124.slice/crio-48e5f718063f753992d0e636d3dd74ad4c481301f54925379298b3b86e7630c7 WatchSource:0}: Error finding container 48e5f718063f753992d0e636d3dd74ad4c481301f54925379298b3b86e7630c7: Status 404 returned error can't find the container with id 48e5f718063f753992d0e636d3dd74ad4c481301f54925379298b3b86e7630c7 Feb 17 13:22:16 crc kubenswrapper[4955]: I0217 13:22:16.401069 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:22:16 crc kubenswrapper[4955]: I0217 13:22:16.428729 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-nfdxn"] Feb 17 13:22:16 crc kubenswrapper[4955]: I0217 13:22:16.428800 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-48hsq"] Feb 17 13:22:16 crc kubenswrapper[4955]: I0217 13:22:16.440250 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-k2p8k"] Feb 17 13:22:16 crc kubenswrapper[4955]: I0217 13:22:16.675470 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.095070 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"da645eab-8b87-4ef3-b10a-2401382e5f19","Type":"ContainerStarted","Data":"2d3fac841ce18e4cd8a3365a3ead9d39042057d73f9a4cf4702818fcb2e60376"} Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.111567 4955 generic.go:334] "Generic (PLEG): container finished" podID="04855a3a-d072-45f0-95d1-fb2f93264a52" containerID="1ccc5c07865713311e4a6ff4900e7ac30db4dd26f40d38d4b3cfffbbe15bb459" exitCode=0 Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.111838 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-k2p8k" event={"ID":"04855a3a-d072-45f0-95d1-fb2f93264a52","Type":"ContainerDied","Data":"1ccc5c07865713311e4a6ff4900e7ac30db4dd26f40d38d4b3cfffbbe15bb459"} Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.111960 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-k2p8k" event={"ID":"04855a3a-d072-45f0-95d1-fb2f93264a52","Type":"ContainerStarted","Data":"4626844e4ee4838aff936e347470c7f6aa19dd504201e84862aae8029154230c"} Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.139976 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-w6s4s" event={"ID":"e1c2d0c5-7647-447f-a1a6-b37d8c09310c","Type":"ContainerStarted","Data":"92880165e23e247c24ba8d154bdc1ba73ca37f070bbe970c913f1f598aeb8b97"} Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.206249 4955 generic.go:334] "Generic (PLEG): container finished" podID="e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2" containerID="6d7fef0f80b497c4e99cb7b6ddb559ffc0c3a43dafa57a226539ec7d6491a5c4" exitCode=0 Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.207702 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-wr4x8" event={"ID":"e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2","Type":"ContainerDied","Data":"6d7fef0f80b497c4e99cb7b6ddb559ffc0c3a43dafa57a226539ec7d6491a5c4"} Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.246205 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-76cdc74c45-lss64" event={"ID":"dd349c9d-f6b6-48b5-a22f-e2c1d81f6711","Type":"ContainerStarted","Data":"25358c9f4c2ef72c3ffa95ec983ed5222c638d8f3bc40656d219b1179d9404cf"} Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.281999 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-48hsq" event={"ID":"be85ffb8-5128-4903-9f7d-cdd462a44b95","Type":"ContainerStarted","Data":"e80f7012019a470099c47cd6309d9795aca80788cbfc35c46df2b84756b7c03c"} Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.282252 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-48hsq" event={"ID":"be85ffb8-5128-4903-9f7d-cdd462a44b95","Type":"ContainerStarted","Data":"e3f3541a8fde56e369a10bd7c665bd910b9c221d3590924d2410e30e09fd93e5"} Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.286312 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01547c5a-5ca7-435b-ba75-52e075800cf3","Type":"ContainerStarted","Data":"c1577238904bde655336171e1f902b973ec8981862e0b663c6e6dacfeb580d1d"} Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.300086 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nfdxn" event={"ID":"700f73a1-f56d-4d4e-9822-438472980124","Type":"ContainerStarted","Data":"48e5f718063f753992d0e636d3dd74ad4c481301f54925379298b3b86e7630c7"} Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.305682 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4pqz5" event={"ID":"fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e","Type":"ContainerStarted","Data":"902464c41c5c22295b964e133a26a55d00752d1e0938fe4d135fcb53ce02936a"} Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.322759 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-48hsq" podStartSLOduration=3.322735637 podStartE2EDuration="3.322735637s" podCreationTimestamp="2026-02-17 13:22:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:22:17.318448895 +0000 UTC m=+1075.841178438" watchObservedRunningTime="2026-02-17 13:22:17.322735637 +0000 UTC m=+1075.845465180" Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.347934 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6cf59496f7-6ntrd" event={"ID":"d1b9c8f8-6829-4eaf-b5bf-48121eaaf818","Type":"ContainerStarted","Data":"53baeb1e2441390fb1afebdf726ab4dd0cf54f830b271eb797ad7942b592192d"} Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.374363 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-tr2q9" event={"ID":"1d695519-3143-4f53-ab46-5823bc04dc0d","Type":"ContainerStarted","Data":"a22683ecd358c1850e1abe1fd3c5d7a3a150311bc260e6a97b49ad753395900b"} Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.418880 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.443903 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6cf59496f7-6ntrd"] Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.453509 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-4pqz5" podStartSLOduration=3.4534851460000002 podStartE2EDuration="3.453485146s" podCreationTimestamp="2026-02-17 13:22:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:22:17.397304244 +0000 UTC m=+1075.920033787" watchObservedRunningTime="2026-02-17 13:22:17.453485146 +0000 UTC m=+1075.976214679" Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.513631 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5bbc974f4f-sf5xt"] Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.521829 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5bbc974f4f-sf5xt" Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.530222 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.535935 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7eea91a6-8688-4c16-a45e-5ec47ba5fba6-config-data\") pod \"horizon-5bbc974f4f-sf5xt\" (UID: \"7eea91a6-8688-4c16-a45e-5ec47ba5fba6\") " pod="openstack/horizon-5bbc974f4f-sf5xt" Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.536003 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7eea91a6-8688-4c16-a45e-5ec47ba5fba6-logs\") pod \"horizon-5bbc974f4f-sf5xt\" (UID: \"7eea91a6-8688-4c16-a45e-5ec47ba5fba6\") " pod="openstack/horizon-5bbc974f4f-sf5xt" Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.536058 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7eea91a6-8688-4c16-a45e-5ec47ba5fba6-horizon-secret-key\") pod \"horizon-5bbc974f4f-sf5xt\" (UID: \"7eea91a6-8688-4c16-a45e-5ec47ba5fba6\") " pod="openstack/horizon-5bbc974f4f-sf5xt" Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.536110 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7eea91a6-8688-4c16-a45e-5ec47ba5fba6-scripts\") pod \"horizon-5bbc974f4f-sf5xt\" (UID: \"7eea91a6-8688-4c16-a45e-5ec47ba5fba6\") " pod="openstack/horizon-5bbc974f4f-sf5xt" Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.536136 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8zpw\" (UniqueName: \"kubernetes.io/projected/7eea91a6-8688-4c16-a45e-5ec47ba5fba6-kube-api-access-n8zpw\") pod \"horizon-5bbc974f4f-sf5xt\" (UID: \"7eea91a6-8688-4c16-a45e-5ec47ba5fba6\") " pod="openstack/horizon-5bbc974f4f-sf5xt" Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.539679 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.559154 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5bbc974f4f-sf5xt"] Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.578596 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.637585 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7eea91a6-8688-4c16-a45e-5ec47ba5fba6-config-data\") pod \"horizon-5bbc974f4f-sf5xt\" (UID: \"7eea91a6-8688-4c16-a45e-5ec47ba5fba6\") " pod="openstack/horizon-5bbc974f4f-sf5xt" Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.637644 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7eea91a6-8688-4c16-a45e-5ec47ba5fba6-logs\") pod \"horizon-5bbc974f4f-sf5xt\" (UID: \"7eea91a6-8688-4c16-a45e-5ec47ba5fba6\") " pod="openstack/horizon-5bbc974f4f-sf5xt" Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.637704 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7eea91a6-8688-4c16-a45e-5ec47ba5fba6-horizon-secret-key\") pod \"horizon-5bbc974f4f-sf5xt\" (UID: \"7eea91a6-8688-4c16-a45e-5ec47ba5fba6\") " pod="openstack/horizon-5bbc974f4f-sf5xt" Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.637756 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7eea91a6-8688-4c16-a45e-5ec47ba5fba6-scripts\") pod \"horizon-5bbc974f4f-sf5xt\" (UID: \"7eea91a6-8688-4c16-a45e-5ec47ba5fba6\") " pod="openstack/horizon-5bbc974f4f-sf5xt" Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.637800 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8zpw\" (UniqueName: \"kubernetes.io/projected/7eea91a6-8688-4c16-a45e-5ec47ba5fba6-kube-api-access-n8zpw\") pod \"horizon-5bbc974f4f-sf5xt\" (UID: \"7eea91a6-8688-4c16-a45e-5ec47ba5fba6\") " pod="openstack/horizon-5bbc974f4f-sf5xt" Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.640083 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7eea91a6-8688-4c16-a45e-5ec47ba5fba6-config-data\") pod \"horizon-5bbc974f4f-sf5xt\" (UID: \"7eea91a6-8688-4c16-a45e-5ec47ba5fba6\") " pod="openstack/horizon-5bbc974f4f-sf5xt" Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.640369 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7eea91a6-8688-4c16-a45e-5ec47ba5fba6-logs\") pod \"horizon-5bbc974f4f-sf5xt\" (UID: \"7eea91a6-8688-4c16-a45e-5ec47ba5fba6\") " pod="openstack/horizon-5bbc974f4f-sf5xt" Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.647852 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7eea91a6-8688-4c16-a45e-5ec47ba5fba6-horizon-secret-key\") pod \"horizon-5bbc974f4f-sf5xt\" (UID: \"7eea91a6-8688-4c16-a45e-5ec47ba5fba6\") " pod="openstack/horizon-5bbc974f4f-sf5xt" Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.648410 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7eea91a6-8688-4c16-a45e-5ec47ba5fba6-scripts\") pod \"horizon-5bbc974f4f-sf5xt\" (UID: \"7eea91a6-8688-4c16-a45e-5ec47ba5fba6\") " pod="openstack/horizon-5bbc974f4f-sf5xt" Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.661637 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8zpw\" (UniqueName: \"kubernetes.io/projected/7eea91a6-8688-4c16-a45e-5ec47ba5fba6-kube-api-access-n8zpw\") pod \"horizon-5bbc974f4f-sf5xt\" (UID: \"7eea91a6-8688-4c16-a45e-5ec47ba5fba6\") " pod="openstack/horizon-5bbc974f4f-sf5xt" Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.803511 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5bbc974f4f-sf5xt" Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.910675 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-wr4x8" Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.942240 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2-config\") pod \"e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2\" (UID: \"e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2\") " Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.942304 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2-dns-svc\") pod \"e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2\" (UID: \"e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2\") " Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.942331 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2-ovsdbserver-nb\") pod \"e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2\" (UID: \"e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2\") " Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.942396 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2-ovsdbserver-sb\") pod \"e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2\" (UID: \"e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2\") " Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.942728 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tx2q4\" (UniqueName: \"kubernetes.io/projected/e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2-kube-api-access-tx2q4\") pod \"e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2\" (UID: \"e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2\") " Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.942809 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2-dns-swift-storage-0\") pod \"e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2\" (UID: \"e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2\") " Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.963650 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2-kube-api-access-tx2q4" (OuterVolumeSpecName: "kube-api-access-tx2q4") pod "e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2" (UID: "e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2"). InnerVolumeSpecName "kube-api-access-tx2q4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.975867 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2" (UID: "e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.983250 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2" (UID: "e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.986615 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2" (UID: "e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:22:17 crc kubenswrapper[4955]: I0217 13:22:17.991610 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2-config" (OuterVolumeSpecName: "config") pod "e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2" (UID: "e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:22:18 crc kubenswrapper[4955]: I0217 13:22:18.009665 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2" (UID: "e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:22:18 crc kubenswrapper[4955]: I0217 13:22:18.050930 4955 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:18 crc kubenswrapper[4955]: I0217 13:22:18.050963 4955 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:18 crc kubenswrapper[4955]: I0217 13:22:18.050975 4955 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:18 crc kubenswrapper[4955]: I0217 13:22:18.050987 4955 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:18 crc kubenswrapper[4955]: I0217 13:22:18.050999 4955 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:18 crc kubenswrapper[4955]: I0217 13:22:18.051011 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tx2q4\" (UniqueName: \"kubernetes.io/projected/e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2-kube-api-access-tx2q4\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:18 crc kubenswrapper[4955]: I0217 13:22:18.392641 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"da645eab-8b87-4ef3-b10a-2401382e5f19","Type":"ContainerStarted","Data":"b8f981b0a2fe3a0b405a3a8aa4970b4b4317d072ab6d7f36bddc9e1dcea04865"} Feb 17 13:22:18 crc kubenswrapper[4955]: I0217 13:22:18.399585 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-k2p8k" event={"ID":"04855a3a-d072-45f0-95d1-fb2f93264a52","Type":"ContainerStarted","Data":"0627cccc799eb4174c02bd70233cd70409dd7bb8aca5c6794565243148b8f7b4"} Feb 17 13:22:18 crc kubenswrapper[4955]: I0217 13:22:18.400738 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-56df8fb6b7-k2p8k" Feb 17 13:22:18 crc kubenswrapper[4955]: I0217 13:22:18.404592 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9911635b-d76a-43da-910d-2c5c82bf2900","Type":"ContainerStarted","Data":"1845f86f5437df26461a6c1a2b6782d0cd00d809dd6d3c7a10ed6af39f93c933"} Feb 17 13:22:18 crc kubenswrapper[4955]: I0217 13:22:18.425988 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-56df8fb6b7-k2p8k" podStartSLOduration=3.425963266 podStartE2EDuration="3.425963266s" podCreationTimestamp="2026-02-17 13:22:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:22:18.420358026 +0000 UTC m=+1076.943087569" watchObservedRunningTime="2026-02-17 13:22:18.425963266 +0000 UTC m=+1076.948692809" Feb 17 13:22:18 crc kubenswrapper[4955]: I0217 13:22:18.445537 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-wr4x8" Feb 17 13:22:18 crc kubenswrapper[4955]: I0217 13:22:18.445610 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-wr4x8" event={"ID":"e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2","Type":"ContainerDied","Data":"7b04bb2991822ab6b30890c24ae21f5ba300b29b93fe9874af0f493aeaeb4b0a"} Feb 17 13:22:18 crc kubenswrapper[4955]: I0217 13:22:18.445646 4955 scope.go:117] "RemoveContainer" containerID="6d7fef0f80b497c4e99cb7b6ddb559ffc0c3a43dafa57a226539ec7d6491a5c4" Feb 17 13:22:18 crc kubenswrapper[4955]: I0217 13:22:18.505980 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-wr4x8"] Feb 17 13:22:18 crc kubenswrapper[4955]: I0217 13:22:18.519489 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-wr4x8"] Feb 17 13:22:18 crc kubenswrapper[4955]: I0217 13:22:18.564236 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5bbc974f4f-sf5xt"] Feb 17 13:22:18 crc kubenswrapper[4955]: W0217 13:22:18.631475 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7eea91a6_8688_4c16_a45e_5ec47ba5fba6.slice/crio-674cc704ff6179471a37780d640752c3ad36343953dca4555ec93b61a6d59b1a WatchSource:0}: Error finding container 674cc704ff6179471a37780d640752c3ad36343953dca4555ec93b61a6d59b1a: Status 404 returned error can't find the container with id 674cc704ff6179471a37780d640752c3ad36343953dca4555ec93b61a6d59b1a Feb 17 13:22:19 crc kubenswrapper[4955]: I0217 13:22:19.494598 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="da645eab-8b87-4ef3-b10a-2401382e5f19" containerName="glance-log" containerID="cri-o://b8f981b0a2fe3a0b405a3a8aa4970b4b4317d072ab6d7f36bddc9e1dcea04865" gracePeriod=30 Feb 17 13:22:19 crc kubenswrapper[4955]: I0217 13:22:19.495258 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="da645eab-8b87-4ef3-b10a-2401382e5f19" containerName="glance-httpd" containerID="cri-o://173d0eacfb715e15e0a3b896d811361689cd29b9f416c896d3e0fb0d1aa3c81b" gracePeriod=30 Feb 17 13:22:19 crc kubenswrapper[4955]: I0217 13:22:19.494818 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"da645eab-8b87-4ef3-b10a-2401382e5f19","Type":"ContainerStarted","Data":"173d0eacfb715e15e0a3b896d811361689cd29b9f416c896d3e0fb0d1aa3c81b"} Feb 17 13:22:19 crc kubenswrapper[4955]: I0217 13:22:19.521601 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5bbc974f4f-sf5xt" event={"ID":"7eea91a6-8688-4c16-a45e-5ec47ba5fba6","Type":"ContainerStarted","Data":"674cc704ff6179471a37780d640752c3ad36343953dca4555ec93b61a6d59b1a"} Feb 17 13:22:19 crc kubenswrapper[4955]: I0217 13:22:19.523002 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.522981867 podStartE2EDuration="4.522981867s" podCreationTimestamp="2026-02-17 13:22:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:22:19.521051112 +0000 UTC m=+1078.043780655" watchObservedRunningTime="2026-02-17 13:22:19.522981867 +0000 UTC m=+1078.045711410" Feb 17 13:22:19 crc kubenswrapper[4955]: I0217 13:22:19.530972 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9911635b-d76a-43da-910d-2c5c82bf2900","Type":"ContainerStarted","Data":"34b9043b43dcfcf4e9bd8fe5dd18d0625fa3489ecc3028391af8b117f5445333"} Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.202217 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.243086 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2" path="/var/lib/kubelet/pods/e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2/volumes" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.329652 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"da645eab-8b87-4ef3-b10a-2401382e5f19\" (UID: \"da645eab-8b87-4ef3-b10a-2401382e5f19\") " Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.329762 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da645eab-8b87-4ef3-b10a-2401382e5f19-scripts\") pod \"da645eab-8b87-4ef3-b10a-2401382e5f19\" (UID: \"da645eab-8b87-4ef3-b10a-2401382e5f19\") " Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.329811 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2gcxk\" (UniqueName: \"kubernetes.io/projected/da645eab-8b87-4ef3-b10a-2401382e5f19-kube-api-access-2gcxk\") pod \"da645eab-8b87-4ef3-b10a-2401382e5f19\" (UID: \"da645eab-8b87-4ef3-b10a-2401382e5f19\") " Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.329845 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da645eab-8b87-4ef3-b10a-2401382e5f19-logs\") pod \"da645eab-8b87-4ef3-b10a-2401382e5f19\" (UID: \"da645eab-8b87-4ef3-b10a-2401382e5f19\") " Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.329941 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da645eab-8b87-4ef3-b10a-2401382e5f19-config-data\") pod \"da645eab-8b87-4ef3-b10a-2401382e5f19\" (UID: \"da645eab-8b87-4ef3-b10a-2401382e5f19\") " Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.329978 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/da645eab-8b87-4ef3-b10a-2401382e5f19-httpd-run\") pod \"da645eab-8b87-4ef3-b10a-2401382e5f19\" (UID: \"da645eab-8b87-4ef3-b10a-2401382e5f19\") " Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.329993 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/da645eab-8b87-4ef3-b10a-2401382e5f19-internal-tls-certs\") pod \"da645eab-8b87-4ef3-b10a-2401382e5f19\" (UID: \"da645eab-8b87-4ef3-b10a-2401382e5f19\") " Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.330012 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da645eab-8b87-4ef3-b10a-2401382e5f19-combined-ca-bundle\") pod \"da645eab-8b87-4ef3-b10a-2401382e5f19\" (UID: \"da645eab-8b87-4ef3-b10a-2401382e5f19\") " Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.331422 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da645eab-8b87-4ef3-b10a-2401382e5f19-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "da645eab-8b87-4ef3-b10a-2401382e5f19" (UID: "da645eab-8b87-4ef3-b10a-2401382e5f19"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.335076 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da645eab-8b87-4ef3-b10a-2401382e5f19-logs" (OuterVolumeSpecName: "logs") pod "da645eab-8b87-4ef3-b10a-2401382e5f19" (UID: "da645eab-8b87-4ef3-b10a-2401382e5f19"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.336115 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da645eab-8b87-4ef3-b10a-2401382e5f19-kube-api-access-2gcxk" (OuterVolumeSpecName: "kube-api-access-2gcxk") pod "da645eab-8b87-4ef3-b10a-2401382e5f19" (UID: "da645eab-8b87-4ef3-b10a-2401382e5f19"). InnerVolumeSpecName "kube-api-access-2gcxk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.336350 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "da645eab-8b87-4ef3-b10a-2401382e5f19" (UID: "da645eab-8b87-4ef3-b10a-2401382e5f19"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.338643 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da645eab-8b87-4ef3-b10a-2401382e5f19-scripts" (OuterVolumeSpecName: "scripts") pod "da645eab-8b87-4ef3-b10a-2401382e5f19" (UID: "da645eab-8b87-4ef3-b10a-2401382e5f19"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.383087 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da645eab-8b87-4ef3-b10a-2401382e5f19-config-data" (OuterVolumeSpecName: "config-data") pod "da645eab-8b87-4ef3-b10a-2401382e5f19" (UID: "da645eab-8b87-4ef3-b10a-2401382e5f19"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.388010 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da645eab-8b87-4ef3-b10a-2401382e5f19-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "da645eab-8b87-4ef3-b10a-2401382e5f19" (UID: "da645eab-8b87-4ef3-b10a-2401382e5f19"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.396341 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da645eab-8b87-4ef3-b10a-2401382e5f19-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "da645eab-8b87-4ef3-b10a-2401382e5f19" (UID: "da645eab-8b87-4ef3-b10a-2401382e5f19"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.436418 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2gcxk\" (UniqueName: \"kubernetes.io/projected/da645eab-8b87-4ef3-b10a-2401382e5f19-kube-api-access-2gcxk\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.436459 4955 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da645eab-8b87-4ef3-b10a-2401382e5f19-logs\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.436474 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da645eab-8b87-4ef3-b10a-2401382e5f19-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.436486 4955 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/da645eab-8b87-4ef3-b10a-2401382e5f19-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.436497 4955 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/da645eab-8b87-4ef3-b10a-2401382e5f19-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.436507 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da645eab-8b87-4ef3-b10a-2401382e5f19-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.436545 4955 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.436558 4955 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da645eab-8b87-4ef3-b10a-2401382e5f19-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.456096 4955 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.538929 4955 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.552057 4955 generic.go:334] "Generic (PLEG): container finished" podID="da645eab-8b87-4ef3-b10a-2401382e5f19" containerID="173d0eacfb715e15e0a3b896d811361689cd29b9f416c896d3e0fb0d1aa3c81b" exitCode=143 Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.552149 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.552167 4955 generic.go:334] "Generic (PLEG): container finished" podID="da645eab-8b87-4ef3-b10a-2401382e5f19" containerID="b8f981b0a2fe3a0b405a3a8aa4970b4b4317d072ab6d7f36bddc9e1dcea04865" exitCode=143 Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.552126 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"da645eab-8b87-4ef3-b10a-2401382e5f19","Type":"ContainerDied","Data":"173d0eacfb715e15e0a3b896d811361689cd29b9f416c896d3e0fb0d1aa3c81b"} Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.552251 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"da645eab-8b87-4ef3-b10a-2401382e5f19","Type":"ContainerDied","Data":"b8f981b0a2fe3a0b405a3a8aa4970b4b4317d072ab6d7f36bddc9e1dcea04865"} Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.552268 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"da645eab-8b87-4ef3-b10a-2401382e5f19","Type":"ContainerDied","Data":"2d3fac841ce18e4cd8a3365a3ead9d39042057d73f9a4cf4702818fcb2e60376"} Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.552286 4955 scope.go:117] "RemoveContainer" containerID="173d0eacfb715e15e0a3b896d811361689cd29b9f416c896d3e0fb0d1aa3c81b" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.556304 4955 generic.go:334] "Generic (PLEG): container finished" podID="fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e" containerID="902464c41c5c22295b964e133a26a55d00752d1e0938fe4d135fcb53ce02936a" exitCode=0 Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.556369 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4pqz5" event={"ID":"fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e","Type":"ContainerDied","Data":"902464c41c5c22295b964e133a26a55d00752d1e0938fe4d135fcb53ce02936a"} Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.567907 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="9911635b-d76a-43da-910d-2c5c82bf2900" containerName="glance-log" containerID="cri-o://34b9043b43dcfcf4e9bd8fe5dd18d0625fa3489ecc3028391af8b117f5445333" gracePeriod=30 Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.568206 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9911635b-d76a-43da-910d-2c5c82bf2900","Type":"ContainerStarted","Data":"099f1b679ed3eccbe094cddcb02aacc1522fee8432993309a7fb48be275c2a98"} Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.568276 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="9911635b-d76a-43da-910d-2c5c82bf2900" containerName="glance-httpd" containerID="cri-o://099f1b679ed3eccbe094cddcb02aacc1522fee8432993309a7fb48be275c2a98" gracePeriod=30 Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.600427 4955 scope.go:117] "RemoveContainer" containerID="b8f981b0a2fe3a0b405a3a8aa4970b4b4317d072ab6d7f36bddc9e1dcea04865" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.627303 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.627279357 podStartE2EDuration="5.627279357s" podCreationTimestamp="2026-02-17 13:22:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:22:20.607992027 +0000 UTC m=+1079.130721570" watchObservedRunningTime="2026-02-17 13:22:20.627279357 +0000 UTC m=+1079.150008900" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.646928 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.678866 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.684332 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 17 13:22:20 crc kubenswrapper[4955]: E0217 13:22:20.684656 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da645eab-8b87-4ef3-b10a-2401382e5f19" containerName="glance-httpd" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.684667 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="da645eab-8b87-4ef3-b10a-2401382e5f19" containerName="glance-httpd" Feb 17 13:22:20 crc kubenswrapper[4955]: E0217 13:22:20.684682 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da645eab-8b87-4ef3-b10a-2401382e5f19" containerName="glance-log" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.684688 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="da645eab-8b87-4ef3-b10a-2401382e5f19" containerName="glance-log" Feb 17 13:22:20 crc kubenswrapper[4955]: E0217 13:22:20.684717 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2" containerName="init" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.684723 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2" containerName="init" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.684872 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="da645eab-8b87-4ef3-b10a-2401382e5f19" containerName="glance-log" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.684881 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8bcff51-1ddc-4dfe-b45e-33aa814c5ef2" containerName="init" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.684888 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="da645eab-8b87-4ef3-b10a-2401382e5f19" containerName="glance-httpd" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.685687 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.687984 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.692373 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.705509 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.745110 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9aac5b59-a6ed-43cc-888d-a583384f1ede-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"9aac5b59-a6ed-43cc-888d-a583384f1ede\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.745235 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"9aac5b59-a6ed-43cc-888d-a583384f1ede\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.745265 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9aac5b59-a6ed-43cc-888d-a583384f1ede-scripts\") pod \"glance-default-internal-api-0\" (UID: \"9aac5b59-a6ed-43cc-888d-a583384f1ede\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.745317 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9aac5b59-a6ed-43cc-888d-a583384f1ede-config-data\") pod \"glance-default-internal-api-0\" (UID: \"9aac5b59-a6ed-43cc-888d-a583384f1ede\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.745403 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9aac5b59-a6ed-43cc-888d-a583384f1ede-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"9aac5b59-a6ed-43cc-888d-a583384f1ede\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.745450 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9aac5b59-a6ed-43cc-888d-a583384f1ede-logs\") pod \"glance-default-internal-api-0\" (UID: \"9aac5b59-a6ed-43cc-888d-a583384f1ede\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.745512 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfs67\" (UniqueName: \"kubernetes.io/projected/9aac5b59-a6ed-43cc-888d-a583384f1ede-kube-api-access-mfs67\") pod \"glance-default-internal-api-0\" (UID: \"9aac5b59-a6ed-43cc-888d-a583384f1ede\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.745612 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9aac5b59-a6ed-43cc-888d-a583384f1ede-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"9aac5b59-a6ed-43cc-888d-a583384f1ede\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.749497 4955 scope.go:117] "RemoveContainer" containerID="173d0eacfb715e15e0a3b896d811361689cd29b9f416c896d3e0fb0d1aa3c81b" Feb 17 13:22:20 crc kubenswrapper[4955]: E0217 13:22:20.756033 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"173d0eacfb715e15e0a3b896d811361689cd29b9f416c896d3e0fb0d1aa3c81b\": container with ID starting with 173d0eacfb715e15e0a3b896d811361689cd29b9f416c896d3e0fb0d1aa3c81b not found: ID does not exist" containerID="173d0eacfb715e15e0a3b896d811361689cd29b9f416c896d3e0fb0d1aa3c81b" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.756093 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"173d0eacfb715e15e0a3b896d811361689cd29b9f416c896d3e0fb0d1aa3c81b"} err="failed to get container status \"173d0eacfb715e15e0a3b896d811361689cd29b9f416c896d3e0fb0d1aa3c81b\": rpc error: code = NotFound desc = could not find container \"173d0eacfb715e15e0a3b896d811361689cd29b9f416c896d3e0fb0d1aa3c81b\": container with ID starting with 173d0eacfb715e15e0a3b896d811361689cd29b9f416c896d3e0fb0d1aa3c81b not found: ID does not exist" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.756124 4955 scope.go:117] "RemoveContainer" containerID="b8f981b0a2fe3a0b405a3a8aa4970b4b4317d072ab6d7f36bddc9e1dcea04865" Feb 17 13:22:20 crc kubenswrapper[4955]: E0217 13:22:20.757156 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8f981b0a2fe3a0b405a3a8aa4970b4b4317d072ab6d7f36bddc9e1dcea04865\": container with ID starting with b8f981b0a2fe3a0b405a3a8aa4970b4b4317d072ab6d7f36bddc9e1dcea04865 not found: ID does not exist" containerID="b8f981b0a2fe3a0b405a3a8aa4970b4b4317d072ab6d7f36bddc9e1dcea04865" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.757199 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8f981b0a2fe3a0b405a3a8aa4970b4b4317d072ab6d7f36bddc9e1dcea04865"} err="failed to get container status \"b8f981b0a2fe3a0b405a3a8aa4970b4b4317d072ab6d7f36bddc9e1dcea04865\": rpc error: code = NotFound desc = could not find container \"b8f981b0a2fe3a0b405a3a8aa4970b4b4317d072ab6d7f36bddc9e1dcea04865\": container with ID starting with b8f981b0a2fe3a0b405a3a8aa4970b4b4317d072ab6d7f36bddc9e1dcea04865 not found: ID does not exist" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.757253 4955 scope.go:117] "RemoveContainer" containerID="173d0eacfb715e15e0a3b896d811361689cd29b9f416c896d3e0fb0d1aa3c81b" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.757729 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"173d0eacfb715e15e0a3b896d811361689cd29b9f416c896d3e0fb0d1aa3c81b"} err="failed to get container status \"173d0eacfb715e15e0a3b896d811361689cd29b9f416c896d3e0fb0d1aa3c81b\": rpc error: code = NotFound desc = could not find container \"173d0eacfb715e15e0a3b896d811361689cd29b9f416c896d3e0fb0d1aa3c81b\": container with ID starting with 173d0eacfb715e15e0a3b896d811361689cd29b9f416c896d3e0fb0d1aa3c81b not found: ID does not exist" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.757881 4955 scope.go:117] "RemoveContainer" containerID="b8f981b0a2fe3a0b405a3a8aa4970b4b4317d072ab6d7f36bddc9e1dcea04865" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.763944 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8f981b0a2fe3a0b405a3a8aa4970b4b4317d072ab6d7f36bddc9e1dcea04865"} err="failed to get container status \"b8f981b0a2fe3a0b405a3a8aa4970b4b4317d072ab6d7f36bddc9e1dcea04865\": rpc error: code = NotFound desc = could not find container \"b8f981b0a2fe3a0b405a3a8aa4970b4b4317d072ab6d7f36bddc9e1dcea04865\": container with ID starting with b8f981b0a2fe3a0b405a3a8aa4970b4b4317d072ab6d7f36bddc9e1dcea04865 not found: ID does not exist" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.847133 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9aac5b59-a6ed-43cc-888d-a583384f1ede-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"9aac5b59-a6ed-43cc-888d-a583384f1ede\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.847223 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9aac5b59-a6ed-43cc-888d-a583384f1ede-logs\") pod \"glance-default-internal-api-0\" (UID: \"9aac5b59-a6ed-43cc-888d-a583384f1ede\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.847250 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfs67\" (UniqueName: \"kubernetes.io/projected/9aac5b59-a6ed-43cc-888d-a583384f1ede-kube-api-access-mfs67\") pod \"glance-default-internal-api-0\" (UID: \"9aac5b59-a6ed-43cc-888d-a583384f1ede\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.847318 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9aac5b59-a6ed-43cc-888d-a583384f1ede-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"9aac5b59-a6ed-43cc-888d-a583384f1ede\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.847403 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9aac5b59-a6ed-43cc-888d-a583384f1ede-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"9aac5b59-a6ed-43cc-888d-a583384f1ede\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.847428 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"9aac5b59-a6ed-43cc-888d-a583384f1ede\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.847442 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9aac5b59-a6ed-43cc-888d-a583384f1ede-scripts\") pod \"glance-default-internal-api-0\" (UID: \"9aac5b59-a6ed-43cc-888d-a583384f1ede\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.847456 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9aac5b59-a6ed-43cc-888d-a583384f1ede-config-data\") pod \"glance-default-internal-api-0\" (UID: \"9aac5b59-a6ed-43cc-888d-a583384f1ede\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.848191 4955 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"9aac5b59-a6ed-43cc-888d-a583384f1ede\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.848830 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9aac5b59-a6ed-43cc-888d-a583384f1ede-logs\") pod \"glance-default-internal-api-0\" (UID: \"9aac5b59-a6ed-43cc-888d-a583384f1ede\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.849067 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9aac5b59-a6ed-43cc-888d-a583384f1ede-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"9aac5b59-a6ed-43cc-888d-a583384f1ede\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.859960 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9aac5b59-a6ed-43cc-888d-a583384f1ede-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"9aac5b59-a6ed-43cc-888d-a583384f1ede\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.859970 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9aac5b59-a6ed-43cc-888d-a583384f1ede-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"9aac5b59-a6ed-43cc-888d-a583384f1ede\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.860809 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9aac5b59-a6ed-43cc-888d-a583384f1ede-config-data\") pod \"glance-default-internal-api-0\" (UID: \"9aac5b59-a6ed-43cc-888d-a583384f1ede\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.865374 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfs67\" (UniqueName: \"kubernetes.io/projected/9aac5b59-a6ed-43cc-888d-a583384f1ede-kube-api-access-mfs67\") pod \"glance-default-internal-api-0\" (UID: \"9aac5b59-a6ed-43cc-888d-a583384f1ede\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.865413 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9aac5b59-a6ed-43cc-888d-a583384f1ede-scripts\") pod \"glance-default-internal-api-0\" (UID: \"9aac5b59-a6ed-43cc-888d-a583384f1ede\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:20 crc kubenswrapper[4955]: I0217 13:22:20.888961 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"9aac5b59-a6ed-43cc-888d-a583384f1ede\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.021917 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.315958 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.368373 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9911635b-d76a-43da-910d-2c5c82bf2900-httpd-run\") pod \"9911635b-d76a-43da-910d-2c5c82bf2900\" (UID: \"9911635b-d76a-43da-910d-2c5c82bf2900\") " Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.368414 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9911635b-d76a-43da-910d-2c5c82bf2900-config-data\") pod \"9911635b-d76a-43da-910d-2c5c82bf2900\" (UID: \"9911635b-d76a-43da-910d-2c5c82bf2900\") " Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.368484 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9911635b-d76a-43da-910d-2c5c82bf2900-public-tls-certs\") pod \"9911635b-d76a-43da-910d-2c5c82bf2900\" (UID: \"9911635b-d76a-43da-910d-2c5c82bf2900\") " Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.368531 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9911635b-d76a-43da-910d-2c5c82bf2900-logs\") pod \"9911635b-d76a-43da-910d-2c5c82bf2900\" (UID: \"9911635b-d76a-43da-910d-2c5c82bf2900\") " Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.368563 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"9911635b-d76a-43da-910d-2c5c82bf2900\" (UID: \"9911635b-d76a-43da-910d-2c5c82bf2900\") " Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.368585 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9911635b-d76a-43da-910d-2c5c82bf2900-scripts\") pod \"9911635b-d76a-43da-910d-2c5c82bf2900\" (UID: \"9911635b-d76a-43da-910d-2c5c82bf2900\") " Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.368644 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9911635b-d76a-43da-910d-2c5c82bf2900-combined-ca-bundle\") pod \"9911635b-d76a-43da-910d-2c5c82bf2900\" (UID: \"9911635b-d76a-43da-910d-2c5c82bf2900\") " Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.368686 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snqv6\" (UniqueName: \"kubernetes.io/projected/9911635b-d76a-43da-910d-2c5c82bf2900-kube-api-access-snqv6\") pod \"9911635b-d76a-43da-910d-2c5c82bf2900\" (UID: \"9911635b-d76a-43da-910d-2c5c82bf2900\") " Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.368899 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9911635b-d76a-43da-910d-2c5c82bf2900-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "9911635b-d76a-43da-910d-2c5c82bf2900" (UID: "9911635b-d76a-43da-910d-2c5c82bf2900"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.368923 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9911635b-d76a-43da-910d-2c5c82bf2900-logs" (OuterVolumeSpecName: "logs") pod "9911635b-d76a-43da-910d-2c5c82bf2900" (UID: "9911635b-d76a-43da-910d-2c5c82bf2900"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.369288 4955 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9911635b-d76a-43da-910d-2c5c82bf2900-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.369302 4955 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9911635b-d76a-43da-910d-2c5c82bf2900-logs\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.379686 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9911635b-d76a-43da-910d-2c5c82bf2900-kube-api-access-snqv6" (OuterVolumeSpecName: "kube-api-access-snqv6") pod "9911635b-d76a-43da-910d-2c5c82bf2900" (UID: "9911635b-d76a-43da-910d-2c5c82bf2900"). InnerVolumeSpecName "kube-api-access-snqv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.379925 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9911635b-d76a-43da-910d-2c5c82bf2900-scripts" (OuterVolumeSpecName: "scripts") pod "9911635b-d76a-43da-910d-2c5c82bf2900" (UID: "9911635b-d76a-43da-910d-2c5c82bf2900"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.382403 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "9911635b-d76a-43da-910d-2c5c82bf2900" (UID: "9911635b-d76a-43da-910d-2c5c82bf2900"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.418117 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9911635b-d76a-43da-910d-2c5c82bf2900-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9911635b-d76a-43da-910d-2c5c82bf2900" (UID: "9911635b-d76a-43da-910d-2c5c82bf2900"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.433868 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9911635b-d76a-43da-910d-2c5c82bf2900-config-data" (OuterVolumeSpecName: "config-data") pod "9911635b-d76a-43da-910d-2c5c82bf2900" (UID: "9911635b-d76a-43da-910d-2c5c82bf2900"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.443342 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9911635b-d76a-43da-910d-2c5c82bf2900-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "9911635b-d76a-43da-910d-2c5c82bf2900" (UID: "9911635b-d76a-43da-910d-2c5c82bf2900"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.470716 4955 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9911635b-d76a-43da-910d-2c5c82bf2900-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.470768 4955 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.470792 4955 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9911635b-d76a-43da-910d-2c5c82bf2900-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.470802 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9911635b-d76a-43da-910d-2c5c82bf2900-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.470818 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snqv6\" (UniqueName: \"kubernetes.io/projected/9911635b-d76a-43da-910d-2c5c82bf2900-kube-api-access-snqv6\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.470828 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9911635b-d76a-43da-910d-2c5c82bf2900-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.489438 4955 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.572003 4955 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.581327 4955 generic.go:334] "Generic (PLEG): container finished" podID="9911635b-d76a-43da-910d-2c5c82bf2900" containerID="099f1b679ed3eccbe094cddcb02aacc1522fee8432993309a7fb48be275c2a98" exitCode=0 Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.581356 4955 generic.go:334] "Generic (PLEG): container finished" podID="9911635b-d76a-43da-910d-2c5c82bf2900" containerID="34b9043b43dcfcf4e9bd8fe5dd18d0625fa3489ecc3028391af8b117f5445333" exitCode=143 Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.581383 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.581397 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9911635b-d76a-43da-910d-2c5c82bf2900","Type":"ContainerDied","Data":"099f1b679ed3eccbe094cddcb02aacc1522fee8432993309a7fb48be275c2a98"} Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.581448 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9911635b-d76a-43da-910d-2c5c82bf2900","Type":"ContainerDied","Data":"34b9043b43dcfcf4e9bd8fe5dd18d0625fa3489ecc3028391af8b117f5445333"} Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.581463 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9911635b-d76a-43da-910d-2c5c82bf2900","Type":"ContainerDied","Data":"1845f86f5437df26461a6c1a2b6782d0cd00d809dd6d3c7a10ed6af39f93c933"} Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.581492 4955 scope.go:117] "RemoveContainer" containerID="099f1b679ed3eccbe094cddcb02aacc1522fee8432993309a7fb48be275c2a98" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.637770 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.667079 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.680189 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 17 13:22:21 crc kubenswrapper[4955]: E0217 13:22:21.680622 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9911635b-d76a-43da-910d-2c5c82bf2900" containerName="glance-log" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.680638 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="9911635b-d76a-43da-910d-2c5c82bf2900" containerName="glance-log" Feb 17 13:22:21 crc kubenswrapper[4955]: E0217 13:22:21.680659 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9911635b-d76a-43da-910d-2c5c82bf2900" containerName="glance-httpd" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.680668 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="9911635b-d76a-43da-910d-2c5c82bf2900" containerName="glance-httpd" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.680896 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="9911635b-d76a-43da-910d-2c5c82bf2900" containerName="glance-httpd" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.680913 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="9911635b-d76a-43da-910d-2c5c82bf2900" containerName="glance-log" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.682046 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.684702 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.685492 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.689485 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.775546 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"f7a87a1f-b8cc-442c-be49-3117c5fa1702\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.775585 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7a87a1f-b8cc-442c-be49-3117c5fa1702-scripts\") pod \"glance-default-external-api-0\" (UID: \"f7a87a1f-b8cc-442c-be49-3117c5fa1702\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.775611 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7a87a1f-b8cc-442c-be49-3117c5fa1702-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f7a87a1f-b8cc-442c-be49-3117c5fa1702\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.775634 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7a87a1f-b8cc-442c-be49-3117c5fa1702-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f7a87a1f-b8cc-442c-be49-3117c5fa1702\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.775665 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpmz2\" (UniqueName: \"kubernetes.io/projected/f7a87a1f-b8cc-442c-be49-3117c5fa1702-kube-api-access-hpmz2\") pod \"glance-default-external-api-0\" (UID: \"f7a87a1f-b8cc-442c-be49-3117c5fa1702\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.775802 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7a87a1f-b8cc-442c-be49-3117c5fa1702-logs\") pod \"glance-default-external-api-0\" (UID: \"f7a87a1f-b8cc-442c-be49-3117c5fa1702\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.775956 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7a87a1f-b8cc-442c-be49-3117c5fa1702-config-data\") pod \"glance-default-external-api-0\" (UID: \"f7a87a1f-b8cc-442c-be49-3117c5fa1702\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.775984 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f7a87a1f-b8cc-442c-be49-3117c5fa1702-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f7a87a1f-b8cc-442c-be49-3117c5fa1702\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.877270 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7a87a1f-b8cc-442c-be49-3117c5fa1702-config-data\") pod \"glance-default-external-api-0\" (UID: \"f7a87a1f-b8cc-442c-be49-3117c5fa1702\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.877584 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f7a87a1f-b8cc-442c-be49-3117c5fa1702-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f7a87a1f-b8cc-442c-be49-3117c5fa1702\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.877652 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"f7a87a1f-b8cc-442c-be49-3117c5fa1702\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.877670 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7a87a1f-b8cc-442c-be49-3117c5fa1702-scripts\") pod \"glance-default-external-api-0\" (UID: \"f7a87a1f-b8cc-442c-be49-3117c5fa1702\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.877716 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7a87a1f-b8cc-442c-be49-3117c5fa1702-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f7a87a1f-b8cc-442c-be49-3117c5fa1702\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.877743 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7a87a1f-b8cc-442c-be49-3117c5fa1702-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f7a87a1f-b8cc-442c-be49-3117c5fa1702\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.877774 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpmz2\" (UniqueName: \"kubernetes.io/projected/f7a87a1f-b8cc-442c-be49-3117c5fa1702-kube-api-access-hpmz2\") pod \"glance-default-external-api-0\" (UID: \"f7a87a1f-b8cc-442c-be49-3117c5fa1702\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.877829 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7a87a1f-b8cc-442c-be49-3117c5fa1702-logs\") pod \"glance-default-external-api-0\" (UID: \"f7a87a1f-b8cc-442c-be49-3117c5fa1702\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.878537 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7a87a1f-b8cc-442c-be49-3117c5fa1702-logs\") pod \"glance-default-external-api-0\" (UID: \"f7a87a1f-b8cc-442c-be49-3117c5fa1702\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.879213 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f7a87a1f-b8cc-442c-be49-3117c5fa1702-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f7a87a1f-b8cc-442c-be49-3117c5fa1702\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.879336 4955 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"f7a87a1f-b8cc-442c-be49-3117c5fa1702\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-external-api-0" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.899268 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7a87a1f-b8cc-442c-be49-3117c5fa1702-scripts\") pod \"glance-default-external-api-0\" (UID: \"f7a87a1f-b8cc-442c-be49-3117c5fa1702\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.901149 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7a87a1f-b8cc-442c-be49-3117c5fa1702-config-data\") pod \"glance-default-external-api-0\" (UID: \"f7a87a1f-b8cc-442c-be49-3117c5fa1702\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.912488 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7a87a1f-b8cc-442c-be49-3117c5fa1702-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f7a87a1f-b8cc-442c-be49-3117c5fa1702\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.915599 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7a87a1f-b8cc-442c-be49-3117c5fa1702-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f7a87a1f-b8cc-442c-be49-3117c5fa1702\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.924906 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpmz2\" (UniqueName: \"kubernetes.io/projected/f7a87a1f-b8cc-442c-be49-3117c5fa1702-kube-api-access-hpmz2\") pod \"glance-default-external-api-0\" (UID: \"f7a87a1f-b8cc-442c-be49-3117c5fa1702\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:21 crc kubenswrapper[4955]: I0217 13:22:21.981064 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"f7a87a1f-b8cc-442c-be49-3117c5fa1702\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:22 crc kubenswrapper[4955]: I0217 13:22:22.041188 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 17 13:22:22 crc kubenswrapper[4955]: I0217 13:22:22.238908 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9911635b-d76a-43da-910d-2c5c82bf2900" path="/var/lib/kubelet/pods/9911635b-d76a-43da-910d-2c5c82bf2900/volumes" Feb 17 13:22:22 crc kubenswrapper[4955]: I0217 13:22:22.240470 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da645eab-8b87-4ef3-b10a-2401382e5f19" path="/var/lib/kubelet/pods/da645eab-8b87-4ef3-b10a-2401382e5f19/volumes" Feb 17 13:22:22 crc kubenswrapper[4955]: I0217 13:22:22.560582 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.253690 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-76cdc74c45-lss64"] Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.288163 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.304874 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5788cc4448-245bs"] Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.306237 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5788cc4448-245bs" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.308115 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.320438 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5788cc4448-245bs"] Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.403163 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.414195 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf-scripts\") pod \"horizon-5788cc4448-245bs\" (UID: \"1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf\") " pod="openstack/horizon-5788cc4448-245bs" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.414272 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf-config-data\") pod \"horizon-5788cc4448-245bs\" (UID: \"1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf\") " pod="openstack/horizon-5788cc4448-245bs" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.414294 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf-combined-ca-bundle\") pod \"horizon-5788cc4448-245bs\" (UID: \"1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf\") " pod="openstack/horizon-5788cc4448-245bs" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.414347 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhbs6\" (UniqueName: \"kubernetes.io/projected/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf-kube-api-access-jhbs6\") pod \"horizon-5788cc4448-245bs\" (UID: \"1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf\") " pod="openstack/horizon-5788cc4448-245bs" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.414367 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf-horizon-secret-key\") pod \"horizon-5788cc4448-245bs\" (UID: \"1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf\") " pod="openstack/horizon-5788cc4448-245bs" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.414421 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf-logs\") pod \"horizon-5788cc4448-245bs\" (UID: \"1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf\") " pod="openstack/horizon-5788cc4448-245bs" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.414438 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf-horizon-tls-certs\") pod \"horizon-5788cc4448-245bs\" (UID: \"1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf\") " pod="openstack/horizon-5788cc4448-245bs" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.424841 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5bbc974f4f-sf5xt"] Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.457270 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7fffbcf7cb-wr74b"] Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.459326 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7fffbcf7cb-wr74b" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.467116 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7fffbcf7cb-wr74b"] Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.515760 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf-config-data\") pod \"horizon-5788cc4448-245bs\" (UID: \"1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf\") " pod="openstack/horizon-5788cc4448-245bs" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.515829 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf-combined-ca-bundle\") pod \"horizon-5788cc4448-245bs\" (UID: \"1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf\") " pod="openstack/horizon-5788cc4448-245bs" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.515886 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6cb09d2-ad7d-4486-a41d-89023789fd87-horizon-tls-certs\") pod \"horizon-7fffbcf7cb-wr74b\" (UID: \"d6cb09d2-ad7d-4486-a41d-89023789fd87\") " pod="openstack/horizon-7fffbcf7cb-wr74b" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.515906 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6cb09d2-ad7d-4486-a41d-89023789fd87-combined-ca-bundle\") pod \"horizon-7fffbcf7cb-wr74b\" (UID: \"d6cb09d2-ad7d-4486-a41d-89023789fd87\") " pod="openstack/horizon-7fffbcf7cb-wr74b" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.515928 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhbs6\" (UniqueName: \"kubernetes.io/projected/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf-kube-api-access-jhbs6\") pod \"horizon-5788cc4448-245bs\" (UID: \"1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf\") " pod="openstack/horizon-5788cc4448-245bs" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.515949 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d6cb09d2-ad7d-4486-a41d-89023789fd87-horizon-secret-key\") pod \"horizon-7fffbcf7cb-wr74b\" (UID: \"d6cb09d2-ad7d-4486-a41d-89023789fd87\") " pod="openstack/horizon-7fffbcf7cb-wr74b" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.515968 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf-horizon-secret-key\") pod \"horizon-5788cc4448-245bs\" (UID: \"1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf\") " pod="openstack/horizon-5788cc4448-245bs" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.516027 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf-logs\") pod \"horizon-5788cc4448-245bs\" (UID: \"1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf\") " pod="openstack/horizon-5788cc4448-245bs" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.516048 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf-horizon-tls-certs\") pod \"horizon-5788cc4448-245bs\" (UID: \"1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf\") " pod="openstack/horizon-5788cc4448-245bs" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.516068 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf-scripts\") pod \"horizon-5788cc4448-245bs\" (UID: \"1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf\") " pod="openstack/horizon-5788cc4448-245bs" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.516088 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d6cb09d2-ad7d-4486-a41d-89023789fd87-scripts\") pod \"horizon-7fffbcf7cb-wr74b\" (UID: \"d6cb09d2-ad7d-4486-a41d-89023789fd87\") " pod="openstack/horizon-7fffbcf7cb-wr74b" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.516110 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d6cb09d2-ad7d-4486-a41d-89023789fd87-config-data\") pod \"horizon-7fffbcf7cb-wr74b\" (UID: \"d6cb09d2-ad7d-4486-a41d-89023789fd87\") " pod="openstack/horizon-7fffbcf7cb-wr74b" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.516132 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jqbf\" (UniqueName: \"kubernetes.io/projected/d6cb09d2-ad7d-4486-a41d-89023789fd87-kube-api-access-2jqbf\") pod \"horizon-7fffbcf7cb-wr74b\" (UID: \"d6cb09d2-ad7d-4486-a41d-89023789fd87\") " pod="openstack/horizon-7fffbcf7cb-wr74b" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.516155 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6cb09d2-ad7d-4486-a41d-89023789fd87-logs\") pod \"horizon-7fffbcf7cb-wr74b\" (UID: \"d6cb09d2-ad7d-4486-a41d-89023789fd87\") " pod="openstack/horizon-7fffbcf7cb-wr74b" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.516502 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf-logs\") pod \"horizon-5788cc4448-245bs\" (UID: \"1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf\") " pod="openstack/horizon-5788cc4448-245bs" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.517164 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf-scripts\") pod \"horizon-5788cc4448-245bs\" (UID: \"1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf\") " pod="openstack/horizon-5788cc4448-245bs" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.521294 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf-horizon-tls-certs\") pod \"horizon-5788cc4448-245bs\" (UID: \"1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf\") " pod="openstack/horizon-5788cc4448-245bs" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.528518 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf-combined-ca-bundle\") pod \"horizon-5788cc4448-245bs\" (UID: \"1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf\") " pod="openstack/horizon-5788cc4448-245bs" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.529386 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf-config-data\") pod \"horizon-5788cc4448-245bs\" (UID: \"1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf\") " pod="openstack/horizon-5788cc4448-245bs" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.531734 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhbs6\" (UniqueName: \"kubernetes.io/projected/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf-kube-api-access-jhbs6\") pod \"horizon-5788cc4448-245bs\" (UID: \"1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf\") " pod="openstack/horizon-5788cc4448-245bs" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.542557 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf-horizon-secret-key\") pod \"horizon-5788cc4448-245bs\" (UID: \"1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf\") " pod="openstack/horizon-5788cc4448-245bs" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.618103 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d6cb09d2-ad7d-4486-a41d-89023789fd87-scripts\") pod \"horizon-7fffbcf7cb-wr74b\" (UID: \"d6cb09d2-ad7d-4486-a41d-89023789fd87\") " pod="openstack/horizon-7fffbcf7cb-wr74b" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.618734 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d6cb09d2-ad7d-4486-a41d-89023789fd87-scripts\") pod \"horizon-7fffbcf7cb-wr74b\" (UID: \"d6cb09d2-ad7d-4486-a41d-89023789fd87\") " pod="openstack/horizon-7fffbcf7cb-wr74b" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.618858 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d6cb09d2-ad7d-4486-a41d-89023789fd87-config-data\") pod \"horizon-7fffbcf7cb-wr74b\" (UID: \"d6cb09d2-ad7d-4486-a41d-89023789fd87\") " pod="openstack/horizon-7fffbcf7cb-wr74b" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.620007 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d6cb09d2-ad7d-4486-a41d-89023789fd87-config-data\") pod \"horizon-7fffbcf7cb-wr74b\" (UID: \"d6cb09d2-ad7d-4486-a41d-89023789fd87\") " pod="openstack/horizon-7fffbcf7cb-wr74b" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.618893 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jqbf\" (UniqueName: \"kubernetes.io/projected/d6cb09d2-ad7d-4486-a41d-89023789fd87-kube-api-access-2jqbf\") pod \"horizon-7fffbcf7cb-wr74b\" (UID: \"d6cb09d2-ad7d-4486-a41d-89023789fd87\") " pod="openstack/horizon-7fffbcf7cb-wr74b" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.620110 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6cb09d2-ad7d-4486-a41d-89023789fd87-logs\") pod \"horizon-7fffbcf7cb-wr74b\" (UID: \"d6cb09d2-ad7d-4486-a41d-89023789fd87\") " pod="openstack/horizon-7fffbcf7cb-wr74b" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.620224 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6cb09d2-ad7d-4486-a41d-89023789fd87-horizon-tls-certs\") pod \"horizon-7fffbcf7cb-wr74b\" (UID: \"d6cb09d2-ad7d-4486-a41d-89023789fd87\") " pod="openstack/horizon-7fffbcf7cb-wr74b" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.620254 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6cb09d2-ad7d-4486-a41d-89023789fd87-combined-ca-bundle\") pod \"horizon-7fffbcf7cb-wr74b\" (UID: \"d6cb09d2-ad7d-4486-a41d-89023789fd87\") " pod="openstack/horizon-7fffbcf7cb-wr74b" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.620300 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d6cb09d2-ad7d-4486-a41d-89023789fd87-horizon-secret-key\") pod \"horizon-7fffbcf7cb-wr74b\" (UID: \"d6cb09d2-ad7d-4486-a41d-89023789fd87\") " pod="openstack/horizon-7fffbcf7cb-wr74b" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.624670 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d6cb09d2-ad7d-4486-a41d-89023789fd87-horizon-secret-key\") pod \"horizon-7fffbcf7cb-wr74b\" (UID: \"d6cb09d2-ad7d-4486-a41d-89023789fd87\") " pod="openstack/horizon-7fffbcf7cb-wr74b" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.625723 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6cb09d2-ad7d-4486-a41d-89023789fd87-logs\") pod \"horizon-7fffbcf7cb-wr74b\" (UID: \"d6cb09d2-ad7d-4486-a41d-89023789fd87\") " pod="openstack/horizon-7fffbcf7cb-wr74b" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.627436 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6cb09d2-ad7d-4486-a41d-89023789fd87-horizon-tls-certs\") pod \"horizon-7fffbcf7cb-wr74b\" (UID: \"d6cb09d2-ad7d-4486-a41d-89023789fd87\") " pod="openstack/horizon-7fffbcf7cb-wr74b" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.629993 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6cb09d2-ad7d-4486-a41d-89023789fd87-combined-ca-bundle\") pod \"horizon-7fffbcf7cb-wr74b\" (UID: \"d6cb09d2-ad7d-4486-a41d-89023789fd87\") " pod="openstack/horizon-7fffbcf7cb-wr74b" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.636226 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jqbf\" (UniqueName: \"kubernetes.io/projected/d6cb09d2-ad7d-4486-a41d-89023789fd87-kube-api-access-2jqbf\") pod \"horizon-7fffbcf7cb-wr74b\" (UID: \"d6cb09d2-ad7d-4486-a41d-89023789fd87\") " pod="openstack/horizon-7fffbcf7cb-wr74b" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.648091 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5788cc4448-245bs" Feb 17 13:22:23 crc kubenswrapper[4955]: I0217 13:22:23.782828 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7fffbcf7cb-wr74b" Feb 17 13:22:25 crc kubenswrapper[4955]: I0217 13:22:25.527000 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-56df8fb6b7-k2p8k" Feb 17 13:22:25 crc kubenswrapper[4955]: I0217 13:22:25.585137 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-xxclj"] Feb 17 13:22:25 crc kubenswrapper[4955]: I0217 13:22:25.585351 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5f59b8f679-xxclj" podUID="7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a" containerName="dnsmasq-dns" containerID="cri-o://4fcb9bb9cd8261f98d8dc33fd2041813f105d13915a97619409dd56f1dd69a52" gracePeriod=10 Feb 17 13:22:26 crc kubenswrapper[4955]: I0217 13:22:26.655071 4955 generic.go:334] "Generic (PLEG): container finished" podID="7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a" containerID="4fcb9bb9cd8261f98d8dc33fd2041813f105d13915a97619409dd56f1dd69a52" exitCode=0 Feb 17 13:22:26 crc kubenswrapper[4955]: I0217 13:22:26.655127 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-xxclj" event={"ID":"7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a","Type":"ContainerDied","Data":"4fcb9bb9cd8261f98d8dc33fd2041813f105d13915a97619409dd56f1dd69a52"} Feb 17 13:22:29 crc kubenswrapper[4955]: I0217 13:22:29.333076 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f59b8f679-xxclj" podUID="7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.126:5353: connect: connection refused" Feb 17 13:22:34 crc kubenswrapper[4955]: E0217 13:22:34.239755 4955 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Feb 17 13:22:34 crc kubenswrapper[4955]: E0217 13:22:34.240429 4955 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n669h59ch5b8h5dbh649h5d8hbh4hch89h576h564h59bh54bh64fh84h57bh6dh568h559h85h5c4h5f7hf7h5ch575h5bhddh558h7fh675h686q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n8zpw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-5bbc974f4f-sf5xt_openstack(7eea91a6-8688-4c16-a45e-5ec47ba5fba6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 17 13:22:34 crc kubenswrapper[4955]: E0217 13:22:34.243272 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-5bbc974f4f-sf5xt" podUID="7eea91a6-8688-4c16-a45e-5ec47ba5fba6" Feb 17 13:22:34 crc kubenswrapper[4955]: I0217 13:22:34.331380 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f59b8f679-xxclj" podUID="7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.126:5353: connect: connection refused" Feb 17 13:22:34 crc kubenswrapper[4955]: I0217 13:22:34.574769 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 13:22:34 crc kubenswrapper[4955]: I0217 13:22:34.574859 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 13:22:34 crc kubenswrapper[4955]: W0217 13:22:34.591575 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9aac5b59_a6ed_43cc_888d_a583384f1ede.slice/crio-14485051310b551cede05b8013894ece75fab9d72684e688123b780cefe44c59 WatchSource:0}: Error finding container 14485051310b551cede05b8013894ece75fab9d72684e688123b780cefe44c59: Status 404 returned error can't find the container with id 14485051310b551cede05b8013894ece75fab9d72684e688123b780cefe44c59 Feb 17 13:22:34 crc kubenswrapper[4955]: E0217 13:22:34.599925 4955 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Feb 17 13:22:34 crc kubenswrapper[4955]: E0217 13:22:34.600058 4955 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5f4h89h569hbdh5dh5dh66hb5h7hfh95h9fh587h575h679hffh58dh64fh66chbchb5h54h544h79hbfh674h58h555h6bh646h596h95q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4xcff,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(01547c5a-5ca7-435b-ba75-52e075800cf3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 17 13:22:34 crc kubenswrapper[4955]: I0217 13:22:34.737035 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9aac5b59-a6ed-43cc-888d-a583384f1ede","Type":"ContainerStarted","Data":"14485051310b551cede05b8013894ece75fab9d72684e688123b780cefe44c59"} Feb 17 13:22:39 crc kubenswrapper[4955]: I0217 13:22:39.331236 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f59b8f679-xxclj" podUID="7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.126:5353: connect: connection refused" Feb 17 13:22:39 crc kubenswrapper[4955]: I0217 13:22:39.331865 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5f59b8f679-xxclj" Feb 17 13:22:43 crc kubenswrapper[4955]: I0217 13:22:43.252158 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4pqz5" Feb 17 13:22:43 crc kubenswrapper[4955]: I0217 13:22:43.389911 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e-fernet-keys\") pod \"fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e\" (UID: \"fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e\") " Feb 17 13:22:43 crc kubenswrapper[4955]: I0217 13:22:43.389992 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fj7rl\" (UniqueName: \"kubernetes.io/projected/fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e-kube-api-access-fj7rl\") pod \"fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e\" (UID: \"fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e\") " Feb 17 13:22:43 crc kubenswrapper[4955]: I0217 13:22:43.390017 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e-credential-keys\") pod \"fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e\" (UID: \"fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e\") " Feb 17 13:22:43 crc kubenswrapper[4955]: I0217 13:22:43.390124 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e-combined-ca-bundle\") pod \"fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e\" (UID: \"fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e\") " Feb 17 13:22:43 crc kubenswrapper[4955]: I0217 13:22:43.390154 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e-scripts\") pod \"fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e\" (UID: \"fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e\") " Feb 17 13:22:43 crc kubenswrapper[4955]: I0217 13:22:43.390220 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e-config-data\") pod \"fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e\" (UID: \"fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e\") " Feb 17 13:22:43 crc kubenswrapper[4955]: I0217 13:22:43.396833 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e-kube-api-access-fj7rl" (OuterVolumeSpecName: "kube-api-access-fj7rl") pod "fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e" (UID: "fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e"). InnerVolumeSpecName "kube-api-access-fj7rl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:22:43 crc kubenswrapper[4955]: I0217 13:22:43.398495 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e" (UID: "fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:22:43 crc kubenswrapper[4955]: I0217 13:22:43.399088 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e" (UID: "fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:22:43 crc kubenswrapper[4955]: I0217 13:22:43.399218 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e-scripts" (OuterVolumeSpecName: "scripts") pod "fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e" (UID: "fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:22:43 crc kubenswrapper[4955]: I0217 13:22:43.420135 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e" (UID: "fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:22:43 crc kubenswrapper[4955]: I0217 13:22:43.420324 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e-config-data" (OuterVolumeSpecName: "config-data") pod "fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e" (UID: "fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:22:43 crc kubenswrapper[4955]: I0217 13:22:43.492276 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fj7rl\" (UniqueName: \"kubernetes.io/projected/fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e-kube-api-access-fj7rl\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:43 crc kubenswrapper[4955]: I0217 13:22:43.492660 4955 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:43 crc kubenswrapper[4955]: I0217 13:22:43.492675 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:43 crc kubenswrapper[4955]: I0217 13:22:43.492690 4955 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:43 crc kubenswrapper[4955]: I0217 13:22:43.492704 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:43 crc kubenswrapper[4955]: I0217 13:22:43.492715 4955 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:43 crc kubenswrapper[4955]: E0217 13:22:43.715684 4955 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Feb 17 13:22:43 crc kubenswrapper[4955]: E0217 13:22:43.716050 4955 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rl7sq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-tr2q9_openstack(1d695519-3143-4f53-ab46-5823bc04dc0d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 17 13:22:43 crc kubenswrapper[4955]: E0217 13:22:43.717290 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-tr2q9" podUID="1d695519-3143-4f53-ab46-5823bc04dc0d" Feb 17 13:22:43 crc kubenswrapper[4955]: I0217 13:22:43.792560 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5bbc974f4f-sf5xt" Feb 17 13:22:43 crc kubenswrapper[4955]: I0217 13:22:43.810149 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5bbc974f4f-sf5xt" event={"ID":"7eea91a6-8688-4c16-a45e-5ec47ba5fba6","Type":"ContainerDied","Data":"674cc704ff6179471a37780d640752c3ad36343953dca4555ec93b61a6d59b1a"} Feb 17 13:22:43 crc kubenswrapper[4955]: I0217 13:22:43.810245 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5bbc974f4f-sf5xt" Feb 17 13:22:43 crc kubenswrapper[4955]: I0217 13:22:43.815613 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4pqz5" event={"ID":"fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e","Type":"ContainerDied","Data":"fc805665a5a34766d800b2d06f2a7ef6aa2e74c1684cae4a273c47ffcc73940a"} Feb 17 13:22:43 crc kubenswrapper[4955]: I0217 13:22:43.815653 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc805665a5a34766d800b2d06f2a7ef6aa2e74c1684cae4a273c47ffcc73940a" Feb 17 13:22:43 crc kubenswrapper[4955]: I0217 13:22:43.815702 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4pqz5" Feb 17 13:22:43 crc kubenswrapper[4955]: E0217 13:22:43.818387 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-tr2q9" podUID="1d695519-3143-4f53-ab46-5823bc04dc0d" Feb 17 13:22:43 crc kubenswrapper[4955]: I0217 13:22:43.905981 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7eea91a6-8688-4c16-a45e-5ec47ba5fba6-config-data\") pod \"7eea91a6-8688-4c16-a45e-5ec47ba5fba6\" (UID: \"7eea91a6-8688-4c16-a45e-5ec47ba5fba6\") " Feb 17 13:22:43 crc kubenswrapper[4955]: I0217 13:22:43.906103 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7eea91a6-8688-4c16-a45e-5ec47ba5fba6-scripts\") pod \"7eea91a6-8688-4c16-a45e-5ec47ba5fba6\" (UID: \"7eea91a6-8688-4c16-a45e-5ec47ba5fba6\") " Feb 17 13:22:43 crc kubenswrapper[4955]: I0217 13:22:43.906139 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n8zpw\" (UniqueName: \"kubernetes.io/projected/7eea91a6-8688-4c16-a45e-5ec47ba5fba6-kube-api-access-n8zpw\") pod \"7eea91a6-8688-4c16-a45e-5ec47ba5fba6\" (UID: \"7eea91a6-8688-4c16-a45e-5ec47ba5fba6\") " Feb 17 13:22:43 crc kubenswrapper[4955]: I0217 13:22:43.906284 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7eea91a6-8688-4c16-a45e-5ec47ba5fba6-horizon-secret-key\") pod \"7eea91a6-8688-4c16-a45e-5ec47ba5fba6\" (UID: \"7eea91a6-8688-4c16-a45e-5ec47ba5fba6\") " Feb 17 13:22:43 crc kubenswrapper[4955]: I0217 13:22:43.906857 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7eea91a6-8688-4c16-a45e-5ec47ba5fba6-logs\") pod \"7eea91a6-8688-4c16-a45e-5ec47ba5fba6\" (UID: \"7eea91a6-8688-4c16-a45e-5ec47ba5fba6\") " Feb 17 13:22:43 crc kubenswrapper[4955]: I0217 13:22:43.906898 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7eea91a6-8688-4c16-a45e-5ec47ba5fba6-scripts" (OuterVolumeSpecName: "scripts") pod "7eea91a6-8688-4c16-a45e-5ec47ba5fba6" (UID: "7eea91a6-8688-4c16-a45e-5ec47ba5fba6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:22:43 crc kubenswrapper[4955]: I0217 13:22:43.906905 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7eea91a6-8688-4c16-a45e-5ec47ba5fba6-config-data" (OuterVolumeSpecName: "config-data") pod "7eea91a6-8688-4c16-a45e-5ec47ba5fba6" (UID: "7eea91a6-8688-4c16-a45e-5ec47ba5fba6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:22:43 crc kubenswrapper[4955]: I0217 13:22:43.907351 4955 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7eea91a6-8688-4c16-a45e-5ec47ba5fba6-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:43 crc kubenswrapper[4955]: I0217 13:22:43.907375 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7eea91a6-8688-4c16-a45e-5ec47ba5fba6-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:43 crc kubenswrapper[4955]: I0217 13:22:43.907340 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7eea91a6-8688-4c16-a45e-5ec47ba5fba6-logs" (OuterVolumeSpecName: "logs") pod "7eea91a6-8688-4c16-a45e-5ec47ba5fba6" (UID: "7eea91a6-8688-4c16-a45e-5ec47ba5fba6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:22:43 crc kubenswrapper[4955]: I0217 13:22:43.911240 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7eea91a6-8688-4c16-a45e-5ec47ba5fba6-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "7eea91a6-8688-4c16-a45e-5ec47ba5fba6" (UID: "7eea91a6-8688-4c16-a45e-5ec47ba5fba6"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:22:43 crc kubenswrapper[4955]: I0217 13:22:43.913477 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7eea91a6-8688-4c16-a45e-5ec47ba5fba6-kube-api-access-n8zpw" (OuterVolumeSpecName: "kube-api-access-n8zpw") pod "7eea91a6-8688-4c16-a45e-5ec47ba5fba6" (UID: "7eea91a6-8688-4c16-a45e-5ec47ba5fba6"). InnerVolumeSpecName "kube-api-access-n8zpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:22:44 crc kubenswrapper[4955]: I0217 13:22:44.008944 4955 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7eea91a6-8688-4c16-a45e-5ec47ba5fba6-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:44 crc kubenswrapper[4955]: I0217 13:22:44.009052 4955 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7eea91a6-8688-4c16-a45e-5ec47ba5fba6-logs\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:44 crc kubenswrapper[4955]: I0217 13:22:44.009067 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n8zpw\" (UniqueName: \"kubernetes.io/projected/7eea91a6-8688-4c16-a45e-5ec47ba5fba6-kube-api-access-n8zpw\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:44 crc kubenswrapper[4955]: I0217 13:22:44.198852 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5bbc974f4f-sf5xt"] Feb 17 13:22:44 crc kubenswrapper[4955]: I0217 13:22:44.207803 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5bbc974f4f-sf5xt"] Feb 17 13:22:44 crc kubenswrapper[4955]: I0217 13:22:44.235908 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7eea91a6-8688-4c16-a45e-5ec47ba5fba6" path="/var/lib/kubelet/pods/7eea91a6-8688-4c16-a45e-5ec47ba5fba6/volumes" Feb 17 13:22:44 crc kubenswrapper[4955]: I0217 13:22:44.332672 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-4pqz5"] Feb 17 13:22:44 crc kubenswrapper[4955]: I0217 13:22:44.340674 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-4pqz5"] Feb 17 13:22:44 crc kubenswrapper[4955]: I0217 13:22:44.439649 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-sstln"] Feb 17 13:22:44 crc kubenswrapper[4955]: E0217 13:22:44.440160 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e" containerName="keystone-bootstrap" Feb 17 13:22:44 crc kubenswrapper[4955]: I0217 13:22:44.440178 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e" containerName="keystone-bootstrap" Feb 17 13:22:44 crc kubenswrapper[4955]: I0217 13:22:44.440398 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e" containerName="keystone-bootstrap" Feb 17 13:22:44 crc kubenswrapper[4955]: I0217 13:22:44.441159 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sstln" Feb 17 13:22:44 crc kubenswrapper[4955]: I0217 13:22:44.443106 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 17 13:22:44 crc kubenswrapper[4955]: I0217 13:22:44.443223 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 17 13:22:44 crc kubenswrapper[4955]: I0217 13:22:44.443398 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-qh96v" Feb 17 13:22:44 crc kubenswrapper[4955]: I0217 13:22:44.443698 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 17 13:22:44 crc kubenswrapper[4955]: I0217 13:22:44.445260 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 17 13:22:44 crc kubenswrapper[4955]: I0217 13:22:44.486380 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-sstln"] Feb 17 13:22:44 crc kubenswrapper[4955]: I0217 13:22:44.518534 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50465113-9c90-4a32-8485-0b976374107a-combined-ca-bundle\") pod \"keystone-bootstrap-sstln\" (UID: \"50465113-9c90-4a32-8485-0b976374107a\") " pod="openstack/keystone-bootstrap-sstln" Feb 17 13:22:44 crc kubenswrapper[4955]: I0217 13:22:44.518760 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50465113-9c90-4a32-8485-0b976374107a-scripts\") pod \"keystone-bootstrap-sstln\" (UID: \"50465113-9c90-4a32-8485-0b976374107a\") " pod="openstack/keystone-bootstrap-sstln" Feb 17 13:22:44 crc kubenswrapper[4955]: I0217 13:22:44.518881 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/50465113-9c90-4a32-8485-0b976374107a-credential-keys\") pod \"keystone-bootstrap-sstln\" (UID: \"50465113-9c90-4a32-8485-0b976374107a\") " pod="openstack/keystone-bootstrap-sstln" Feb 17 13:22:44 crc kubenswrapper[4955]: I0217 13:22:44.518949 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/50465113-9c90-4a32-8485-0b976374107a-fernet-keys\") pod \"keystone-bootstrap-sstln\" (UID: \"50465113-9c90-4a32-8485-0b976374107a\") " pod="openstack/keystone-bootstrap-sstln" Feb 17 13:22:44 crc kubenswrapper[4955]: I0217 13:22:44.519036 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50465113-9c90-4a32-8485-0b976374107a-config-data\") pod \"keystone-bootstrap-sstln\" (UID: \"50465113-9c90-4a32-8485-0b976374107a\") " pod="openstack/keystone-bootstrap-sstln" Feb 17 13:22:44 crc kubenswrapper[4955]: I0217 13:22:44.519177 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xxgd\" (UniqueName: \"kubernetes.io/projected/50465113-9c90-4a32-8485-0b976374107a-kube-api-access-9xxgd\") pod \"keystone-bootstrap-sstln\" (UID: \"50465113-9c90-4a32-8485-0b976374107a\") " pod="openstack/keystone-bootstrap-sstln" Feb 17 13:22:44 crc kubenswrapper[4955]: I0217 13:22:44.628900 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xxgd\" (UniqueName: \"kubernetes.io/projected/50465113-9c90-4a32-8485-0b976374107a-kube-api-access-9xxgd\") pod \"keystone-bootstrap-sstln\" (UID: \"50465113-9c90-4a32-8485-0b976374107a\") " pod="openstack/keystone-bootstrap-sstln" Feb 17 13:22:44 crc kubenswrapper[4955]: I0217 13:22:44.628980 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50465113-9c90-4a32-8485-0b976374107a-combined-ca-bundle\") pod \"keystone-bootstrap-sstln\" (UID: \"50465113-9c90-4a32-8485-0b976374107a\") " pod="openstack/keystone-bootstrap-sstln" Feb 17 13:22:44 crc kubenswrapper[4955]: I0217 13:22:44.629010 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50465113-9c90-4a32-8485-0b976374107a-scripts\") pod \"keystone-bootstrap-sstln\" (UID: \"50465113-9c90-4a32-8485-0b976374107a\") " pod="openstack/keystone-bootstrap-sstln" Feb 17 13:22:44 crc kubenswrapper[4955]: I0217 13:22:44.629034 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/50465113-9c90-4a32-8485-0b976374107a-credential-keys\") pod \"keystone-bootstrap-sstln\" (UID: \"50465113-9c90-4a32-8485-0b976374107a\") " pod="openstack/keystone-bootstrap-sstln" Feb 17 13:22:44 crc kubenswrapper[4955]: I0217 13:22:44.629057 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/50465113-9c90-4a32-8485-0b976374107a-fernet-keys\") pod \"keystone-bootstrap-sstln\" (UID: \"50465113-9c90-4a32-8485-0b976374107a\") " pod="openstack/keystone-bootstrap-sstln" Feb 17 13:22:44 crc kubenswrapper[4955]: I0217 13:22:44.629094 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50465113-9c90-4a32-8485-0b976374107a-config-data\") pod \"keystone-bootstrap-sstln\" (UID: \"50465113-9c90-4a32-8485-0b976374107a\") " pod="openstack/keystone-bootstrap-sstln" Feb 17 13:22:44 crc kubenswrapper[4955]: I0217 13:22:44.632660 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50465113-9c90-4a32-8485-0b976374107a-scripts\") pod \"keystone-bootstrap-sstln\" (UID: \"50465113-9c90-4a32-8485-0b976374107a\") " pod="openstack/keystone-bootstrap-sstln" Feb 17 13:22:44 crc kubenswrapper[4955]: I0217 13:22:44.632933 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50465113-9c90-4a32-8485-0b976374107a-config-data\") pod \"keystone-bootstrap-sstln\" (UID: \"50465113-9c90-4a32-8485-0b976374107a\") " pod="openstack/keystone-bootstrap-sstln" Feb 17 13:22:44 crc kubenswrapper[4955]: I0217 13:22:44.634584 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/50465113-9c90-4a32-8485-0b976374107a-credential-keys\") pod \"keystone-bootstrap-sstln\" (UID: \"50465113-9c90-4a32-8485-0b976374107a\") " pod="openstack/keystone-bootstrap-sstln" Feb 17 13:22:44 crc kubenswrapper[4955]: I0217 13:22:44.636378 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/50465113-9c90-4a32-8485-0b976374107a-fernet-keys\") pod \"keystone-bootstrap-sstln\" (UID: \"50465113-9c90-4a32-8485-0b976374107a\") " pod="openstack/keystone-bootstrap-sstln" Feb 17 13:22:44 crc kubenswrapper[4955]: I0217 13:22:44.642532 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50465113-9c90-4a32-8485-0b976374107a-combined-ca-bundle\") pod \"keystone-bootstrap-sstln\" (UID: \"50465113-9c90-4a32-8485-0b976374107a\") " pod="openstack/keystone-bootstrap-sstln" Feb 17 13:22:44 crc kubenswrapper[4955]: I0217 13:22:44.649173 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xxgd\" (UniqueName: \"kubernetes.io/projected/50465113-9c90-4a32-8485-0b976374107a-kube-api-access-9xxgd\") pod \"keystone-bootstrap-sstln\" (UID: \"50465113-9c90-4a32-8485-0b976374107a\") " pod="openstack/keystone-bootstrap-sstln" Feb 17 13:22:44 crc kubenswrapper[4955]: I0217 13:22:44.796857 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sstln" Feb 17 13:22:44 crc kubenswrapper[4955]: I0217 13:22:44.949707 4955 scope.go:117] "RemoveContainer" containerID="34b9043b43dcfcf4e9bd8fe5dd18d0625fa3489ecc3028391af8b117f5445333" Feb 17 13:22:44 crc kubenswrapper[4955]: E0217 13:22:44.991165 4955 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Feb 17 13:22:44 crc kubenswrapper[4955]: E0217 13:22:44.991332 4955 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6w2b6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-w6s4s_openstack(e1c2d0c5-7647-447f-a1a6-b37d8c09310c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 17 13:22:44 crc kubenswrapper[4955]: E0217 13:22:44.993363 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-w6s4s" podUID="e1c2d0c5-7647-447f-a1a6-b37d8c09310c" Feb 17 13:22:45 crc kubenswrapper[4955]: I0217 13:22:45.051495 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-xxclj" Feb 17 13:22:45 crc kubenswrapper[4955]: I0217 13:22:45.136363 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a-dns-svc\") pod \"7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a\" (UID: \"7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a\") " Feb 17 13:22:45 crc kubenswrapper[4955]: I0217 13:22:45.136538 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlkwb\" (UniqueName: \"kubernetes.io/projected/7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a-kube-api-access-vlkwb\") pod \"7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a\" (UID: \"7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a\") " Feb 17 13:22:45 crc kubenswrapper[4955]: I0217 13:22:45.136571 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a-ovsdbserver-nb\") pod \"7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a\" (UID: \"7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a\") " Feb 17 13:22:45 crc kubenswrapper[4955]: I0217 13:22:45.136674 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a-config\") pod \"7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a\" (UID: \"7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a\") " Feb 17 13:22:45 crc kubenswrapper[4955]: I0217 13:22:45.136695 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a-dns-swift-storage-0\") pod \"7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a\" (UID: \"7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a\") " Feb 17 13:22:45 crc kubenswrapper[4955]: I0217 13:22:45.136715 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a-ovsdbserver-sb\") pod \"7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a\" (UID: \"7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a\") " Feb 17 13:22:45 crc kubenswrapper[4955]: I0217 13:22:45.140150 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a-kube-api-access-vlkwb" (OuterVolumeSpecName: "kube-api-access-vlkwb") pod "7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a" (UID: "7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a"). InnerVolumeSpecName "kube-api-access-vlkwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:22:45 crc kubenswrapper[4955]: I0217 13:22:45.182595 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a" (UID: "7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:22:45 crc kubenswrapper[4955]: I0217 13:22:45.183990 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a" (UID: "7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:22:45 crc kubenswrapper[4955]: I0217 13:22:45.186555 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a-config" (OuterVolumeSpecName: "config") pod "7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a" (UID: "7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:22:45 crc kubenswrapper[4955]: I0217 13:22:45.198282 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a" (UID: "7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:22:45 crc kubenswrapper[4955]: I0217 13:22:45.199537 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a" (UID: "7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:22:45 crc kubenswrapper[4955]: I0217 13:22:45.239085 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlkwb\" (UniqueName: \"kubernetes.io/projected/7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a-kube-api-access-vlkwb\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:45 crc kubenswrapper[4955]: I0217 13:22:45.239118 4955 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:45 crc kubenswrapper[4955]: I0217 13:22:45.239133 4955 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:45 crc kubenswrapper[4955]: I0217 13:22:45.239144 4955 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:45 crc kubenswrapper[4955]: I0217 13:22:45.239157 4955 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:45 crc kubenswrapper[4955]: I0217 13:22:45.239169 4955 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:45 crc kubenswrapper[4955]: I0217 13:22:45.555566 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5788cc4448-245bs"] Feb 17 13:22:45 crc kubenswrapper[4955]: I0217 13:22:45.582312 4955 scope.go:117] "RemoveContainer" containerID="099f1b679ed3eccbe094cddcb02aacc1522fee8432993309a7fb48be275c2a98" Feb 17 13:22:45 crc kubenswrapper[4955]: E0217 13:22:45.583121 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"099f1b679ed3eccbe094cddcb02aacc1522fee8432993309a7fb48be275c2a98\": container with ID starting with 099f1b679ed3eccbe094cddcb02aacc1522fee8432993309a7fb48be275c2a98 not found: ID does not exist" containerID="099f1b679ed3eccbe094cddcb02aacc1522fee8432993309a7fb48be275c2a98" Feb 17 13:22:45 crc kubenswrapper[4955]: I0217 13:22:45.583151 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"099f1b679ed3eccbe094cddcb02aacc1522fee8432993309a7fb48be275c2a98"} err="failed to get container status \"099f1b679ed3eccbe094cddcb02aacc1522fee8432993309a7fb48be275c2a98\": rpc error: code = NotFound desc = could not find container \"099f1b679ed3eccbe094cddcb02aacc1522fee8432993309a7fb48be275c2a98\": container with ID starting with 099f1b679ed3eccbe094cddcb02aacc1522fee8432993309a7fb48be275c2a98 not found: ID does not exist" Feb 17 13:22:45 crc kubenswrapper[4955]: I0217 13:22:45.583189 4955 scope.go:117] "RemoveContainer" containerID="34b9043b43dcfcf4e9bd8fe5dd18d0625fa3489ecc3028391af8b117f5445333" Feb 17 13:22:45 crc kubenswrapper[4955]: E0217 13:22:45.585022 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34b9043b43dcfcf4e9bd8fe5dd18d0625fa3489ecc3028391af8b117f5445333\": container with ID starting with 34b9043b43dcfcf4e9bd8fe5dd18d0625fa3489ecc3028391af8b117f5445333 not found: ID does not exist" containerID="34b9043b43dcfcf4e9bd8fe5dd18d0625fa3489ecc3028391af8b117f5445333" Feb 17 13:22:45 crc kubenswrapper[4955]: I0217 13:22:45.585046 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34b9043b43dcfcf4e9bd8fe5dd18d0625fa3489ecc3028391af8b117f5445333"} err="failed to get container status \"34b9043b43dcfcf4e9bd8fe5dd18d0625fa3489ecc3028391af8b117f5445333\": rpc error: code = NotFound desc = could not find container \"34b9043b43dcfcf4e9bd8fe5dd18d0625fa3489ecc3028391af8b117f5445333\": container with ID starting with 34b9043b43dcfcf4e9bd8fe5dd18d0625fa3489ecc3028391af8b117f5445333 not found: ID does not exist" Feb 17 13:22:45 crc kubenswrapper[4955]: I0217 13:22:45.585060 4955 scope.go:117] "RemoveContainer" containerID="099f1b679ed3eccbe094cddcb02aacc1522fee8432993309a7fb48be275c2a98" Feb 17 13:22:45 crc kubenswrapper[4955]: I0217 13:22:45.586930 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"099f1b679ed3eccbe094cddcb02aacc1522fee8432993309a7fb48be275c2a98"} err="failed to get container status \"099f1b679ed3eccbe094cddcb02aacc1522fee8432993309a7fb48be275c2a98\": rpc error: code = NotFound desc = could not find container \"099f1b679ed3eccbe094cddcb02aacc1522fee8432993309a7fb48be275c2a98\": container with ID starting with 099f1b679ed3eccbe094cddcb02aacc1522fee8432993309a7fb48be275c2a98 not found: ID does not exist" Feb 17 13:22:45 crc kubenswrapper[4955]: I0217 13:22:45.586951 4955 scope.go:117] "RemoveContainer" containerID="34b9043b43dcfcf4e9bd8fe5dd18d0625fa3489ecc3028391af8b117f5445333" Feb 17 13:22:45 crc kubenswrapper[4955]: I0217 13:22:45.587262 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34b9043b43dcfcf4e9bd8fe5dd18d0625fa3489ecc3028391af8b117f5445333"} err="failed to get container status \"34b9043b43dcfcf4e9bd8fe5dd18d0625fa3489ecc3028391af8b117f5445333\": rpc error: code = NotFound desc = could not find container \"34b9043b43dcfcf4e9bd8fe5dd18d0625fa3489ecc3028391af8b117f5445333\": container with ID starting with 34b9043b43dcfcf4e9bd8fe5dd18d0625fa3489ecc3028391af8b117f5445333 not found: ID does not exist" Feb 17 13:22:45 crc kubenswrapper[4955]: I0217 13:22:45.844106 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-xxclj" event={"ID":"7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a","Type":"ContainerDied","Data":"4e0642b790a638c92b3b5b870c8743a91d4bac05432f296cf349cd9141d42f23"} Feb 17 13:22:45 crc kubenswrapper[4955]: I0217 13:22:45.844297 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-xxclj" Feb 17 13:22:45 crc kubenswrapper[4955]: I0217 13:22:45.848925 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 17 13:22:45 crc kubenswrapper[4955]: I0217 13:22:45.848963 4955 scope.go:117] "RemoveContainer" containerID="4fcb9bb9cd8261f98d8dc33fd2041813f105d13915a97619409dd56f1dd69a52" Feb 17 13:22:45 crc kubenswrapper[4955]: I0217 13:22:45.864341 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5788cc4448-245bs" event={"ID":"1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf","Type":"ContainerStarted","Data":"937a259e06c16c941790cb4506e60ff90561c28dea54c31c6838afbcee91cd93"} Feb 17 13:22:45 crc kubenswrapper[4955]: E0217 13:22:45.887849 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-w6s4s" podUID="e1c2d0c5-7647-447f-a1a6-b37d8c09310c" Feb 17 13:22:45 crc kubenswrapper[4955]: I0217 13:22:45.926908 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-xxclj"] Feb 17 13:22:45 crc kubenswrapper[4955]: I0217 13:22:45.929279 4955 scope.go:117] "RemoveContainer" containerID="3adb0f9fd4de50ae4f78609af2df357877052a387dab551eb76a7960ec86083e" Feb 17 13:22:45 crc kubenswrapper[4955]: I0217 13:22:45.934033 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-xxclj"] Feb 17 13:22:46 crc kubenswrapper[4955]: I0217 13:22:46.048746 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7fffbcf7cb-wr74b"] Feb 17 13:22:46 crc kubenswrapper[4955]: W0217 13:22:46.062006 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6cb09d2_ad7d_4486_a41d_89023789fd87.slice/crio-00a7dd88ff34d2583d7dcdae613b365f889bfff0b3bcef94d70cf8b91b9427d4 WatchSource:0}: Error finding container 00a7dd88ff34d2583d7dcdae613b365f889bfff0b3bcef94d70cf8b91b9427d4: Status 404 returned error can't find the container with id 00a7dd88ff34d2583d7dcdae613b365f889bfff0b3bcef94d70cf8b91b9427d4 Feb 17 13:22:46 crc kubenswrapper[4955]: I0217 13:22:46.129565 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-sstln"] Feb 17 13:22:46 crc kubenswrapper[4955]: W0217 13:22:46.141255 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod50465113_9c90_4a32_8485_0b976374107a.slice/crio-a28c861a30341d835e4ab72c2b940e34200bbe33289646383214d8602c2d637f WatchSource:0}: Error finding container a28c861a30341d835e4ab72c2b940e34200bbe33289646383214d8602c2d637f: Status 404 returned error can't find the container with id a28c861a30341d835e4ab72c2b940e34200bbe33289646383214d8602c2d637f Feb 17 13:22:46 crc kubenswrapper[4955]: I0217 13:22:46.242588 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a" path="/var/lib/kubelet/pods/7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a/volumes" Feb 17 13:22:46 crc kubenswrapper[4955]: I0217 13:22:46.243320 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e" path="/var/lib/kubelet/pods/fc4bdfa1-7fa1-40db-a5fd-eb1e2efb0f1e/volumes" Feb 17 13:22:46 crc kubenswrapper[4955]: I0217 13:22:46.888389 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7fffbcf7cb-wr74b" event={"ID":"d6cb09d2-ad7d-4486-a41d-89023789fd87","Type":"ContainerStarted","Data":"00a7dd88ff34d2583d7dcdae613b365f889bfff0b3bcef94d70cf8b91b9427d4"} Feb 17 13:22:46 crc kubenswrapper[4955]: I0217 13:22:46.897485 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-76cdc74c45-lss64" event={"ID":"dd349c9d-f6b6-48b5-a22f-e2c1d81f6711","Type":"ContainerStarted","Data":"2bd1a1e5ef749430ee242be08db7aa848aff639e569b11a60f2d47e7252bc145"} Feb 17 13:22:46 crc kubenswrapper[4955]: I0217 13:22:46.905578 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5788cc4448-245bs" event={"ID":"1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf","Type":"ContainerStarted","Data":"52cdaba09384ae938ed1a6b17e232b9095c1468ccdd7386221ad6a01d6be8ff4"} Feb 17 13:22:46 crc kubenswrapper[4955]: I0217 13:22:46.905640 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5788cc4448-245bs" event={"ID":"1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf","Type":"ContainerStarted","Data":"32e18fff4a97bffae1a71c95b67ec8d234531f3a2c24d149ec79a131c6c4a75a"} Feb 17 13:22:46 crc kubenswrapper[4955]: I0217 13:22:46.933412 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5788cc4448-245bs" podStartSLOduration=23.933394323 podStartE2EDuration="23.933394323s" podCreationTimestamp="2026-02-17 13:22:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:22:46.932052725 +0000 UTC m=+1105.454782288" watchObservedRunningTime="2026-02-17 13:22:46.933394323 +0000 UTC m=+1105.456123876" Feb 17 13:22:46 crc kubenswrapper[4955]: I0217 13:22:46.947454 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6cf59496f7-6ntrd" event={"ID":"d1b9c8f8-6829-4eaf-b5bf-48121eaaf818","Type":"ContainerStarted","Data":"3f4ac0f91d02f64d61272851b2a1ac47fab311746345b16f0c2110dbf11b51b0"} Feb 17 13:22:46 crc kubenswrapper[4955]: I0217 13:22:46.947510 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6cf59496f7-6ntrd" event={"ID":"d1b9c8f8-6829-4eaf-b5bf-48121eaaf818","Type":"ContainerStarted","Data":"9fc523848c08f25598fae2f5e3fad537b2578bef89b29e3368d51587809c9183"} Feb 17 13:22:46 crc kubenswrapper[4955]: I0217 13:22:46.947669 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6cf59496f7-6ntrd" podUID="d1b9c8f8-6829-4eaf-b5bf-48121eaaf818" containerName="horizon-log" containerID="cri-o://9fc523848c08f25598fae2f5e3fad537b2578bef89b29e3368d51587809c9183" gracePeriod=30 Feb 17 13:22:46 crc kubenswrapper[4955]: I0217 13:22:46.948317 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6cf59496f7-6ntrd" podUID="d1b9c8f8-6829-4eaf-b5bf-48121eaaf818" containerName="horizon" containerID="cri-o://3f4ac0f91d02f64d61272851b2a1ac47fab311746345b16f0c2110dbf11b51b0" gracePeriod=30 Feb 17 13:22:46 crc kubenswrapper[4955]: I0217 13:22:46.956138 4955 generic.go:334] "Generic (PLEG): container finished" podID="be85ffb8-5128-4903-9f7d-cdd462a44b95" containerID="e80f7012019a470099c47cd6309d9795aca80788cbfc35c46df2b84756b7c03c" exitCode=0 Feb 17 13:22:46 crc kubenswrapper[4955]: I0217 13:22:46.956213 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-48hsq" event={"ID":"be85ffb8-5128-4903-9f7d-cdd462a44b95","Type":"ContainerDied","Data":"e80f7012019a470099c47cd6309d9795aca80788cbfc35c46df2b84756b7c03c"} Feb 17 13:22:46 crc kubenswrapper[4955]: I0217 13:22:46.958951 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sstln" event={"ID":"50465113-9c90-4a32-8485-0b976374107a","Type":"ContainerStarted","Data":"a28c861a30341d835e4ab72c2b940e34200bbe33289646383214d8602c2d637f"} Feb 17 13:22:46 crc kubenswrapper[4955]: I0217 13:22:46.962041 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f7a87a1f-b8cc-442c-be49-3117c5fa1702","Type":"ContainerStarted","Data":"44dd02ceec20b585af6b7888400614ce1eb3c3f8a19de650b1b77ee6a9b9ea4f"} Feb 17 13:22:46 crc kubenswrapper[4955]: I0217 13:22:46.963504 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01547c5a-5ca7-435b-ba75-52e075800cf3","Type":"ContainerStarted","Data":"6773574f1d1d99cdf57467650dff95afe5982095ed615f3e81c7756f97852f50"} Feb 17 13:22:46 crc kubenswrapper[4955]: I0217 13:22:46.995493 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nfdxn" event={"ID":"700f73a1-f56d-4d4e-9822-438472980124","Type":"ContainerStarted","Data":"d3b62bc3199101ee1ff8f5e21dd262a136664fda289119da42be6248400fa0c1"} Feb 17 13:22:47 crc kubenswrapper[4955]: I0217 13:22:47.013478 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6cf59496f7-6ntrd" podStartSLOduration=3.79331136 podStartE2EDuration="33.013446597s" podCreationTimestamp="2026-02-17 13:22:14 +0000 UTC" firstStartedPulling="2026-02-17 13:22:16.340602942 +0000 UTC m=+1074.863332485" lastFinishedPulling="2026-02-17 13:22:45.560738179 +0000 UTC m=+1104.083467722" observedRunningTime="2026-02-17 13:22:46.969842503 +0000 UTC m=+1105.492572066" watchObservedRunningTime="2026-02-17 13:22:47.013446597 +0000 UTC m=+1105.536176140" Feb 17 13:22:47 crc kubenswrapper[4955]: I0217 13:22:47.045324 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-sstln" podStartSLOduration=3.045303575 podStartE2EDuration="3.045303575s" podCreationTimestamp="2026-02-17 13:22:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:22:47.00726002 +0000 UTC m=+1105.529989563" watchObservedRunningTime="2026-02-17 13:22:47.045303575 +0000 UTC m=+1105.568033118" Feb 17 13:22:47 crc kubenswrapper[4955]: I0217 13:22:47.066608 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-nfdxn" podStartSLOduration=4.517791676 podStartE2EDuration="33.066589063s" podCreationTimestamp="2026-02-17 13:22:14 +0000 UTC" firstStartedPulling="2026-02-17 13:22:16.403943749 +0000 UTC m=+1074.926673282" lastFinishedPulling="2026-02-17 13:22:44.952741126 +0000 UTC m=+1103.475470669" observedRunningTime="2026-02-17 13:22:47.056519165 +0000 UTC m=+1105.579248708" watchObservedRunningTime="2026-02-17 13:22:47.066589063 +0000 UTC m=+1105.589318606" Feb 17 13:22:48 crc kubenswrapper[4955]: I0217 13:22:48.011906 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f7a87a1f-b8cc-442c-be49-3117c5fa1702","Type":"ContainerStarted","Data":"eab65bbb894bdbaff270eba43ab03ab3b5906adf6217d1f6f4cc38b9548f0409"} Feb 17 13:22:48 crc kubenswrapper[4955]: I0217 13:22:48.012266 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f7a87a1f-b8cc-442c-be49-3117c5fa1702","Type":"ContainerStarted","Data":"f701df1e2a1c187537cf765f73483b67e4d2c702e82909a34de83798851616ed"} Feb 17 13:22:48 crc kubenswrapper[4955]: I0217 13:22:48.011996 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="f7a87a1f-b8cc-442c-be49-3117c5fa1702" containerName="glance-log" containerID="cri-o://eab65bbb894bdbaff270eba43ab03ab3b5906adf6217d1f6f4cc38b9548f0409" gracePeriod=30 Feb 17 13:22:48 crc kubenswrapper[4955]: I0217 13:22:48.012399 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="f7a87a1f-b8cc-442c-be49-3117c5fa1702" containerName="glance-httpd" containerID="cri-o://f701df1e2a1c187537cf765f73483b67e4d2c702e82909a34de83798851616ed" gracePeriod=30 Feb 17 13:22:48 crc kubenswrapper[4955]: I0217 13:22:48.020135 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-76cdc74c45-lss64" event={"ID":"dd349c9d-f6b6-48b5-a22f-e2c1d81f6711","Type":"ContainerStarted","Data":"be4213dde918fe6799bdc14d9565d677ae84d633e730632a7b5c4918dff901ca"} Feb 17 13:22:48 crc kubenswrapper[4955]: I0217 13:22:48.020625 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-76cdc74c45-lss64" podUID="dd349c9d-f6b6-48b5-a22f-e2c1d81f6711" containerName="horizon-log" containerID="cri-o://2bd1a1e5ef749430ee242be08db7aa848aff639e569b11a60f2d47e7252bc145" gracePeriod=30 Feb 17 13:22:48 crc kubenswrapper[4955]: I0217 13:22:48.020738 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-76cdc74c45-lss64" podUID="dd349c9d-f6b6-48b5-a22f-e2c1d81f6711" containerName="horizon" containerID="cri-o://be4213dde918fe6799bdc14d9565d677ae84d633e730632a7b5c4918dff901ca" gracePeriod=30 Feb 17 13:22:48 crc kubenswrapper[4955]: I0217 13:22:48.028003 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9aac5b59-a6ed-43cc-888d-a583384f1ede","Type":"ContainerStarted","Data":"df3dd502c352e95f89bf2961a4a7a9f0013aa610f64820fd42fa4ee03ab9a39a"} Feb 17 13:22:48 crc kubenswrapper[4955]: I0217 13:22:48.028041 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9aac5b59-a6ed-43cc-888d-a583384f1ede","Type":"ContainerStarted","Data":"b2401de1a0e5e2b73a6f7477690436326790cf89aff6a488e201041265f3f6b7"} Feb 17 13:22:48 crc kubenswrapper[4955]: I0217 13:22:48.028174 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="9aac5b59-a6ed-43cc-888d-a583384f1ede" containerName="glance-log" containerID="cri-o://df3dd502c352e95f89bf2961a4a7a9f0013aa610f64820fd42fa4ee03ab9a39a" gracePeriod=30 Feb 17 13:22:48 crc kubenswrapper[4955]: I0217 13:22:48.028279 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="9aac5b59-a6ed-43cc-888d-a583384f1ede" containerName="glance-httpd" containerID="cri-o://b2401de1a0e5e2b73a6f7477690436326790cf89aff6a488e201041265f3f6b7" gracePeriod=30 Feb 17 13:22:48 crc kubenswrapper[4955]: I0217 13:22:48.034519 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sstln" event={"ID":"50465113-9c90-4a32-8485-0b976374107a","Type":"ContainerStarted","Data":"cd6b303add2963b8be95890e398356f20f61c8a63f52dd9ccbe7e3a9586c3c6d"} Feb 17 13:22:48 crc kubenswrapper[4955]: I0217 13:22:48.040200 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7fffbcf7cb-wr74b" event={"ID":"d6cb09d2-ad7d-4486-a41d-89023789fd87","Type":"ContainerStarted","Data":"4f73bcd0f66cb72a557180bee4a7929fb6b83ea1698c00ec0c790c79e87470c5"} Feb 17 13:22:48 crc kubenswrapper[4955]: I0217 13:22:48.040263 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7fffbcf7cb-wr74b" event={"ID":"d6cb09d2-ad7d-4486-a41d-89023789fd87","Type":"ContainerStarted","Data":"ff87aa06377dd9a2d46ca2479a2c68e672ef4b038729f8cd38e4c6a83b5f9279"} Feb 17 13:22:48 crc kubenswrapper[4955]: I0217 13:22:48.058807 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=27.058766433 podStartE2EDuration="27.058766433s" podCreationTimestamp="2026-02-17 13:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:22:48.040042559 +0000 UTC m=+1106.562772102" watchObservedRunningTime="2026-02-17 13:22:48.058766433 +0000 UTC m=+1106.581495976" Feb 17 13:22:48 crc kubenswrapper[4955]: I0217 13:22:48.089912 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=28.089889631 podStartE2EDuration="28.089889631s" podCreationTimestamp="2026-02-17 13:22:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:22:48.060177114 +0000 UTC m=+1106.582906657" watchObservedRunningTime="2026-02-17 13:22:48.089889631 +0000 UTC m=+1106.612619194" Feb 17 13:22:48 crc kubenswrapper[4955]: I0217 13:22:48.117823 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-76cdc74c45-lss64" podStartSLOduration=5.429926583 podStartE2EDuration="34.117786127s" podCreationTimestamp="2026-02-17 13:22:14 +0000 UTC" firstStartedPulling="2026-02-17 13:22:16.289432732 +0000 UTC m=+1074.812162275" lastFinishedPulling="2026-02-17 13:22:44.977292276 +0000 UTC m=+1103.500021819" observedRunningTime="2026-02-17 13:22:48.086805023 +0000 UTC m=+1106.609534576" watchObservedRunningTime="2026-02-17 13:22:48.117786127 +0000 UTC m=+1106.640515670" Feb 17 13:22:48 crc kubenswrapper[4955]: I0217 13:22:48.132308 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7fffbcf7cb-wr74b" podStartSLOduration=25.13228557 podStartE2EDuration="25.13228557s" podCreationTimestamp="2026-02-17 13:22:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:22:48.111299892 +0000 UTC m=+1106.634029425" watchObservedRunningTime="2026-02-17 13:22:48.13228557 +0000 UTC m=+1106.655015113" Feb 17 13:22:48 crc kubenswrapper[4955]: I0217 13:22:48.432453 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-48hsq" Feb 17 13:22:48 crc kubenswrapper[4955]: I0217 13:22:48.499710 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be85ffb8-5128-4903-9f7d-cdd462a44b95-combined-ca-bundle\") pod \"be85ffb8-5128-4903-9f7d-cdd462a44b95\" (UID: \"be85ffb8-5128-4903-9f7d-cdd462a44b95\") " Feb 17 13:22:48 crc kubenswrapper[4955]: I0217 13:22:48.499817 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7qhbv\" (UniqueName: \"kubernetes.io/projected/be85ffb8-5128-4903-9f7d-cdd462a44b95-kube-api-access-7qhbv\") pod \"be85ffb8-5128-4903-9f7d-cdd462a44b95\" (UID: \"be85ffb8-5128-4903-9f7d-cdd462a44b95\") " Feb 17 13:22:48 crc kubenswrapper[4955]: I0217 13:22:48.500018 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/be85ffb8-5128-4903-9f7d-cdd462a44b95-config\") pod \"be85ffb8-5128-4903-9f7d-cdd462a44b95\" (UID: \"be85ffb8-5128-4903-9f7d-cdd462a44b95\") " Feb 17 13:22:48 crc kubenswrapper[4955]: I0217 13:22:48.522128 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be85ffb8-5128-4903-9f7d-cdd462a44b95-kube-api-access-7qhbv" (OuterVolumeSpecName: "kube-api-access-7qhbv") pod "be85ffb8-5128-4903-9f7d-cdd462a44b95" (UID: "be85ffb8-5128-4903-9f7d-cdd462a44b95"). InnerVolumeSpecName "kube-api-access-7qhbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:22:48 crc kubenswrapper[4955]: I0217 13:22:48.528595 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be85ffb8-5128-4903-9f7d-cdd462a44b95-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "be85ffb8-5128-4903-9f7d-cdd462a44b95" (UID: "be85ffb8-5128-4903-9f7d-cdd462a44b95"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:22:48 crc kubenswrapper[4955]: I0217 13:22:48.541571 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be85ffb8-5128-4903-9f7d-cdd462a44b95-config" (OuterVolumeSpecName: "config") pod "be85ffb8-5128-4903-9f7d-cdd462a44b95" (UID: "be85ffb8-5128-4903-9f7d-cdd462a44b95"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:22:48 crc kubenswrapper[4955]: I0217 13:22:48.605293 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be85ffb8-5128-4903-9f7d-cdd462a44b95-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:48 crc kubenswrapper[4955]: I0217 13:22:48.605567 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7qhbv\" (UniqueName: \"kubernetes.io/projected/be85ffb8-5128-4903-9f7d-cdd462a44b95-kube-api-access-7qhbv\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:48 crc kubenswrapper[4955]: I0217 13:22:48.605579 4955 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/be85ffb8-5128-4903-9f7d-cdd462a44b95-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.074757 4955 generic.go:334] "Generic (PLEG): container finished" podID="9aac5b59-a6ed-43cc-888d-a583384f1ede" containerID="b2401de1a0e5e2b73a6f7477690436326790cf89aff6a488e201041265f3f6b7" exitCode=0 Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.074804 4955 generic.go:334] "Generic (PLEG): container finished" podID="9aac5b59-a6ed-43cc-888d-a583384f1ede" containerID="df3dd502c352e95f89bf2961a4a7a9f0013aa610f64820fd42fa4ee03ab9a39a" exitCode=143 Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.074844 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9aac5b59-a6ed-43cc-888d-a583384f1ede","Type":"ContainerDied","Data":"b2401de1a0e5e2b73a6f7477690436326790cf89aff6a488e201041265f3f6b7"} Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.074869 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9aac5b59-a6ed-43cc-888d-a583384f1ede","Type":"ContainerDied","Data":"df3dd502c352e95f89bf2961a4a7a9f0013aa610f64820fd42fa4ee03ab9a39a"} Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.077825 4955 generic.go:334] "Generic (PLEG): container finished" podID="700f73a1-f56d-4d4e-9822-438472980124" containerID="d3b62bc3199101ee1ff8f5e21dd262a136664fda289119da42be6248400fa0c1" exitCode=0 Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.077868 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nfdxn" event={"ID":"700f73a1-f56d-4d4e-9822-438472980124","Type":"ContainerDied","Data":"d3b62bc3199101ee1ff8f5e21dd262a136664fda289119da42be6248400fa0c1"} Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.079575 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-48hsq" event={"ID":"be85ffb8-5128-4903-9f7d-cdd462a44b95","Type":"ContainerDied","Data":"e3f3541a8fde56e369a10bd7c665bd910b9c221d3590924d2410e30e09fd93e5"} Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.079598 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e3f3541a8fde56e369a10bd7c665bd910b9c221d3590924d2410e30e09fd93e5" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.079643 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-48hsq" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.100807 4955 generic.go:334] "Generic (PLEG): container finished" podID="f7a87a1f-b8cc-442c-be49-3117c5fa1702" containerID="f701df1e2a1c187537cf765f73483b67e4d2c702e82909a34de83798851616ed" exitCode=0 Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.100836 4955 generic.go:334] "Generic (PLEG): container finished" podID="f7a87a1f-b8cc-442c-be49-3117c5fa1702" containerID="eab65bbb894bdbaff270eba43ab03ab3b5906adf6217d1f6f4cc38b9548f0409" exitCode=143 Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.101477 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f7a87a1f-b8cc-442c-be49-3117c5fa1702","Type":"ContainerDied","Data":"f701df1e2a1c187537cf765f73483b67e4d2c702e82909a34de83798851616ed"} Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.101514 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f7a87a1f-b8cc-442c-be49-3117c5fa1702","Type":"ContainerDied","Data":"eab65bbb894bdbaff270eba43ab03ab3b5906adf6217d1f6f4cc38b9548f0409"} Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.187597 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-hkwx7"] Feb 17 13:22:49 crc kubenswrapper[4955]: E0217 13:22:49.187961 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be85ffb8-5128-4903-9f7d-cdd462a44b95" containerName="neutron-db-sync" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.187988 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="be85ffb8-5128-4903-9f7d-cdd462a44b95" containerName="neutron-db-sync" Feb 17 13:22:49 crc kubenswrapper[4955]: E0217 13:22:49.188005 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a" containerName="dnsmasq-dns" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.188011 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a" containerName="dnsmasq-dns" Feb 17 13:22:49 crc kubenswrapper[4955]: E0217 13:22:49.188029 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a" containerName="init" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.188035 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a" containerName="init" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.188188 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a" containerName="dnsmasq-dns" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.188225 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="be85ffb8-5128-4903-9f7d-cdd462a44b95" containerName="neutron-db-sync" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.189132 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-hkwx7" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.213680 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-hkwx7"] Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.324770 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-746fcf4988-4q7tw"] Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.327275 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-746fcf4988-4q7tw" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.327883 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cca48ad8-a612-4d37-b39d-116707dbef56-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-hkwx7\" (UID: \"cca48ad8-a612-4d37-b39d-116707dbef56\") " pod="openstack/dnsmasq-dns-6b7b667979-hkwx7" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.327932 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cca48ad8-a612-4d37-b39d-116707dbef56-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-hkwx7\" (UID: \"cca48ad8-a612-4d37-b39d-116707dbef56\") " pod="openstack/dnsmasq-dns-6b7b667979-hkwx7" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.327989 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kghk\" (UniqueName: \"kubernetes.io/projected/cca48ad8-a612-4d37-b39d-116707dbef56-kube-api-access-8kghk\") pod \"dnsmasq-dns-6b7b667979-hkwx7\" (UID: \"cca48ad8-a612-4d37-b39d-116707dbef56\") " pod="openstack/dnsmasq-dns-6b7b667979-hkwx7" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.328021 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cca48ad8-a612-4d37-b39d-116707dbef56-dns-svc\") pod \"dnsmasq-dns-6b7b667979-hkwx7\" (UID: \"cca48ad8-a612-4d37-b39d-116707dbef56\") " pod="openstack/dnsmasq-dns-6b7b667979-hkwx7" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.328146 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cca48ad8-a612-4d37-b39d-116707dbef56-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-hkwx7\" (UID: \"cca48ad8-a612-4d37-b39d-116707dbef56\") " pod="openstack/dnsmasq-dns-6b7b667979-hkwx7" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.328186 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cca48ad8-a612-4d37-b39d-116707dbef56-config\") pod \"dnsmasq-dns-6b7b667979-hkwx7\" (UID: \"cca48ad8-a612-4d37-b39d-116707dbef56\") " pod="openstack/dnsmasq-dns-6b7b667979-hkwx7" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.330065 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.330333 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-kcgqq" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.330616 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f59b8f679-xxclj" podUID="7bcb9f4f-14cb-4e80-b7e6-b2e0e84dff3a" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.126:5353: i/o timeout" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.330947 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.333734 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.354457 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-746fcf4988-4q7tw"] Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.430889 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cca48ad8-a612-4d37-b39d-116707dbef56-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-hkwx7\" (UID: \"cca48ad8-a612-4d37-b39d-116707dbef56\") " pod="openstack/dnsmasq-dns-6b7b667979-hkwx7" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.430953 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e-httpd-config\") pod \"neutron-746fcf4988-4q7tw\" (UID: \"47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e\") " pod="openstack/neutron-746fcf4988-4q7tw" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.430985 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e-combined-ca-bundle\") pod \"neutron-746fcf4988-4q7tw\" (UID: \"47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e\") " pod="openstack/neutron-746fcf4988-4q7tw" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.431009 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cca48ad8-a612-4d37-b39d-116707dbef56-config\") pod \"dnsmasq-dns-6b7b667979-hkwx7\" (UID: \"cca48ad8-a612-4d37-b39d-116707dbef56\") " pod="openstack/dnsmasq-dns-6b7b667979-hkwx7" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.431037 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cca48ad8-a612-4d37-b39d-116707dbef56-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-hkwx7\" (UID: \"cca48ad8-a612-4d37-b39d-116707dbef56\") " pod="openstack/dnsmasq-dns-6b7b667979-hkwx7" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.431087 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cca48ad8-a612-4d37-b39d-116707dbef56-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-hkwx7\" (UID: \"cca48ad8-a612-4d37-b39d-116707dbef56\") " pod="openstack/dnsmasq-dns-6b7b667979-hkwx7" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.431122 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e-config\") pod \"neutron-746fcf4988-4q7tw\" (UID: \"47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e\") " pod="openstack/neutron-746fcf4988-4q7tw" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.431153 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-954vg\" (UniqueName: \"kubernetes.io/projected/47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e-kube-api-access-954vg\") pod \"neutron-746fcf4988-4q7tw\" (UID: \"47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e\") " pod="openstack/neutron-746fcf4988-4q7tw" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.431190 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kghk\" (UniqueName: \"kubernetes.io/projected/cca48ad8-a612-4d37-b39d-116707dbef56-kube-api-access-8kghk\") pod \"dnsmasq-dns-6b7b667979-hkwx7\" (UID: \"cca48ad8-a612-4d37-b39d-116707dbef56\") " pod="openstack/dnsmasq-dns-6b7b667979-hkwx7" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.431235 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cca48ad8-a612-4d37-b39d-116707dbef56-dns-svc\") pod \"dnsmasq-dns-6b7b667979-hkwx7\" (UID: \"cca48ad8-a612-4d37-b39d-116707dbef56\") " pod="openstack/dnsmasq-dns-6b7b667979-hkwx7" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.431299 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e-ovndb-tls-certs\") pod \"neutron-746fcf4988-4q7tw\" (UID: \"47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e\") " pod="openstack/neutron-746fcf4988-4q7tw" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.432350 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cca48ad8-a612-4d37-b39d-116707dbef56-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-hkwx7\" (UID: \"cca48ad8-a612-4d37-b39d-116707dbef56\") " pod="openstack/dnsmasq-dns-6b7b667979-hkwx7" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.432757 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cca48ad8-a612-4d37-b39d-116707dbef56-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-hkwx7\" (UID: \"cca48ad8-a612-4d37-b39d-116707dbef56\") " pod="openstack/dnsmasq-dns-6b7b667979-hkwx7" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.433578 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cca48ad8-a612-4d37-b39d-116707dbef56-dns-svc\") pod \"dnsmasq-dns-6b7b667979-hkwx7\" (UID: \"cca48ad8-a612-4d37-b39d-116707dbef56\") " pod="openstack/dnsmasq-dns-6b7b667979-hkwx7" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.437058 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cca48ad8-a612-4d37-b39d-116707dbef56-config\") pod \"dnsmasq-dns-6b7b667979-hkwx7\" (UID: \"cca48ad8-a612-4d37-b39d-116707dbef56\") " pod="openstack/dnsmasq-dns-6b7b667979-hkwx7" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.438191 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cca48ad8-a612-4d37-b39d-116707dbef56-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-hkwx7\" (UID: \"cca48ad8-a612-4d37-b39d-116707dbef56\") " pod="openstack/dnsmasq-dns-6b7b667979-hkwx7" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.462725 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kghk\" (UniqueName: \"kubernetes.io/projected/cca48ad8-a612-4d37-b39d-116707dbef56-kube-api-access-8kghk\") pod \"dnsmasq-dns-6b7b667979-hkwx7\" (UID: \"cca48ad8-a612-4d37-b39d-116707dbef56\") " pod="openstack/dnsmasq-dns-6b7b667979-hkwx7" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.508465 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-hkwx7" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.535771 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e-ovndb-tls-certs\") pod \"neutron-746fcf4988-4q7tw\" (UID: \"47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e\") " pod="openstack/neutron-746fcf4988-4q7tw" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.535876 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e-httpd-config\") pod \"neutron-746fcf4988-4q7tw\" (UID: \"47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e\") " pod="openstack/neutron-746fcf4988-4q7tw" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.535898 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e-combined-ca-bundle\") pod \"neutron-746fcf4988-4q7tw\" (UID: \"47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e\") " pod="openstack/neutron-746fcf4988-4q7tw" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.535938 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e-config\") pod \"neutron-746fcf4988-4q7tw\" (UID: \"47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e\") " pod="openstack/neutron-746fcf4988-4q7tw" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.535962 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-954vg\" (UniqueName: \"kubernetes.io/projected/47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e-kube-api-access-954vg\") pod \"neutron-746fcf4988-4q7tw\" (UID: \"47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e\") " pod="openstack/neutron-746fcf4988-4q7tw" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.545090 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e-combined-ca-bundle\") pod \"neutron-746fcf4988-4q7tw\" (UID: \"47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e\") " pod="openstack/neutron-746fcf4988-4q7tw" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.548599 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e-httpd-config\") pod \"neutron-746fcf4988-4q7tw\" (UID: \"47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e\") " pod="openstack/neutron-746fcf4988-4q7tw" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.548918 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e-config\") pod \"neutron-746fcf4988-4q7tw\" (UID: \"47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e\") " pod="openstack/neutron-746fcf4988-4q7tw" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.558481 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e-ovndb-tls-certs\") pod \"neutron-746fcf4988-4q7tw\" (UID: \"47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e\") " pod="openstack/neutron-746fcf4988-4q7tw" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.562918 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-954vg\" (UniqueName: \"kubernetes.io/projected/47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e-kube-api-access-954vg\") pod \"neutron-746fcf4988-4q7tw\" (UID: \"47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e\") " pod="openstack/neutron-746fcf4988-4q7tw" Feb 17 13:22:49 crc kubenswrapper[4955]: I0217 13:22:49.654275 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-746fcf4988-4q7tw" Feb 17 13:22:51 crc kubenswrapper[4955]: I0217 13:22:51.022356 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 17 13:22:51 crc kubenswrapper[4955]: I0217 13:22:51.022735 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 17 13:22:51 crc kubenswrapper[4955]: I0217 13:22:51.572649 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-784f78c585-7lvx4"] Feb 17 13:22:51 crc kubenswrapper[4955]: I0217 13:22:51.574389 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-784f78c585-7lvx4" Feb 17 13:22:51 crc kubenswrapper[4955]: I0217 13:22:51.577194 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Feb 17 13:22:51 crc kubenswrapper[4955]: I0217 13:22:51.581537 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Feb 17 13:22:51 crc kubenswrapper[4955]: I0217 13:22:51.598934 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-784f78c585-7lvx4"] Feb 17 13:22:51 crc kubenswrapper[4955]: I0217 13:22:51.714064 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e08d2b73-a720-4a28-bdc0-d8e8008417af-combined-ca-bundle\") pod \"neutron-784f78c585-7lvx4\" (UID: \"e08d2b73-a720-4a28-bdc0-d8e8008417af\") " pod="openstack/neutron-784f78c585-7lvx4" Feb 17 13:22:51 crc kubenswrapper[4955]: I0217 13:22:51.714143 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e08d2b73-a720-4a28-bdc0-d8e8008417af-config\") pod \"neutron-784f78c585-7lvx4\" (UID: \"e08d2b73-a720-4a28-bdc0-d8e8008417af\") " pod="openstack/neutron-784f78c585-7lvx4" Feb 17 13:22:51 crc kubenswrapper[4955]: I0217 13:22:51.714193 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e08d2b73-a720-4a28-bdc0-d8e8008417af-ovndb-tls-certs\") pod \"neutron-784f78c585-7lvx4\" (UID: \"e08d2b73-a720-4a28-bdc0-d8e8008417af\") " pod="openstack/neutron-784f78c585-7lvx4" Feb 17 13:22:51 crc kubenswrapper[4955]: I0217 13:22:51.714310 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e08d2b73-a720-4a28-bdc0-d8e8008417af-public-tls-certs\") pod \"neutron-784f78c585-7lvx4\" (UID: \"e08d2b73-a720-4a28-bdc0-d8e8008417af\") " pod="openstack/neutron-784f78c585-7lvx4" Feb 17 13:22:51 crc kubenswrapper[4955]: I0217 13:22:51.714361 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e08d2b73-a720-4a28-bdc0-d8e8008417af-internal-tls-certs\") pod \"neutron-784f78c585-7lvx4\" (UID: \"e08d2b73-a720-4a28-bdc0-d8e8008417af\") " pod="openstack/neutron-784f78c585-7lvx4" Feb 17 13:22:51 crc kubenswrapper[4955]: I0217 13:22:51.714430 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e08d2b73-a720-4a28-bdc0-d8e8008417af-httpd-config\") pod \"neutron-784f78c585-7lvx4\" (UID: \"e08d2b73-a720-4a28-bdc0-d8e8008417af\") " pod="openstack/neutron-784f78c585-7lvx4" Feb 17 13:22:51 crc kubenswrapper[4955]: I0217 13:22:51.714488 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxpf4\" (UniqueName: \"kubernetes.io/projected/e08d2b73-a720-4a28-bdc0-d8e8008417af-kube-api-access-fxpf4\") pod \"neutron-784f78c585-7lvx4\" (UID: \"e08d2b73-a720-4a28-bdc0-d8e8008417af\") " pod="openstack/neutron-784f78c585-7lvx4" Feb 17 13:22:51 crc kubenswrapper[4955]: I0217 13:22:51.815665 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e08d2b73-a720-4a28-bdc0-d8e8008417af-public-tls-certs\") pod \"neutron-784f78c585-7lvx4\" (UID: \"e08d2b73-a720-4a28-bdc0-d8e8008417af\") " pod="openstack/neutron-784f78c585-7lvx4" Feb 17 13:22:51 crc kubenswrapper[4955]: I0217 13:22:51.815709 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e08d2b73-a720-4a28-bdc0-d8e8008417af-internal-tls-certs\") pod \"neutron-784f78c585-7lvx4\" (UID: \"e08d2b73-a720-4a28-bdc0-d8e8008417af\") " pod="openstack/neutron-784f78c585-7lvx4" Feb 17 13:22:51 crc kubenswrapper[4955]: I0217 13:22:51.815784 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e08d2b73-a720-4a28-bdc0-d8e8008417af-httpd-config\") pod \"neutron-784f78c585-7lvx4\" (UID: \"e08d2b73-a720-4a28-bdc0-d8e8008417af\") " pod="openstack/neutron-784f78c585-7lvx4" Feb 17 13:22:51 crc kubenswrapper[4955]: I0217 13:22:51.815837 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxpf4\" (UniqueName: \"kubernetes.io/projected/e08d2b73-a720-4a28-bdc0-d8e8008417af-kube-api-access-fxpf4\") pod \"neutron-784f78c585-7lvx4\" (UID: \"e08d2b73-a720-4a28-bdc0-d8e8008417af\") " pod="openstack/neutron-784f78c585-7lvx4" Feb 17 13:22:51 crc kubenswrapper[4955]: I0217 13:22:51.815907 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e08d2b73-a720-4a28-bdc0-d8e8008417af-combined-ca-bundle\") pod \"neutron-784f78c585-7lvx4\" (UID: \"e08d2b73-a720-4a28-bdc0-d8e8008417af\") " pod="openstack/neutron-784f78c585-7lvx4" Feb 17 13:22:51 crc kubenswrapper[4955]: I0217 13:22:51.815935 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e08d2b73-a720-4a28-bdc0-d8e8008417af-config\") pod \"neutron-784f78c585-7lvx4\" (UID: \"e08d2b73-a720-4a28-bdc0-d8e8008417af\") " pod="openstack/neutron-784f78c585-7lvx4" Feb 17 13:22:51 crc kubenswrapper[4955]: I0217 13:22:51.815953 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e08d2b73-a720-4a28-bdc0-d8e8008417af-ovndb-tls-certs\") pod \"neutron-784f78c585-7lvx4\" (UID: \"e08d2b73-a720-4a28-bdc0-d8e8008417af\") " pod="openstack/neutron-784f78c585-7lvx4" Feb 17 13:22:51 crc kubenswrapper[4955]: I0217 13:22:51.825481 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e08d2b73-a720-4a28-bdc0-d8e8008417af-public-tls-certs\") pod \"neutron-784f78c585-7lvx4\" (UID: \"e08d2b73-a720-4a28-bdc0-d8e8008417af\") " pod="openstack/neutron-784f78c585-7lvx4" Feb 17 13:22:51 crc kubenswrapper[4955]: I0217 13:22:51.825602 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e08d2b73-a720-4a28-bdc0-d8e8008417af-ovndb-tls-certs\") pod \"neutron-784f78c585-7lvx4\" (UID: \"e08d2b73-a720-4a28-bdc0-d8e8008417af\") " pod="openstack/neutron-784f78c585-7lvx4" Feb 17 13:22:51 crc kubenswrapper[4955]: I0217 13:22:51.827225 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e08d2b73-a720-4a28-bdc0-d8e8008417af-internal-tls-certs\") pod \"neutron-784f78c585-7lvx4\" (UID: \"e08d2b73-a720-4a28-bdc0-d8e8008417af\") " pod="openstack/neutron-784f78c585-7lvx4" Feb 17 13:22:51 crc kubenswrapper[4955]: I0217 13:22:51.828044 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e08d2b73-a720-4a28-bdc0-d8e8008417af-combined-ca-bundle\") pod \"neutron-784f78c585-7lvx4\" (UID: \"e08d2b73-a720-4a28-bdc0-d8e8008417af\") " pod="openstack/neutron-784f78c585-7lvx4" Feb 17 13:22:51 crc kubenswrapper[4955]: I0217 13:22:51.833574 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/e08d2b73-a720-4a28-bdc0-d8e8008417af-config\") pod \"neutron-784f78c585-7lvx4\" (UID: \"e08d2b73-a720-4a28-bdc0-d8e8008417af\") " pod="openstack/neutron-784f78c585-7lvx4" Feb 17 13:22:51 crc kubenswrapper[4955]: I0217 13:22:51.839400 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e08d2b73-a720-4a28-bdc0-d8e8008417af-httpd-config\") pod \"neutron-784f78c585-7lvx4\" (UID: \"e08d2b73-a720-4a28-bdc0-d8e8008417af\") " pod="openstack/neutron-784f78c585-7lvx4" Feb 17 13:22:51 crc kubenswrapper[4955]: I0217 13:22:51.844660 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxpf4\" (UniqueName: \"kubernetes.io/projected/e08d2b73-a720-4a28-bdc0-d8e8008417af-kube-api-access-fxpf4\") pod \"neutron-784f78c585-7lvx4\" (UID: \"e08d2b73-a720-4a28-bdc0-d8e8008417af\") " pod="openstack/neutron-784f78c585-7lvx4" Feb 17 13:22:51 crc kubenswrapper[4955]: I0217 13:22:51.900008 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-784f78c585-7lvx4" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.041599 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.041668 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.136496 4955 generic.go:334] "Generic (PLEG): container finished" podID="50465113-9c90-4a32-8485-0b976374107a" containerID="cd6b303add2963b8be95890e398356f20f61c8a63f52dd9ccbe7e3a9586c3c6d" exitCode=0 Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.136651 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sstln" event={"ID":"50465113-9c90-4a32-8485-0b976374107a","Type":"ContainerDied","Data":"cd6b303add2963b8be95890e398356f20f61c8a63f52dd9ccbe7e3a9586c3c6d"} Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.393452 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.403864 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nfdxn" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.442248 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.532007 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/700f73a1-f56d-4d4e-9822-438472980124-logs\") pod \"700f73a1-f56d-4d4e-9822-438472980124\" (UID: \"700f73a1-f56d-4d4e-9822-438472980124\") " Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.532063 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hpmz2\" (UniqueName: \"kubernetes.io/projected/f7a87a1f-b8cc-442c-be49-3117c5fa1702-kube-api-access-hpmz2\") pod \"f7a87a1f-b8cc-442c-be49-3117c5fa1702\" (UID: \"f7a87a1f-b8cc-442c-be49-3117c5fa1702\") " Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.532126 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f7a87a1f-b8cc-442c-be49-3117c5fa1702-httpd-run\") pod \"f7a87a1f-b8cc-442c-be49-3117c5fa1702\" (UID: \"f7a87a1f-b8cc-442c-be49-3117c5fa1702\") " Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.532152 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7a87a1f-b8cc-442c-be49-3117c5fa1702-combined-ca-bundle\") pod \"f7a87a1f-b8cc-442c-be49-3117c5fa1702\" (UID: \"f7a87a1f-b8cc-442c-be49-3117c5fa1702\") " Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.532194 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/700f73a1-f56d-4d4e-9822-438472980124-combined-ca-bundle\") pod \"700f73a1-f56d-4d4e-9822-438472980124\" (UID: \"700f73a1-f56d-4d4e-9822-438472980124\") " Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.532268 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/700f73a1-f56d-4d4e-9822-438472980124-config-data\") pod \"700f73a1-f56d-4d4e-9822-438472980124\" (UID: \"700f73a1-f56d-4d4e-9822-438472980124\") " Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.532337 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7a87a1f-b8cc-442c-be49-3117c5fa1702-logs\") pod \"f7a87a1f-b8cc-442c-be49-3117c5fa1702\" (UID: \"f7a87a1f-b8cc-442c-be49-3117c5fa1702\") " Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.532360 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7a87a1f-b8cc-442c-be49-3117c5fa1702-scripts\") pod \"f7a87a1f-b8cc-442c-be49-3117c5fa1702\" (UID: \"f7a87a1f-b8cc-442c-be49-3117c5fa1702\") " Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.532405 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7a87a1f-b8cc-442c-be49-3117c5fa1702-public-tls-certs\") pod \"f7a87a1f-b8cc-442c-be49-3117c5fa1702\" (UID: \"f7a87a1f-b8cc-442c-be49-3117c5fa1702\") " Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.532450 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/700f73a1-f56d-4d4e-9822-438472980124-scripts\") pod \"700f73a1-f56d-4d4e-9822-438472980124\" (UID: \"700f73a1-f56d-4d4e-9822-438472980124\") " Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.532500 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4ck2\" (UniqueName: \"kubernetes.io/projected/700f73a1-f56d-4d4e-9822-438472980124-kube-api-access-s4ck2\") pod \"700f73a1-f56d-4d4e-9822-438472980124\" (UID: \"700f73a1-f56d-4d4e-9822-438472980124\") " Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.532545 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7a87a1f-b8cc-442c-be49-3117c5fa1702-config-data\") pod \"f7a87a1f-b8cc-442c-be49-3117c5fa1702\" (UID: \"f7a87a1f-b8cc-442c-be49-3117c5fa1702\") " Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.532573 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"f7a87a1f-b8cc-442c-be49-3117c5fa1702\" (UID: \"f7a87a1f-b8cc-442c-be49-3117c5fa1702\") " Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.533165 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/700f73a1-f56d-4d4e-9822-438472980124-logs" (OuterVolumeSpecName: "logs") pod "700f73a1-f56d-4d4e-9822-438472980124" (UID: "700f73a1-f56d-4d4e-9822-438472980124"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.533363 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7a87a1f-b8cc-442c-be49-3117c5fa1702-logs" (OuterVolumeSpecName: "logs") pod "f7a87a1f-b8cc-442c-be49-3117c5fa1702" (UID: "f7a87a1f-b8cc-442c-be49-3117c5fa1702"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.533979 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7a87a1f-b8cc-442c-be49-3117c5fa1702-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f7a87a1f-b8cc-442c-be49-3117c5fa1702" (UID: "f7a87a1f-b8cc-442c-be49-3117c5fa1702"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.536293 4955 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7a87a1f-b8cc-442c-be49-3117c5fa1702-logs\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.536318 4955 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/700f73a1-f56d-4d4e-9822-438472980124-logs\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.536330 4955 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f7a87a1f-b8cc-442c-be49-3117c5fa1702-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.540078 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7a87a1f-b8cc-442c-be49-3117c5fa1702-kube-api-access-hpmz2" (OuterVolumeSpecName: "kube-api-access-hpmz2") pod "f7a87a1f-b8cc-442c-be49-3117c5fa1702" (UID: "f7a87a1f-b8cc-442c-be49-3117c5fa1702"). InnerVolumeSpecName "kube-api-access-hpmz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.550938 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/700f73a1-f56d-4d4e-9822-438472980124-scripts" (OuterVolumeSpecName: "scripts") pod "700f73a1-f56d-4d4e-9822-438472980124" (UID: "700f73a1-f56d-4d4e-9822-438472980124"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.552186 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7a87a1f-b8cc-442c-be49-3117c5fa1702-scripts" (OuterVolumeSpecName: "scripts") pod "f7a87a1f-b8cc-442c-be49-3117c5fa1702" (UID: "f7a87a1f-b8cc-442c-be49-3117c5fa1702"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.572073 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/700f73a1-f56d-4d4e-9822-438472980124-kube-api-access-s4ck2" (OuterVolumeSpecName: "kube-api-access-s4ck2") pod "700f73a1-f56d-4d4e-9822-438472980124" (UID: "700f73a1-f56d-4d4e-9822-438472980124"). InnerVolumeSpecName "kube-api-access-s4ck2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.574004 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "f7a87a1f-b8cc-442c-be49-3117c5fa1702" (UID: "f7a87a1f-b8cc-442c-be49-3117c5fa1702"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.579720 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7a87a1f-b8cc-442c-be49-3117c5fa1702-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f7a87a1f-b8cc-442c-be49-3117c5fa1702" (UID: "f7a87a1f-b8cc-442c-be49-3117c5fa1702"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.601020 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/700f73a1-f56d-4d4e-9822-438472980124-config-data" (OuterVolumeSpecName: "config-data") pod "700f73a1-f56d-4d4e-9822-438472980124" (UID: "700f73a1-f56d-4d4e-9822-438472980124"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.603261 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7a87a1f-b8cc-442c-be49-3117c5fa1702-config-data" (OuterVolumeSpecName: "config-data") pod "f7a87a1f-b8cc-442c-be49-3117c5fa1702" (UID: "f7a87a1f-b8cc-442c-be49-3117c5fa1702"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.605519 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/700f73a1-f56d-4d4e-9822-438472980124-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "700f73a1-f56d-4d4e-9822-438472980124" (UID: "700f73a1-f56d-4d4e-9822-438472980124"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.611272 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7a87a1f-b8cc-442c-be49-3117c5fa1702-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "f7a87a1f-b8cc-442c-be49-3117c5fa1702" (UID: "f7a87a1f-b8cc-442c-be49-3117c5fa1702"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.639577 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9aac5b59-a6ed-43cc-888d-a583384f1ede-httpd-run\") pod \"9aac5b59-a6ed-43cc-888d-a583384f1ede\" (UID: \"9aac5b59-a6ed-43cc-888d-a583384f1ede\") " Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.639632 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9aac5b59-a6ed-43cc-888d-a583384f1ede-internal-tls-certs\") pod \"9aac5b59-a6ed-43cc-888d-a583384f1ede\" (UID: \"9aac5b59-a6ed-43cc-888d-a583384f1ede\") " Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.639681 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9aac5b59-a6ed-43cc-888d-a583384f1ede-config-data\") pod \"9aac5b59-a6ed-43cc-888d-a583384f1ede\" (UID: \"9aac5b59-a6ed-43cc-888d-a583384f1ede\") " Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.639708 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9aac5b59-a6ed-43cc-888d-a583384f1ede-logs\") pod \"9aac5b59-a6ed-43cc-888d-a583384f1ede\" (UID: \"9aac5b59-a6ed-43cc-888d-a583384f1ede\") " Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.639772 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9aac5b59-a6ed-43cc-888d-a583384f1ede-scripts\") pod \"9aac5b59-a6ed-43cc-888d-a583384f1ede\" (UID: \"9aac5b59-a6ed-43cc-888d-a583384f1ede\") " Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.639813 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9aac5b59-a6ed-43cc-888d-a583384f1ede-combined-ca-bundle\") pod \"9aac5b59-a6ed-43cc-888d-a583384f1ede\" (UID: \"9aac5b59-a6ed-43cc-888d-a583384f1ede\") " Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.639884 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"9aac5b59-a6ed-43cc-888d-a583384f1ede\" (UID: \"9aac5b59-a6ed-43cc-888d-a583384f1ede\") " Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.639939 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mfs67\" (UniqueName: \"kubernetes.io/projected/9aac5b59-a6ed-43cc-888d-a583384f1ede-kube-api-access-mfs67\") pod \"9aac5b59-a6ed-43cc-888d-a583384f1ede\" (UID: \"9aac5b59-a6ed-43cc-888d-a583384f1ede\") " Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.640486 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4ck2\" (UniqueName: \"kubernetes.io/projected/700f73a1-f56d-4d4e-9822-438472980124-kube-api-access-s4ck2\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.640505 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7a87a1f-b8cc-442c-be49-3117c5fa1702-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.640531 4955 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.640543 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hpmz2\" (UniqueName: \"kubernetes.io/projected/f7a87a1f-b8cc-442c-be49-3117c5fa1702-kube-api-access-hpmz2\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.640555 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7a87a1f-b8cc-442c-be49-3117c5fa1702-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.640567 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/700f73a1-f56d-4d4e-9822-438472980124-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.640589 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/700f73a1-f56d-4d4e-9822-438472980124-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.640601 4955 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7a87a1f-b8cc-442c-be49-3117c5fa1702-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.640612 4955 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7a87a1f-b8cc-442c-be49-3117c5fa1702-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.640622 4955 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/700f73a1-f56d-4d4e-9822-438472980124-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.640756 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9aac5b59-a6ed-43cc-888d-a583384f1ede-logs" (OuterVolumeSpecName: "logs") pod "9aac5b59-a6ed-43cc-888d-a583384f1ede" (UID: "9aac5b59-a6ed-43cc-888d-a583384f1ede"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.640990 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9aac5b59-a6ed-43cc-888d-a583384f1ede-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "9aac5b59-a6ed-43cc-888d-a583384f1ede" (UID: "9aac5b59-a6ed-43cc-888d-a583384f1ede"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.646989 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9aac5b59-a6ed-43cc-888d-a583384f1ede-kube-api-access-mfs67" (OuterVolumeSpecName: "kube-api-access-mfs67") pod "9aac5b59-a6ed-43cc-888d-a583384f1ede" (UID: "9aac5b59-a6ed-43cc-888d-a583384f1ede"). InnerVolumeSpecName "kube-api-access-mfs67". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.653452 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9aac5b59-a6ed-43cc-888d-a583384f1ede-scripts" (OuterVolumeSpecName: "scripts") pod "9aac5b59-a6ed-43cc-888d-a583384f1ede" (UID: "9aac5b59-a6ed-43cc-888d-a583384f1ede"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.656109 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "9aac5b59-a6ed-43cc-888d-a583384f1ede" (UID: "9aac5b59-a6ed-43cc-888d-a583384f1ede"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.663404 4955 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.670736 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9aac5b59-a6ed-43cc-888d-a583384f1ede-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9aac5b59-a6ed-43cc-888d-a583384f1ede" (UID: "9aac5b59-a6ed-43cc-888d-a583384f1ede"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.699411 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9aac5b59-a6ed-43cc-888d-a583384f1ede-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "9aac5b59-a6ed-43cc-888d-a583384f1ede" (UID: "9aac5b59-a6ed-43cc-888d-a583384f1ede"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.711055 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9aac5b59-a6ed-43cc-888d-a583384f1ede-config-data" (OuterVolumeSpecName: "config-data") pod "9aac5b59-a6ed-43cc-888d-a583384f1ede" (UID: "9aac5b59-a6ed-43cc-888d-a583384f1ede"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.744269 4955 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9aac5b59-a6ed-43cc-888d-a583384f1ede-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.744305 4955 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9aac5b59-a6ed-43cc-888d-a583384f1ede-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.744331 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9aac5b59-a6ed-43cc-888d-a583384f1ede-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.744345 4955 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9aac5b59-a6ed-43cc-888d-a583384f1ede-logs\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.744355 4955 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9aac5b59-a6ed-43cc-888d-a583384f1ede-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.744366 4955 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.744377 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9aac5b59-a6ed-43cc-888d-a583384f1ede-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.744405 4955 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.744417 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mfs67\" (UniqueName: \"kubernetes.io/projected/9aac5b59-a6ed-43cc-888d-a583384f1ede-kube-api-access-mfs67\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.767750 4955 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.845841 4955 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Feb 17 13:22:52 crc kubenswrapper[4955]: I0217 13:22:52.962724 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-746fcf4988-4q7tw"] Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.146437 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f7a87a1f-b8cc-442c-be49-3117c5fa1702","Type":"ContainerDied","Data":"44dd02ceec20b585af6b7888400614ce1eb3c3f8a19de650b1b77ee6a9b9ea4f"} Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.146849 4955 scope.go:117] "RemoveContainer" containerID="f701df1e2a1c187537cf765f73483b67e4d2c702e82909a34de83798851616ed" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.146907 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.153878 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9aac5b59-a6ed-43cc-888d-a583384f1ede","Type":"ContainerDied","Data":"14485051310b551cede05b8013894ece75fab9d72684e688123b780cefe44c59"} Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.153920 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.156764 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nfdxn" event={"ID":"700f73a1-f56d-4d4e-9822-438472980124","Type":"ContainerDied","Data":"48e5f718063f753992d0e636d3dd74ad4c481301f54925379298b3b86e7630c7"} Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.156835 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="48e5f718063f753992d0e636d3dd74ad4c481301f54925379298b3b86e7630c7" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.156810 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nfdxn" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.214843 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.228318 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.290542 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 17 13:22:53 crc kubenswrapper[4955]: E0217 13:22:53.290997 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aac5b59-a6ed-43cc-888d-a583384f1ede" containerName="glance-httpd" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.291017 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aac5b59-a6ed-43cc-888d-a583384f1ede" containerName="glance-httpd" Feb 17 13:22:53 crc kubenswrapper[4955]: E0217 13:22:53.291031 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="700f73a1-f56d-4d4e-9822-438472980124" containerName="placement-db-sync" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.291037 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="700f73a1-f56d-4d4e-9822-438472980124" containerName="placement-db-sync" Feb 17 13:22:53 crc kubenswrapper[4955]: E0217 13:22:53.291054 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7a87a1f-b8cc-442c-be49-3117c5fa1702" containerName="glance-log" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.291059 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7a87a1f-b8cc-442c-be49-3117c5fa1702" containerName="glance-log" Feb 17 13:22:53 crc kubenswrapper[4955]: E0217 13:22:53.291070 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aac5b59-a6ed-43cc-888d-a583384f1ede" containerName="glance-log" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.291076 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aac5b59-a6ed-43cc-888d-a583384f1ede" containerName="glance-log" Feb 17 13:22:53 crc kubenswrapper[4955]: E0217 13:22:53.291099 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7a87a1f-b8cc-442c-be49-3117c5fa1702" containerName="glance-httpd" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.291106 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7a87a1f-b8cc-442c-be49-3117c5fa1702" containerName="glance-httpd" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.291282 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="9aac5b59-a6ed-43cc-888d-a583384f1ede" containerName="glance-httpd" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.291317 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7a87a1f-b8cc-442c-be49-3117c5fa1702" containerName="glance-log" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.291332 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7a87a1f-b8cc-442c-be49-3117c5fa1702" containerName="glance-httpd" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.291347 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="9aac5b59-a6ed-43cc-888d-a583384f1ede" containerName="glance-log" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.291371 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="700f73a1-f56d-4d4e-9822-438472980124" containerName="placement-db-sync" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.292297 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.296594 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.296984 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.297151 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-d6l2j" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.297286 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.316825 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.332847 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.349860 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.361872 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.363526 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.375193 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.375448 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.424173 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.458290 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"03690a1e-7d74-482e-be94-78c5074595b6\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.458349 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssjqh\" (UniqueName: \"kubernetes.io/projected/03690a1e-7d74-482e-be94-78c5074595b6-kube-api-access-ssjqh\") pod \"glance-default-external-api-0\" (UID: \"03690a1e-7d74-482e-be94-78c5074595b6\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.458378 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03690a1e-7d74-482e-be94-78c5074595b6-scripts\") pod \"glance-default-external-api-0\" (UID: \"03690a1e-7d74-482e-be94-78c5074595b6\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.458402 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/03690a1e-7d74-482e-be94-78c5074595b6-logs\") pod \"glance-default-external-api-0\" (UID: \"03690a1e-7d74-482e-be94-78c5074595b6\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.458432 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/03690a1e-7d74-482e-be94-78c5074595b6-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"03690a1e-7d74-482e-be94-78c5074595b6\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.458455 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03690a1e-7d74-482e-be94-78c5074595b6-config-data\") pod \"glance-default-external-api-0\" (UID: \"03690a1e-7d74-482e-be94-78c5074595b6\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.458502 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03690a1e-7d74-482e-be94-78c5074595b6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"03690a1e-7d74-482e-be94-78c5074595b6\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.458541 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/03690a1e-7d74-482e-be94-78c5074595b6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"03690a1e-7d74-482e-be94-78c5074595b6\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.559970 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/03690a1e-7d74-482e-be94-78c5074595b6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"03690a1e-7d74-482e-be94-78c5074595b6\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.560040 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5da72340-6ea2-42dd-855b-dec4f29f061c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5da72340-6ea2-42dd-855b-dec4f29f061c\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.560072 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"03690a1e-7d74-482e-be94-78c5074595b6\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.560101 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5da72340-6ea2-42dd-855b-dec4f29f061c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5da72340-6ea2-42dd-855b-dec4f29f061c\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.560133 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssjqh\" (UniqueName: \"kubernetes.io/projected/03690a1e-7d74-482e-be94-78c5074595b6-kube-api-access-ssjqh\") pod \"glance-default-external-api-0\" (UID: \"03690a1e-7d74-482e-be94-78c5074595b6\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.560163 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48w4x\" (UniqueName: \"kubernetes.io/projected/5da72340-6ea2-42dd-855b-dec4f29f061c-kube-api-access-48w4x\") pod \"glance-default-internal-api-0\" (UID: \"5da72340-6ea2-42dd-855b-dec4f29f061c\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.560188 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03690a1e-7d74-482e-be94-78c5074595b6-scripts\") pod \"glance-default-external-api-0\" (UID: \"03690a1e-7d74-482e-be94-78c5074595b6\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.560220 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/03690a1e-7d74-482e-be94-78c5074595b6-logs\") pod \"glance-default-external-api-0\" (UID: \"03690a1e-7d74-482e-be94-78c5074595b6\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.560244 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"5da72340-6ea2-42dd-855b-dec4f29f061c\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.560269 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5da72340-6ea2-42dd-855b-dec4f29f061c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5da72340-6ea2-42dd-855b-dec4f29f061c\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.560300 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/03690a1e-7d74-482e-be94-78c5074595b6-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"03690a1e-7d74-482e-be94-78c5074595b6\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.560323 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5da72340-6ea2-42dd-855b-dec4f29f061c-logs\") pod \"glance-default-internal-api-0\" (UID: \"5da72340-6ea2-42dd-855b-dec4f29f061c\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.560352 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03690a1e-7d74-482e-be94-78c5074595b6-config-data\") pod \"glance-default-external-api-0\" (UID: \"03690a1e-7d74-482e-be94-78c5074595b6\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.560397 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5da72340-6ea2-42dd-855b-dec4f29f061c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5da72340-6ea2-42dd-855b-dec4f29f061c\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.560424 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5da72340-6ea2-42dd-855b-dec4f29f061c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5da72340-6ea2-42dd-855b-dec4f29f061c\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.560468 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03690a1e-7d74-482e-be94-78c5074595b6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"03690a1e-7d74-482e-be94-78c5074595b6\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.561378 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/03690a1e-7d74-482e-be94-78c5074595b6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"03690a1e-7d74-482e-be94-78c5074595b6\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.561388 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/03690a1e-7d74-482e-be94-78c5074595b6-logs\") pod \"glance-default-external-api-0\" (UID: \"03690a1e-7d74-482e-be94-78c5074595b6\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.561586 4955 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"03690a1e-7d74-482e-be94-78c5074595b6\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-external-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.578229 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03690a1e-7d74-482e-be94-78c5074595b6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"03690a1e-7d74-482e-be94-78c5074595b6\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.591188 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-845ffbbf5d-ntk67"] Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.593023 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03690a1e-7d74-482e-be94-78c5074595b6-scripts\") pod \"glance-default-external-api-0\" (UID: \"03690a1e-7d74-482e-be94-78c5074595b6\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.599591 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/03690a1e-7d74-482e-be94-78c5074595b6-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"03690a1e-7d74-482e-be94-78c5074595b6\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.600705 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-845ffbbf5d-ntk67" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.607601 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.607836 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-llchf" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.608505 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssjqh\" (UniqueName: \"kubernetes.io/projected/03690a1e-7d74-482e-be94-78c5074595b6-kube-api-access-ssjqh\") pod \"glance-default-external-api-0\" (UID: \"03690a1e-7d74-482e-be94-78c5074595b6\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.608700 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.608936 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.609687 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.610107 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03690a1e-7d74-482e-be94-78c5074595b6-config-data\") pod \"glance-default-external-api-0\" (UID: \"03690a1e-7d74-482e-be94-78c5074595b6\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.612103 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-845ffbbf5d-ntk67"] Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.649877 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5788cc4448-245bs" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.660457 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-5788cc4448-245bs" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.661576 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48w4x\" (UniqueName: \"kubernetes.io/projected/5da72340-6ea2-42dd-855b-dec4f29f061c-kube-api-access-48w4x\") pod \"glance-default-internal-api-0\" (UID: \"5da72340-6ea2-42dd-855b-dec4f29f061c\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.661639 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"5da72340-6ea2-42dd-855b-dec4f29f061c\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.661672 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5da72340-6ea2-42dd-855b-dec4f29f061c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5da72340-6ea2-42dd-855b-dec4f29f061c\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.661714 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5da72340-6ea2-42dd-855b-dec4f29f061c-logs\") pod \"glance-default-internal-api-0\" (UID: \"5da72340-6ea2-42dd-855b-dec4f29f061c\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.661970 4955 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"5da72340-6ea2-42dd-855b-dec4f29f061c\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.664218 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5da72340-6ea2-42dd-855b-dec4f29f061c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5da72340-6ea2-42dd-855b-dec4f29f061c\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.664237 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5da72340-6ea2-42dd-855b-dec4f29f061c-logs\") pod \"glance-default-internal-api-0\" (UID: \"5da72340-6ea2-42dd-855b-dec4f29f061c\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.664280 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5da72340-6ea2-42dd-855b-dec4f29f061c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5da72340-6ea2-42dd-855b-dec4f29f061c\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.665104 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5da72340-6ea2-42dd-855b-dec4f29f061c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5da72340-6ea2-42dd-855b-dec4f29f061c\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.665466 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5da72340-6ea2-42dd-855b-dec4f29f061c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5da72340-6ea2-42dd-855b-dec4f29f061c\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.665971 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5da72340-6ea2-42dd-855b-dec4f29f061c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5da72340-6ea2-42dd-855b-dec4f29f061c\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.669040 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5da72340-6ea2-42dd-855b-dec4f29f061c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5da72340-6ea2-42dd-855b-dec4f29f061c\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.680011 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5da72340-6ea2-42dd-855b-dec4f29f061c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5da72340-6ea2-42dd-855b-dec4f29f061c\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.680570 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5da72340-6ea2-42dd-855b-dec4f29f061c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5da72340-6ea2-42dd-855b-dec4f29f061c\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.681530 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5da72340-6ea2-42dd-855b-dec4f29f061c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5da72340-6ea2-42dd-855b-dec4f29f061c\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.707555 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48w4x\" (UniqueName: \"kubernetes.io/projected/5da72340-6ea2-42dd-855b-dec4f29f061c-kube-api-access-48w4x\") pod \"glance-default-internal-api-0\" (UID: \"5da72340-6ea2-42dd-855b-dec4f29f061c\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.712549 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"5da72340-6ea2-42dd-855b-dec4f29f061c\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.714934 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"03690a1e-7d74-482e-be94-78c5074595b6\") " pod="openstack/glance-default-external-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.740279 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.768012 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20d6bdbe-1dd5-4790-add1-2ea3f3496315-config-data\") pod \"placement-845ffbbf5d-ntk67\" (UID: \"20d6bdbe-1dd5-4790-add1-2ea3f3496315\") " pod="openstack/placement-845ffbbf5d-ntk67" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.768109 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20d6bdbe-1dd5-4790-add1-2ea3f3496315-combined-ca-bundle\") pod \"placement-845ffbbf5d-ntk67\" (UID: \"20d6bdbe-1dd5-4790-add1-2ea3f3496315\") " pod="openstack/placement-845ffbbf5d-ntk67" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.768156 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/20d6bdbe-1dd5-4790-add1-2ea3f3496315-public-tls-certs\") pod \"placement-845ffbbf5d-ntk67\" (UID: \"20d6bdbe-1dd5-4790-add1-2ea3f3496315\") " pod="openstack/placement-845ffbbf5d-ntk67" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.768195 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/20d6bdbe-1dd5-4790-add1-2ea3f3496315-internal-tls-certs\") pod \"placement-845ffbbf5d-ntk67\" (UID: \"20d6bdbe-1dd5-4790-add1-2ea3f3496315\") " pod="openstack/placement-845ffbbf5d-ntk67" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.768220 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20d6bdbe-1dd5-4790-add1-2ea3f3496315-scripts\") pod \"placement-845ffbbf5d-ntk67\" (UID: \"20d6bdbe-1dd5-4790-add1-2ea3f3496315\") " pod="openstack/placement-845ffbbf5d-ntk67" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.768357 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmctr\" (UniqueName: \"kubernetes.io/projected/20d6bdbe-1dd5-4790-add1-2ea3f3496315-kube-api-access-zmctr\") pod \"placement-845ffbbf5d-ntk67\" (UID: \"20d6bdbe-1dd5-4790-add1-2ea3f3496315\") " pod="openstack/placement-845ffbbf5d-ntk67" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.768460 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20d6bdbe-1dd5-4790-add1-2ea3f3496315-logs\") pod \"placement-845ffbbf5d-ntk67\" (UID: \"20d6bdbe-1dd5-4790-add1-2ea3f3496315\") " pod="openstack/placement-845ffbbf5d-ntk67" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.785930 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7fffbcf7cb-wr74b" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.786084 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7fffbcf7cb-wr74b" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.870187 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmctr\" (UniqueName: \"kubernetes.io/projected/20d6bdbe-1dd5-4790-add1-2ea3f3496315-kube-api-access-zmctr\") pod \"placement-845ffbbf5d-ntk67\" (UID: \"20d6bdbe-1dd5-4790-add1-2ea3f3496315\") " pod="openstack/placement-845ffbbf5d-ntk67" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.870276 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20d6bdbe-1dd5-4790-add1-2ea3f3496315-logs\") pod \"placement-845ffbbf5d-ntk67\" (UID: \"20d6bdbe-1dd5-4790-add1-2ea3f3496315\") " pod="openstack/placement-845ffbbf5d-ntk67" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.870349 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20d6bdbe-1dd5-4790-add1-2ea3f3496315-config-data\") pod \"placement-845ffbbf5d-ntk67\" (UID: \"20d6bdbe-1dd5-4790-add1-2ea3f3496315\") " pod="openstack/placement-845ffbbf5d-ntk67" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.870365 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20d6bdbe-1dd5-4790-add1-2ea3f3496315-combined-ca-bundle\") pod \"placement-845ffbbf5d-ntk67\" (UID: \"20d6bdbe-1dd5-4790-add1-2ea3f3496315\") " pod="openstack/placement-845ffbbf5d-ntk67" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.870390 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/20d6bdbe-1dd5-4790-add1-2ea3f3496315-public-tls-certs\") pod \"placement-845ffbbf5d-ntk67\" (UID: \"20d6bdbe-1dd5-4790-add1-2ea3f3496315\") " pod="openstack/placement-845ffbbf5d-ntk67" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.870415 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/20d6bdbe-1dd5-4790-add1-2ea3f3496315-internal-tls-certs\") pod \"placement-845ffbbf5d-ntk67\" (UID: \"20d6bdbe-1dd5-4790-add1-2ea3f3496315\") " pod="openstack/placement-845ffbbf5d-ntk67" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.870431 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20d6bdbe-1dd5-4790-add1-2ea3f3496315-scripts\") pod \"placement-845ffbbf5d-ntk67\" (UID: \"20d6bdbe-1dd5-4790-add1-2ea3f3496315\") " pod="openstack/placement-845ffbbf5d-ntk67" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.870694 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20d6bdbe-1dd5-4790-add1-2ea3f3496315-logs\") pod \"placement-845ffbbf5d-ntk67\" (UID: \"20d6bdbe-1dd5-4790-add1-2ea3f3496315\") " pod="openstack/placement-845ffbbf5d-ntk67" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.875204 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20d6bdbe-1dd5-4790-add1-2ea3f3496315-scripts\") pod \"placement-845ffbbf5d-ntk67\" (UID: \"20d6bdbe-1dd5-4790-add1-2ea3f3496315\") " pod="openstack/placement-845ffbbf5d-ntk67" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.875748 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20d6bdbe-1dd5-4790-add1-2ea3f3496315-config-data\") pod \"placement-845ffbbf5d-ntk67\" (UID: \"20d6bdbe-1dd5-4790-add1-2ea3f3496315\") " pod="openstack/placement-845ffbbf5d-ntk67" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.879480 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/20d6bdbe-1dd5-4790-add1-2ea3f3496315-public-tls-certs\") pod \"placement-845ffbbf5d-ntk67\" (UID: \"20d6bdbe-1dd5-4790-add1-2ea3f3496315\") " pod="openstack/placement-845ffbbf5d-ntk67" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.889324 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/20d6bdbe-1dd5-4790-add1-2ea3f3496315-internal-tls-certs\") pod \"placement-845ffbbf5d-ntk67\" (UID: \"20d6bdbe-1dd5-4790-add1-2ea3f3496315\") " pod="openstack/placement-845ffbbf5d-ntk67" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.889915 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmctr\" (UniqueName: \"kubernetes.io/projected/20d6bdbe-1dd5-4790-add1-2ea3f3496315-kube-api-access-zmctr\") pod \"placement-845ffbbf5d-ntk67\" (UID: \"20d6bdbe-1dd5-4790-add1-2ea3f3496315\") " pod="openstack/placement-845ffbbf5d-ntk67" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.892586 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20d6bdbe-1dd5-4790-add1-2ea3f3496315-combined-ca-bundle\") pod \"placement-845ffbbf5d-ntk67\" (UID: \"20d6bdbe-1dd5-4790-add1-2ea3f3496315\") " pod="openstack/placement-845ffbbf5d-ntk67" Feb 17 13:22:53 crc kubenswrapper[4955]: I0217 13:22:53.927665 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 17 13:22:54 crc kubenswrapper[4955]: I0217 13:22:54.093649 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-845ffbbf5d-ntk67" Feb 17 13:22:54 crc kubenswrapper[4955]: I0217 13:22:54.239596 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9aac5b59-a6ed-43cc-888d-a583384f1ede" path="/var/lib/kubelet/pods/9aac5b59-a6ed-43cc-888d-a583384f1ede/volumes" Feb 17 13:22:54 crc kubenswrapper[4955]: I0217 13:22:54.240856 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7a87a1f-b8cc-442c-be49-3117c5fa1702" path="/var/lib/kubelet/pods/f7a87a1f-b8cc-442c-be49-3117c5fa1702/volumes" Feb 17 13:22:55 crc kubenswrapper[4955]: I0217 13:22:55.084822 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-76cdc74c45-lss64" Feb 17 13:22:55 crc kubenswrapper[4955]: I0217 13:22:55.246884 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6cf59496f7-6ntrd" Feb 17 13:22:59 crc kubenswrapper[4955]: I0217 13:22:59.945976 4955 scope.go:117] "RemoveContainer" containerID="eab65bbb894bdbaff270eba43ab03ab3b5906adf6217d1f6f4cc38b9548f0409" Feb 17 13:23:00 crc kubenswrapper[4955]: I0217 13:23:00.276334 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-746fcf4988-4q7tw" event={"ID":"47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e","Type":"ContainerStarted","Data":"d94ba57dd3ef3b201c37616bab2b39c14a4bdb073ea40f077bf62a736e4f5253"} Feb 17 13:23:00 crc kubenswrapper[4955]: I0217 13:23:00.287312 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sstln" event={"ID":"50465113-9c90-4a32-8485-0b976374107a","Type":"ContainerDied","Data":"a28c861a30341d835e4ab72c2b940e34200bbe33289646383214d8602c2d637f"} Feb 17 13:23:00 crc kubenswrapper[4955]: I0217 13:23:00.287352 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a28c861a30341d835e4ab72c2b940e34200bbe33289646383214d8602c2d637f" Feb 17 13:23:00 crc kubenswrapper[4955]: I0217 13:23:00.376344 4955 scope.go:117] "RemoveContainer" containerID="b2401de1a0e5e2b73a6f7477690436326790cf89aff6a488e201041265f3f6b7" Feb 17 13:23:00 crc kubenswrapper[4955]: I0217 13:23:00.400306 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sstln" Feb 17 13:23:00 crc kubenswrapper[4955]: I0217 13:23:00.450059 4955 scope.go:117] "RemoveContainer" containerID="df3dd502c352e95f89bf2961a4a7a9f0013aa610f64820fd42fa4ee03ab9a39a" Feb 17 13:23:00 crc kubenswrapper[4955]: I0217 13:23:00.491750 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50465113-9c90-4a32-8485-0b976374107a-scripts\") pod \"50465113-9c90-4a32-8485-0b976374107a\" (UID: \"50465113-9c90-4a32-8485-0b976374107a\") " Feb 17 13:23:00 crc kubenswrapper[4955]: I0217 13:23:00.491826 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50465113-9c90-4a32-8485-0b976374107a-combined-ca-bundle\") pod \"50465113-9c90-4a32-8485-0b976374107a\" (UID: \"50465113-9c90-4a32-8485-0b976374107a\") " Feb 17 13:23:00 crc kubenswrapper[4955]: I0217 13:23:00.491860 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/50465113-9c90-4a32-8485-0b976374107a-credential-keys\") pod \"50465113-9c90-4a32-8485-0b976374107a\" (UID: \"50465113-9c90-4a32-8485-0b976374107a\") " Feb 17 13:23:00 crc kubenswrapper[4955]: I0217 13:23:00.491930 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/50465113-9c90-4a32-8485-0b976374107a-fernet-keys\") pod \"50465113-9c90-4a32-8485-0b976374107a\" (UID: \"50465113-9c90-4a32-8485-0b976374107a\") " Feb 17 13:23:00 crc kubenswrapper[4955]: I0217 13:23:00.491973 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xxgd\" (UniqueName: \"kubernetes.io/projected/50465113-9c90-4a32-8485-0b976374107a-kube-api-access-9xxgd\") pod \"50465113-9c90-4a32-8485-0b976374107a\" (UID: \"50465113-9c90-4a32-8485-0b976374107a\") " Feb 17 13:23:00 crc kubenswrapper[4955]: I0217 13:23:00.492069 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50465113-9c90-4a32-8485-0b976374107a-config-data\") pod \"50465113-9c90-4a32-8485-0b976374107a\" (UID: \"50465113-9c90-4a32-8485-0b976374107a\") " Feb 17 13:23:00 crc kubenswrapper[4955]: I0217 13:23:00.505481 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50465113-9c90-4a32-8485-0b976374107a-scripts" (OuterVolumeSpecName: "scripts") pod "50465113-9c90-4a32-8485-0b976374107a" (UID: "50465113-9c90-4a32-8485-0b976374107a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:00 crc kubenswrapper[4955]: I0217 13:23:00.511568 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50465113-9c90-4a32-8485-0b976374107a-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "50465113-9c90-4a32-8485-0b976374107a" (UID: "50465113-9c90-4a32-8485-0b976374107a"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:00 crc kubenswrapper[4955]: I0217 13:23:00.513952 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50465113-9c90-4a32-8485-0b976374107a-kube-api-access-9xxgd" (OuterVolumeSpecName: "kube-api-access-9xxgd") pod "50465113-9c90-4a32-8485-0b976374107a" (UID: "50465113-9c90-4a32-8485-0b976374107a"). InnerVolumeSpecName "kube-api-access-9xxgd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:23:00 crc kubenswrapper[4955]: I0217 13:23:00.516025 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50465113-9c90-4a32-8485-0b976374107a-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "50465113-9c90-4a32-8485-0b976374107a" (UID: "50465113-9c90-4a32-8485-0b976374107a"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:00 crc kubenswrapper[4955]: I0217 13:23:00.545982 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50465113-9c90-4a32-8485-0b976374107a-config-data" (OuterVolumeSpecName: "config-data") pod "50465113-9c90-4a32-8485-0b976374107a" (UID: "50465113-9c90-4a32-8485-0b976374107a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:00 crc kubenswrapper[4955]: I0217 13:23:00.576713 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-784f78c585-7lvx4"] Feb 17 13:23:00 crc kubenswrapper[4955]: I0217 13:23:00.586902 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50465113-9c90-4a32-8485-0b976374107a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "50465113-9c90-4a32-8485-0b976374107a" (UID: "50465113-9c90-4a32-8485-0b976374107a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:00 crc kubenswrapper[4955]: I0217 13:23:00.594396 4955 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50465113-9c90-4a32-8485-0b976374107a-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:00 crc kubenswrapper[4955]: I0217 13:23:00.594419 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50465113-9c90-4a32-8485-0b976374107a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:00 crc kubenswrapper[4955]: I0217 13:23:00.594431 4955 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/50465113-9c90-4a32-8485-0b976374107a-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:00 crc kubenswrapper[4955]: I0217 13:23:00.594440 4955 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/50465113-9c90-4a32-8485-0b976374107a-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:00 crc kubenswrapper[4955]: I0217 13:23:00.594448 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xxgd\" (UniqueName: \"kubernetes.io/projected/50465113-9c90-4a32-8485-0b976374107a-kube-api-access-9xxgd\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:00 crc kubenswrapper[4955]: I0217 13:23:00.594458 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50465113-9c90-4a32-8485-0b976374107a-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:00 crc kubenswrapper[4955]: I0217 13:23:00.657996 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-hkwx7"] Feb 17 13:23:00 crc kubenswrapper[4955]: W0217 13:23:00.712855 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcca48ad8_a612_4d37_b39d_116707dbef56.slice/crio-91cbb2a12ff82366fed197af6c7e2e6f2aca666921be26f3407ebb01be77dc5d WatchSource:0}: Error finding container 91cbb2a12ff82366fed197af6c7e2e6f2aca666921be26f3407ebb01be77dc5d: Status 404 returned error can't find the container with id 91cbb2a12ff82366fed197af6c7e2e6f2aca666921be26f3407ebb01be77dc5d Feb 17 13:23:00 crc kubenswrapper[4955]: I0217 13:23:00.898853 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-845ffbbf5d-ntk67"] Feb 17 13:23:00 crc kubenswrapper[4955]: I0217 13:23:00.907271 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.064977 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.338753 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-845ffbbf5d-ntk67" event={"ID":"20d6bdbe-1dd5-4790-add1-2ea3f3496315","Type":"ContainerStarted","Data":"fdf006e63d380ac6f9c5df66f616039d3417717224751ebcdf4e4435267586bd"} Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.343677 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01547c5a-5ca7-435b-ba75-52e075800cf3","Type":"ContainerStarted","Data":"0576e30010b483c3666500ba4802f893b0075c3a2f67ddc2df4245d7e05db25a"} Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.346918 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"03690a1e-7d74-482e-be94-78c5074595b6","Type":"ContainerStarted","Data":"0eb8f2a6d19208279ef41122b8a09aee30ec45e12c795c772236ca4839d4e9cf"} Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.352405 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-tr2q9" event={"ID":"1d695519-3143-4f53-ab46-5823bc04dc0d","Type":"ContainerStarted","Data":"823a14c06920720c3445935a196f9ef5e1bcf7bda248b2b9fc3cab24a17e709a"} Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.356608 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5da72340-6ea2-42dd-855b-dec4f29f061c","Type":"ContainerStarted","Data":"e1132736f08468db8c3016fd720a91e1991cb8ec1e936471c62231597135f2ee"} Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.358987 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-hkwx7" event={"ID":"cca48ad8-a612-4d37-b39d-116707dbef56","Type":"ContainerStarted","Data":"91cbb2a12ff82366fed197af6c7e2e6f2aca666921be26f3407ebb01be77dc5d"} Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.380122 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-tr2q9" podStartSLOduration=3.385554469 podStartE2EDuration="47.380103398s" podCreationTimestamp="2026-02-17 13:22:14 +0000 UTC" firstStartedPulling="2026-02-17 13:22:16.301388493 +0000 UTC m=+1074.824118036" lastFinishedPulling="2026-02-17 13:23:00.295937432 +0000 UTC m=+1118.818666965" observedRunningTime="2026-02-17 13:23:01.376615998 +0000 UTC m=+1119.899345541" watchObservedRunningTime="2026-02-17 13:23:01.380103398 +0000 UTC m=+1119.902832941" Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.420950 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-746fcf4988-4q7tw" event={"ID":"47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e","Type":"ContainerStarted","Data":"6df53ab637202b08040479088ed897dde6492ba6c8f79fb9bae89a3e98c0cd26"} Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.421051 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-746fcf4988-4q7tw" event={"ID":"47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e","Type":"ContainerStarted","Data":"6d565c62d191904d43404bc26ded506c5b6c75a7f0d89b1ca2702cf83d5ad913"} Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.421070 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-746fcf4988-4q7tw" Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.437247 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-784f78c585-7lvx4" event={"ID":"e08d2b73-a720-4a28-bdc0-d8e8008417af","Type":"ContainerStarted","Data":"4c25f2067734abe7fc2878ff092cd406523ac1cd49c75130c410877ed3c8a5bb"} Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.437292 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-784f78c585-7lvx4" event={"ID":"e08d2b73-a720-4a28-bdc0-d8e8008417af","Type":"ContainerStarted","Data":"efd2421edee17423c62f62becce62b64eca1b74343b0d02d67e8b07d9939a824"} Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.437291 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sstln" Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.453096 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-746fcf4988-4q7tw" podStartSLOduration=12.453075829 podStartE2EDuration="12.453075829s" podCreationTimestamp="2026-02-17 13:22:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:23:01.442314852 +0000 UTC m=+1119.965044395" watchObservedRunningTime="2026-02-17 13:23:01.453075829 +0000 UTC m=+1119.975805372" Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.632503 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-c5b8c48f6-zr5fn"] Feb 17 13:23:01 crc kubenswrapper[4955]: E0217 13:23:01.633226 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50465113-9c90-4a32-8485-0b976374107a" containerName="keystone-bootstrap" Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.633241 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="50465113-9c90-4a32-8485-0b976374107a" containerName="keystone-bootstrap" Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.633416 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="50465113-9c90-4a32-8485-0b976374107a" containerName="keystone-bootstrap" Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.633972 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-c5b8c48f6-zr5fn" Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.639467 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.639730 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.639728 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.640251 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.640556 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.640752 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-qh96v" Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.643253 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-c5b8c48f6-zr5fn"] Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.717868 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75-fernet-keys\") pod \"keystone-c5b8c48f6-zr5fn\" (UID: \"fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75\") " pod="openstack/keystone-c5b8c48f6-zr5fn" Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.718021 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75-scripts\") pod \"keystone-c5b8c48f6-zr5fn\" (UID: \"fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75\") " pod="openstack/keystone-c5b8c48f6-zr5fn" Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.718046 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75-credential-keys\") pod \"keystone-c5b8c48f6-zr5fn\" (UID: \"fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75\") " pod="openstack/keystone-c5b8c48f6-zr5fn" Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.718072 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75-public-tls-certs\") pod \"keystone-c5b8c48f6-zr5fn\" (UID: \"fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75\") " pod="openstack/keystone-c5b8c48f6-zr5fn" Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.718098 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75-config-data\") pod \"keystone-c5b8c48f6-zr5fn\" (UID: \"fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75\") " pod="openstack/keystone-c5b8c48f6-zr5fn" Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.718128 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjhc5\" (UniqueName: \"kubernetes.io/projected/fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75-kube-api-access-gjhc5\") pod \"keystone-c5b8c48f6-zr5fn\" (UID: \"fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75\") " pod="openstack/keystone-c5b8c48f6-zr5fn" Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.718168 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75-combined-ca-bundle\") pod \"keystone-c5b8c48f6-zr5fn\" (UID: \"fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75\") " pod="openstack/keystone-c5b8c48f6-zr5fn" Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.718199 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75-internal-tls-certs\") pod \"keystone-c5b8c48f6-zr5fn\" (UID: \"fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75\") " pod="openstack/keystone-c5b8c48f6-zr5fn" Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.820909 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75-fernet-keys\") pod \"keystone-c5b8c48f6-zr5fn\" (UID: \"fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75\") " pod="openstack/keystone-c5b8c48f6-zr5fn" Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.820995 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75-scripts\") pod \"keystone-c5b8c48f6-zr5fn\" (UID: \"fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75\") " pod="openstack/keystone-c5b8c48f6-zr5fn" Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.821018 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75-credential-keys\") pod \"keystone-c5b8c48f6-zr5fn\" (UID: \"fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75\") " pod="openstack/keystone-c5b8c48f6-zr5fn" Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.821044 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75-public-tls-certs\") pod \"keystone-c5b8c48f6-zr5fn\" (UID: \"fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75\") " pod="openstack/keystone-c5b8c48f6-zr5fn" Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.821073 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75-config-data\") pod \"keystone-c5b8c48f6-zr5fn\" (UID: \"fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75\") " pod="openstack/keystone-c5b8c48f6-zr5fn" Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.821107 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjhc5\" (UniqueName: \"kubernetes.io/projected/fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75-kube-api-access-gjhc5\") pod \"keystone-c5b8c48f6-zr5fn\" (UID: \"fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75\") " pod="openstack/keystone-c5b8c48f6-zr5fn" Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.821146 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75-combined-ca-bundle\") pod \"keystone-c5b8c48f6-zr5fn\" (UID: \"fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75\") " pod="openstack/keystone-c5b8c48f6-zr5fn" Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.821176 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75-internal-tls-certs\") pod \"keystone-c5b8c48f6-zr5fn\" (UID: \"fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75\") " pod="openstack/keystone-c5b8c48f6-zr5fn" Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.858747 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75-config-data\") pod \"keystone-c5b8c48f6-zr5fn\" (UID: \"fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75\") " pod="openstack/keystone-c5b8c48f6-zr5fn" Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.860150 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75-scripts\") pod \"keystone-c5b8c48f6-zr5fn\" (UID: \"fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75\") " pod="openstack/keystone-c5b8c48f6-zr5fn" Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.860779 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75-combined-ca-bundle\") pod \"keystone-c5b8c48f6-zr5fn\" (UID: \"fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75\") " pod="openstack/keystone-c5b8c48f6-zr5fn" Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.867045 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75-public-tls-certs\") pod \"keystone-c5b8c48f6-zr5fn\" (UID: \"fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75\") " pod="openstack/keystone-c5b8c48f6-zr5fn" Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.867199 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75-fernet-keys\") pod \"keystone-c5b8c48f6-zr5fn\" (UID: \"fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75\") " pod="openstack/keystone-c5b8c48f6-zr5fn" Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.868336 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75-internal-tls-certs\") pod \"keystone-c5b8c48f6-zr5fn\" (UID: \"fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75\") " pod="openstack/keystone-c5b8c48f6-zr5fn" Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.868453 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjhc5\" (UniqueName: \"kubernetes.io/projected/fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75-kube-api-access-gjhc5\") pod \"keystone-c5b8c48f6-zr5fn\" (UID: \"fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75\") " pod="openstack/keystone-c5b8c48f6-zr5fn" Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.868560 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75-credential-keys\") pod \"keystone-c5b8c48f6-zr5fn\" (UID: \"fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75\") " pod="openstack/keystone-c5b8c48f6-zr5fn" Feb 17 13:23:01 crc kubenswrapper[4955]: I0217 13:23:01.971250 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-c5b8c48f6-zr5fn" Feb 17 13:23:02 crc kubenswrapper[4955]: I0217 13:23:02.491395 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-845ffbbf5d-ntk67" event={"ID":"20d6bdbe-1dd5-4790-add1-2ea3f3496315","Type":"ContainerStarted","Data":"f550116db1414f335a606a7e1ad151d2ee5761dfeb702452fed5a5ea730d9b78"} Feb 17 13:23:02 crc kubenswrapper[4955]: I0217 13:23:02.491835 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-845ffbbf5d-ntk67" event={"ID":"20d6bdbe-1dd5-4790-add1-2ea3f3496315","Type":"ContainerStarted","Data":"a20011351ac1920632e1b4557ffb5d35fab17f1d712096d2aa96c1115c17a75a"} Feb 17 13:23:02 crc kubenswrapper[4955]: I0217 13:23:02.492894 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-845ffbbf5d-ntk67" Feb 17 13:23:02 crc kubenswrapper[4955]: I0217 13:23:02.492932 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-845ffbbf5d-ntk67" Feb 17 13:23:02 crc kubenswrapper[4955]: I0217 13:23:02.505092 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5da72340-6ea2-42dd-855b-dec4f29f061c","Type":"ContainerStarted","Data":"29b22587a44a1d3f014b0d32e0afc120c7ed49e76d1b13702dd03d797cd374cc"} Feb 17 13:23:02 crc kubenswrapper[4955]: I0217 13:23:02.515987 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-845ffbbf5d-ntk67" podStartSLOduration=9.515949417 podStartE2EDuration="9.515949417s" podCreationTimestamp="2026-02-17 13:22:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:23:02.514274599 +0000 UTC m=+1121.037004152" watchObservedRunningTime="2026-02-17 13:23:02.515949417 +0000 UTC m=+1121.038678950" Feb 17 13:23:02 crc kubenswrapper[4955]: I0217 13:23:02.524377 4955 generic.go:334] "Generic (PLEG): container finished" podID="cca48ad8-a612-4d37-b39d-116707dbef56" containerID="bc35cbd2f0bc63599385fc63a8fe9deeaaa338b17281a5bec3563abf75ade0cd" exitCode=0 Feb 17 13:23:02 crc kubenswrapper[4955]: I0217 13:23:02.524480 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-hkwx7" event={"ID":"cca48ad8-a612-4d37-b39d-116707dbef56","Type":"ContainerDied","Data":"bc35cbd2f0bc63599385fc63a8fe9deeaaa338b17281a5bec3563abf75ade0cd"} Feb 17 13:23:02 crc kubenswrapper[4955]: I0217 13:23:02.557734 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-w6s4s" event={"ID":"e1c2d0c5-7647-447f-a1a6-b37d8c09310c","Type":"ContainerStarted","Data":"1bbb37c0a5d7b5d107d25c7e66b2ba1d7ded8216b4ba2eedae3310ba71cc8d6a"} Feb 17 13:23:02 crc kubenswrapper[4955]: I0217 13:23:02.569212 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"03690a1e-7d74-482e-be94-78c5074595b6","Type":"ContainerStarted","Data":"76e742fa32650a687849fb64674d1d2f47974c7b99026b751786a56009f0ba57"} Feb 17 13:23:02 crc kubenswrapper[4955]: I0217 13:23:02.592619 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-w6s4s" podStartSLOduration=4.57405371 podStartE2EDuration="48.592597153s" podCreationTimestamp="2026-02-17 13:22:14 +0000 UTC" firstStartedPulling="2026-02-17 13:22:16.272971233 +0000 UTC m=+1074.795700776" lastFinishedPulling="2026-02-17 13:23:00.291514676 +0000 UTC m=+1118.814244219" observedRunningTime="2026-02-17 13:23:02.582702341 +0000 UTC m=+1121.105431894" watchObservedRunningTime="2026-02-17 13:23:02.592597153 +0000 UTC m=+1121.115326706" Feb 17 13:23:02 crc kubenswrapper[4955]: I0217 13:23:02.601930 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-784f78c585-7lvx4" event={"ID":"e08d2b73-a720-4a28-bdc0-d8e8008417af","Type":"ContainerStarted","Data":"edfe665cc4bfb8b12efc35a155058e0faa141827cfcb336ebc28576eb4699582"} Feb 17 13:23:02 crc kubenswrapper[4955]: I0217 13:23:02.602008 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-784f78c585-7lvx4" Feb 17 13:23:02 crc kubenswrapper[4955]: I0217 13:23:02.661662 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-784f78c585-7lvx4" podStartSLOduration=11.661645962 podStartE2EDuration="11.661645962s" podCreationTimestamp="2026-02-17 13:22:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:23:02.647555081 +0000 UTC m=+1121.170284644" watchObservedRunningTime="2026-02-17 13:23:02.661645962 +0000 UTC m=+1121.184375505" Feb 17 13:23:02 crc kubenswrapper[4955]: W0217 13:23:02.666417 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfc7f3f32_cb4f_4b66_8ba6_4e35b2555b75.slice/crio-e424f0e0e4602c294d96b3a8c3d4b426431f04dcc5aa91fad6a0813954453b55 WatchSource:0}: Error finding container e424f0e0e4602c294d96b3a8c3d4b426431f04dcc5aa91fad6a0813954453b55: Status 404 returned error can't find the container with id e424f0e0e4602c294d96b3a8c3d4b426431f04dcc5aa91fad6a0813954453b55 Feb 17 13:23:02 crc kubenswrapper[4955]: I0217 13:23:02.673444 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-c5b8c48f6-zr5fn"] Feb 17 13:23:03 crc kubenswrapper[4955]: I0217 13:23:03.627496 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5da72340-6ea2-42dd-855b-dec4f29f061c","Type":"ContainerStarted","Data":"172f55b5031dad85034b5db9a36dfbaf6236f4dcc6a4d44cdd9d2c42d86253b2"} Feb 17 13:23:03 crc kubenswrapper[4955]: I0217 13:23:03.632155 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-hkwx7" event={"ID":"cca48ad8-a612-4d37-b39d-116707dbef56","Type":"ContainerStarted","Data":"70afb0d3b38304267ebdd70a3406513bfd74b19c6d94549763933370586c836c"} Feb 17 13:23:03 crc kubenswrapper[4955]: I0217 13:23:03.632456 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b7b667979-hkwx7" Feb 17 13:23:03 crc kubenswrapper[4955]: I0217 13:23:03.640378 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"03690a1e-7d74-482e-be94-78c5074595b6","Type":"ContainerStarted","Data":"ddb90ae2171d4c56825cc492fcade33f9f4837d699bc2112954ddeddd9d99958"} Feb 17 13:23:03 crc kubenswrapper[4955]: I0217 13:23:03.648973 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=10.648955955 podStartE2EDuration="10.648955955s" podCreationTimestamp="2026-02-17 13:22:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:23:03.646260118 +0000 UTC m=+1122.168989661" watchObservedRunningTime="2026-02-17 13:23:03.648955955 +0000 UTC m=+1122.171685498" Feb 17 13:23:03 crc kubenswrapper[4955]: I0217 13:23:03.655668 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-c5b8c48f6-zr5fn" event={"ID":"fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75","Type":"ContainerStarted","Data":"72708f96683288310f54c8403c4eb47fb4853f5cd595fd97f896f3bce972c3f8"} Feb 17 13:23:03 crc kubenswrapper[4955]: I0217 13:23:03.655807 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-c5b8c48f6-zr5fn" event={"ID":"fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75","Type":"ContainerStarted","Data":"e424f0e0e4602c294d96b3a8c3d4b426431f04dcc5aa91fad6a0813954453b55"} Feb 17 13:23:03 crc kubenswrapper[4955]: I0217 13:23:03.658458 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-c5b8c48f6-zr5fn" Feb 17 13:23:03 crc kubenswrapper[4955]: I0217 13:23:03.658898 4955 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5788cc4448-245bs" podUID="1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Feb 17 13:23:03 crc kubenswrapper[4955]: I0217 13:23:03.677948 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b7b667979-hkwx7" podStartSLOduration=14.677921911 podStartE2EDuration="14.677921911s" podCreationTimestamp="2026-02-17 13:22:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:23:03.67050612 +0000 UTC m=+1122.193235673" watchObservedRunningTime="2026-02-17 13:23:03.677921911 +0000 UTC m=+1122.200651454" Feb 17 13:23:03 crc kubenswrapper[4955]: I0217 13:23:03.739687 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=10.739665943 podStartE2EDuration="10.739665943s" podCreationTimestamp="2026-02-17 13:22:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:23:03.716591154 +0000 UTC m=+1122.239320717" watchObservedRunningTime="2026-02-17 13:23:03.739665943 +0000 UTC m=+1122.262395496" Feb 17 13:23:03 crc kubenswrapper[4955]: I0217 13:23:03.739849 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-c5b8c48f6-zr5fn" podStartSLOduration=2.739842838 podStartE2EDuration="2.739842838s" podCreationTimestamp="2026-02-17 13:23:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:23:03.737264554 +0000 UTC m=+1122.259994107" watchObservedRunningTime="2026-02-17 13:23:03.739842838 +0000 UTC m=+1122.262572391" Feb 17 13:23:03 crc kubenswrapper[4955]: I0217 13:23:03.745990 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 17 13:23:03 crc kubenswrapper[4955]: I0217 13:23:03.746270 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 17 13:23:03 crc kubenswrapper[4955]: I0217 13:23:03.785982 4955 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7fffbcf7cb-wr74b" podUID="d6cb09d2-ad7d-4486-a41d-89023789fd87" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.150:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.150:8443: connect: connection refused" Feb 17 13:23:03 crc kubenswrapper[4955]: I0217 13:23:03.826158 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 17 13:23:03 crc kubenswrapper[4955]: I0217 13:23:03.826317 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 17 13:23:03 crc kubenswrapper[4955]: I0217 13:23:03.928265 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 17 13:23:03 crc kubenswrapper[4955]: I0217 13:23:03.928307 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 17 13:23:03 crc kubenswrapper[4955]: I0217 13:23:03.991480 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 17 13:23:04 crc kubenswrapper[4955]: I0217 13:23:04.009743 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 17 13:23:04 crc kubenswrapper[4955]: I0217 13:23:04.575143 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 13:23:04 crc kubenswrapper[4955]: I0217 13:23:04.575431 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 13:23:04 crc kubenswrapper[4955]: I0217 13:23:04.670916 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 17 13:23:04 crc kubenswrapper[4955]: I0217 13:23:04.670974 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 17 13:23:04 crc kubenswrapper[4955]: I0217 13:23:04.670992 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 17 13:23:04 crc kubenswrapper[4955]: I0217 13:23:04.671003 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 17 13:23:05 crc kubenswrapper[4955]: I0217 13:23:05.681517 4955 generic.go:334] "Generic (PLEG): container finished" podID="1d695519-3143-4f53-ab46-5823bc04dc0d" containerID="823a14c06920720c3445935a196f9ef5e1bcf7bda248b2b9fc3cab24a17e709a" exitCode=0 Feb 17 13:23:05 crc kubenswrapper[4955]: I0217 13:23:05.681913 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-tr2q9" event={"ID":"1d695519-3143-4f53-ab46-5823bc04dc0d","Type":"ContainerDied","Data":"823a14c06920720c3445935a196f9ef5e1bcf7bda248b2b9fc3cab24a17e709a"} Feb 17 13:23:09 crc kubenswrapper[4955]: I0217 13:23:09.520044 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b7b667979-hkwx7" Feb 17 13:23:09 crc kubenswrapper[4955]: I0217 13:23:09.580851 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-k2p8k"] Feb 17 13:23:09 crc kubenswrapper[4955]: I0217 13:23:09.581069 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-56df8fb6b7-k2p8k" podUID="04855a3a-d072-45f0-95d1-fb2f93264a52" containerName="dnsmasq-dns" containerID="cri-o://0627cccc799eb4174c02bd70233cd70409dd7bb8aca5c6794565243148b8f7b4" gracePeriod=10 Feb 17 13:23:09 crc kubenswrapper[4955]: I0217 13:23:09.731876 4955 generic.go:334] "Generic (PLEG): container finished" podID="04855a3a-d072-45f0-95d1-fb2f93264a52" containerID="0627cccc799eb4174c02bd70233cd70409dd7bb8aca5c6794565243148b8f7b4" exitCode=0 Feb 17 13:23:09 crc kubenswrapper[4955]: I0217 13:23:09.731940 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-k2p8k" event={"ID":"04855a3a-d072-45f0-95d1-fb2f93264a52","Type":"ContainerDied","Data":"0627cccc799eb4174c02bd70233cd70409dd7bb8aca5c6794565243148b8f7b4"} Feb 17 13:23:09 crc kubenswrapper[4955]: I0217 13:23:09.733897 4955 generic.go:334] "Generic (PLEG): container finished" podID="e1c2d0c5-7647-447f-a1a6-b37d8c09310c" containerID="1bbb37c0a5d7b5d107d25c7e66b2ba1d7ded8216b4ba2eedae3310ba71cc8d6a" exitCode=0 Feb 17 13:23:09 crc kubenswrapper[4955]: I0217 13:23:09.733940 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-w6s4s" event={"ID":"e1c2d0c5-7647-447f-a1a6-b37d8c09310c","Type":"ContainerDied","Data":"1bbb37c0a5d7b5d107d25c7e66b2ba1d7ded8216b4ba2eedae3310ba71cc8d6a"} Feb 17 13:23:10 crc kubenswrapper[4955]: I0217 13:23:10.526349 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-56df8fb6b7-k2p8k" podUID="04855a3a-d072-45f0-95d1-fb2f93264a52" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.143:5353: connect: connection refused" Feb 17 13:23:11 crc kubenswrapper[4955]: I0217 13:23:11.047244 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-tr2q9" Feb 17 13:23:11 crc kubenswrapper[4955]: I0217 13:23:11.154351 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rl7sq\" (UniqueName: \"kubernetes.io/projected/1d695519-3143-4f53-ab46-5823bc04dc0d-kube-api-access-rl7sq\") pod \"1d695519-3143-4f53-ab46-5823bc04dc0d\" (UID: \"1d695519-3143-4f53-ab46-5823bc04dc0d\") " Feb 17 13:23:11 crc kubenswrapper[4955]: I0217 13:23:11.154576 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1d695519-3143-4f53-ab46-5823bc04dc0d-db-sync-config-data\") pod \"1d695519-3143-4f53-ab46-5823bc04dc0d\" (UID: \"1d695519-3143-4f53-ab46-5823bc04dc0d\") " Feb 17 13:23:11 crc kubenswrapper[4955]: I0217 13:23:11.154655 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d695519-3143-4f53-ab46-5823bc04dc0d-combined-ca-bundle\") pod \"1d695519-3143-4f53-ab46-5823bc04dc0d\" (UID: \"1d695519-3143-4f53-ab46-5823bc04dc0d\") " Feb 17 13:23:11 crc kubenswrapper[4955]: I0217 13:23:11.163621 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d695519-3143-4f53-ab46-5823bc04dc0d-kube-api-access-rl7sq" (OuterVolumeSpecName: "kube-api-access-rl7sq") pod "1d695519-3143-4f53-ab46-5823bc04dc0d" (UID: "1d695519-3143-4f53-ab46-5823bc04dc0d"). InnerVolumeSpecName "kube-api-access-rl7sq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:23:11 crc kubenswrapper[4955]: I0217 13:23:11.167055 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d695519-3143-4f53-ab46-5823bc04dc0d-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "1d695519-3143-4f53-ab46-5823bc04dc0d" (UID: "1d695519-3143-4f53-ab46-5823bc04dc0d"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:11 crc kubenswrapper[4955]: I0217 13:23:11.192072 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d695519-3143-4f53-ab46-5823bc04dc0d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1d695519-3143-4f53-ab46-5823bc04dc0d" (UID: "1d695519-3143-4f53-ab46-5823bc04dc0d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:11 crc kubenswrapper[4955]: I0217 13:23:11.256630 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d695519-3143-4f53-ab46-5823bc04dc0d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:11 crc kubenswrapper[4955]: I0217 13:23:11.256656 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rl7sq\" (UniqueName: \"kubernetes.io/projected/1d695519-3143-4f53-ab46-5823bc04dc0d-kube-api-access-rl7sq\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:11 crc kubenswrapper[4955]: I0217 13:23:11.256667 4955 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1d695519-3143-4f53-ab46-5823bc04dc0d-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:11 crc kubenswrapper[4955]: I0217 13:23:11.752652 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-tr2q9" event={"ID":"1d695519-3143-4f53-ab46-5823bc04dc0d","Type":"ContainerDied","Data":"a22683ecd358c1850e1abe1fd3c5d7a3a150311bc260e6a97b49ad753395900b"} Feb 17 13:23:11 crc kubenswrapper[4955]: I0217 13:23:11.752699 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a22683ecd358c1850e1abe1fd3c5d7a3a150311bc260e6a97b49ad753395900b" Feb 17 13:23:11 crc kubenswrapper[4955]: I0217 13:23:11.752706 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-tr2q9" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.362850 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-6ff67b7c6c-9dcnd"] Feb 17 13:23:12 crc kubenswrapper[4955]: E0217 13:23:12.363461 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d695519-3143-4f53-ab46-5823bc04dc0d" containerName="barbican-db-sync" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.363473 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d695519-3143-4f53-ab46-5823bc04dc0d" containerName="barbican-db-sync" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.363721 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d695519-3143-4f53-ab46-5823bc04dc0d" containerName="barbican-db-sync" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.364593 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6ff67b7c6c-9dcnd" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.369506 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-tttcl" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.369700 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.369957 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.376588 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6ff67b7c6c-9dcnd"] Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.393958 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-7789497c74-pgff5"] Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.395387 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7789497c74-pgff5" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.404143 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.435055 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7789497c74-pgff5"] Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.483367 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-8z8kz"] Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.485205 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-8z8kz" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.494487 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20f0b6cf-c788-45bd-bdb2-cb2e01608a08-combined-ca-bundle\") pod \"barbican-worker-6ff67b7c6c-9dcnd\" (UID: \"20f0b6cf-c788-45bd-bdb2-cb2e01608a08\") " pod="openstack/barbican-worker-6ff67b7c6c-9dcnd" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.494723 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8468e621-9a50-49f7-ba03-996bc8564f43-config-data-custom\") pod \"barbican-keystone-listener-7789497c74-pgff5\" (UID: \"8468e621-9a50-49f7-ba03-996bc8564f43\") " pod="openstack/barbican-keystone-listener-7789497c74-pgff5" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.494802 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20f0b6cf-c788-45bd-bdb2-cb2e01608a08-logs\") pod \"barbican-worker-6ff67b7c6c-9dcnd\" (UID: \"20f0b6cf-c788-45bd-bdb2-cb2e01608a08\") " pod="openstack/barbican-worker-6ff67b7c6c-9dcnd" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.494831 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxbjb\" (UniqueName: \"kubernetes.io/projected/20f0b6cf-c788-45bd-bdb2-cb2e01608a08-kube-api-access-vxbjb\") pod \"barbican-worker-6ff67b7c6c-9dcnd\" (UID: \"20f0b6cf-c788-45bd-bdb2-cb2e01608a08\") " pod="openstack/barbican-worker-6ff67b7c6c-9dcnd" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.494861 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8468e621-9a50-49f7-ba03-996bc8564f43-config-data\") pod \"barbican-keystone-listener-7789497c74-pgff5\" (UID: \"8468e621-9a50-49f7-ba03-996bc8564f43\") " pod="openstack/barbican-keystone-listener-7789497c74-pgff5" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.494914 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/20f0b6cf-c788-45bd-bdb2-cb2e01608a08-config-data-custom\") pod \"barbican-worker-6ff67b7c6c-9dcnd\" (UID: \"20f0b6cf-c788-45bd-bdb2-cb2e01608a08\") " pod="openstack/barbican-worker-6ff67b7c6c-9dcnd" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.494932 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20f0b6cf-c788-45bd-bdb2-cb2e01608a08-config-data\") pod \"barbican-worker-6ff67b7c6c-9dcnd\" (UID: \"20f0b6cf-c788-45bd-bdb2-cb2e01608a08\") " pod="openstack/barbican-worker-6ff67b7c6c-9dcnd" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.494960 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwnzt\" (UniqueName: \"kubernetes.io/projected/8468e621-9a50-49f7-ba03-996bc8564f43-kube-api-access-hwnzt\") pod \"barbican-keystone-listener-7789497c74-pgff5\" (UID: \"8468e621-9a50-49f7-ba03-996bc8564f43\") " pod="openstack/barbican-keystone-listener-7789497c74-pgff5" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.494982 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8468e621-9a50-49f7-ba03-996bc8564f43-combined-ca-bundle\") pod \"barbican-keystone-listener-7789497c74-pgff5\" (UID: \"8468e621-9a50-49f7-ba03-996bc8564f43\") " pod="openstack/barbican-keystone-listener-7789497c74-pgff5" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.494998 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8468e621-9a50-49f7-ba03-996bc8564f43-logs\") pod \"barbican-keystone-listener-7789497c74-pgff5\" (UID: \"8468e621-9a50-49f7-ba03-996bc8564f43\") " pod="openstack/barbican-keystone-listener-7789497c74-pgff5" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.502442 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-8z8kz"] Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.572501 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-65b6bfc874-zgznk"] Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.578070 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-65b6bfc874-zgznk" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.581052 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.587546 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-65b6bfc874-zgznk"] Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.596471 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/20f0b6cf-c788-45bd-bdb2-cb2e01608a08-config-data-custom\") pod \"barbican-worker-6ff67b7c6c-9dcnd\" (UID: \"20f0b6cf-c788-45bd-bdb2-cb2e01608a08\") " pod="openstack/barbican-worker-6ff67b7c6c-9dcnd" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.596510 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20f0b6cf-c788-45bd-bdb2-cb2e01608a08-config-data\") pod \"barbican-worker-6ff67b7c6c-9dcnd\" (UID: \"20f0b6cf-c788-45bd-bdb2-cb2e01608a08\") " pod="openstack/barbican-worker-6ff67b7c6c-9dcnd" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.596534 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2793f42e-26b2-4115-9ac5-9b6c1f503218-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-8z8kz\" (UID: \"2793f42e-26b2-4115-9ac5-9b6c1f503218\") " pod="openstack/dnsmasq-dns-848cf88cfc-8z8kz" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.596571 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwnzt\" (UniqueName: \"kubernetes.io/projected/8468e621-9a50-49f7-ba03-996bc8564f43-kube-api-access-hwnzt\") pod \"barbican-keystone-listener-7789497c74-pgff5\" (UID: \"8468e621-9a50-49f7-ba03-996bc8564f43\") " pod="openstack/barbican-keystone-listener-7789497c74-pgff5" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.596588 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rwv2\" (UniqueName: \"kubernetes.io/projected/2793f42e-26b2-4115-9ac5-9b6c1f503218-kube-api-access-6rwv2\") pod \"dnsmasq-dns-848cf88cfc-8z8kz\" (UID: \"2793f42e-26b2-4115-9ac5-9b6c1f503218\") " pod="openstack/dnsmasq-dns-848cf88cfc-8z8kz" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.596608 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8468e621-9a50-49f7-ba03-996bc8564f43-combined-ca-bundle\") pod \"barbican-keystone-listener-7789497c74-pgff5\" (UID: \"8468e621-9a50-49f7-ba03-996bc8564f43\") " pod="openstack/barbican-keystone-listener-7789497c74-pgff5" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.596627 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8468e621-9a50-49f7-ba03-996bc8564f43-logs\") pod \"barbican-keystone-listener-7789497c74-pgff5\" (UID: \"8468e621-9a50-49f7-ba03-996bc8564f43\") " pod="openstack/barbican-keystone-listener-7789497c74-pgff5" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.596661 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20f0b6cf-c788-45bd-bdb2-cb2e01608a08-combined-ca-bundle\") pod \"barbican-worker-6ff67b7c6c-9dcnd\" (UID: \"20f0b6cf-c788-45bd-bdb2-cb2e01608a08\") " pod="openstack/barbican-worker-6ff67b7c6c-9dcnd" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.596687 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8468e621-9a50-49f7-ba03-996bc8564f43-config-data-custom\") pod \"barbican-keystone-listener-7789497c74-pgff5\" (UID: \"8468e621-9a50-49f7-ba03-996bc8564f43\") " pod="openstack/barbican-keystone-listener-7789497c74-pgff5" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.596711 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2793f42e-26b2-4115-9ac5-9b6c1f503218-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-8z8kz\" (UID: \"2793f42e-26b2-4115-9ac5-9b6c1f503218\") " pod="openstack/dnsmasq-dns-848cf88cfc-8z8kz" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.596754 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2793f42e-26b2-4115-9ac5-9b6c1f503218-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-8z8kz\" (UID: \"2793f42e-26b2-4115-9ac5-9b6c1f503218\") " pod="openstack/dnsmasq-dns-848cf88cfc-8z8kz" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.596772 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20f0b6cf-c788-45bd-bdb2-cb2e01608a08-logs\") pod \"barbican-worker-6ff67b7c6c-9dcnd\" (UID: \"20f0b6cf-c788-45bd-bdb2-cb2e01608a08\") " pod="openstack/barbican-worker-6ff67b7c6c-9dcnd" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.596870 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2793f42e-26b2-4115-9ac5-9b6c1f503218-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-8z8kz\" (UID: \"2793f42e-26b2-4115-9ac5-9b6c1f503218\") " pod="openstack/dnsmasq-dns-848cf88cfc-8z8kz" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.596893 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxbjb\" (UniqueName: \"kubernetes.io/projected/20f0b6cf-c788-45bd-bdb2-cb2e01608a08-kube-api-access-vxbjb\") pod \"barbican-worker-6ff67b7c6c-9dcnd\" (UID: \"20f0b6cf-c788-45bd-bdb2-cb2e01608a08\") " pod="openstack/barbican-worker-6ff67b7c6c-9dcnd" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.596923 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2793f42e-26b2-4115-9ac5-9b6c1f503218-config\") pod \"dnsmasq-dns-848cf88cfc-8z8kz\" (UID: \"2793f42e-26b2-4115-9ac5-9b6c1f503218\") " pod="openstack/dnsmasq-dns-848cf88cfc-8z8kz" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.596938 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8468e621-9a50-49f7-ba03-996bc8564f43-config-data\") pod \"barbican-keystone-listener-7789497c74-pgff5\" (UID: \"8468e621-9a50-49f7-ba03-996bc8564f43\") " pod="openstack/barbican-keystone-listener-7789497c74-pgff5" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.597417 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20f0b6cf-c788-45bd-bdb2-cb2e01608a08-logs\") pod \"barbican-worker-6ff67b7c6c-9dcnd\" (UID: \"20f0b6cf-c788-45bd-bdb2-cb2e01608a08\") " pod="openstack/barbican-worker-6ff67b7c6c-9dcnd" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.598076 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8468e621-9a50-49f7-ba03-996bc8564f43-logs\") pod \"barbican-keystone-listener-7789497c74-pgff5\" (UID: \"8468e621-9a50-49f7-ba03-996bc8564f43\") " pod="openstack/barbican-keystone-listener-7789497c74-pgff5" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.602519 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8468e621-9a50-49f7-ba03-996bc8564f43-combined-ca-bundle\") pod \"barbican-keystone-listener-7789497c74-pgff5\" (UID: \"8468e621-9a50-49f7-ba03-996bc8564f43\") " pod="openstack/barbican-keystone-listener-7789497c74-pgff5" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.602772 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/20f0b6cf-c788-45bd-bdb2-cb2e01608a08-config-data-custom\") pod \"barbican-worker-6ff67b7c6c-9dcnd\" (UID: \"20f0b6cf-c788-45bd-bdb2-cb2e01608a08\") " pod="openstack/barbican-worker-6ff67b7c6c-9dcnd" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.603303 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20f0b6cf-c788-45bd-bdb2-cb2e01608a08-config-data\") pod \"barbican-worker-6ff67b7c6c-9dcnd\" (UID: \"20f0b6cf-c788-45bd-bdb2-cb2e01608a08\") " pod="openstack/barbican-worker-6ff67b7c6c-9dcnd" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.609958 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8468e621-9a50-49f7-ba03-996bc8564f43-config-data-custom\") pod \"barbican-keystone-listener-7789497c74-pgff5\" (UID: \"8468e621-9a50-49f7-ba03-996bc8564f43\") " pod="openstack/barbican-keystone-listener-7789497c74-pgff5" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.620006 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxbjb\" (UniqueName: \"kubernetes.io/projected/20f0b6cf-c788-45bd-bdb2-cb2e01608a08-kube-api-access-vxbjb\") pod \"barbican-worker-6ff67b7c6c-9dcnd\" (UID: \"20f0b6cf-c788-45bd-bdb2-cb2e01608a08\") " pod="openstack/barbican-worker-6ff67b7c6c-9dcnd" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.620557 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20f0b6cf-c788-45bd-bdb2-cb2e01608a08-combined-ca-bundle\") pod \"barbican-worker-6ff67b7c6c-9dcnd\" (UID: \"20f0b6cf-c788-45bd-bdb2-cb2e01608a08\") " pod="openstack/barbican-worker-6ff67b7c6c-9dcnd" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.621043 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwnzt\" (UniqueName: \"kubernetes.io/projected/8468e621-9a50-49f7-ba03-996bc8564f43-kube-api-access-hwnzt\") pod \"barbican-keystone-listener-7789497c74-pgff5\" (UID: \"8468e621-9a50-49f7-ba03-996bc8564f43\") " pod="openstack/barbican-keystone-listener-7789497c74-pgff5" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.621752 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8468e621-9a50-49f7-ba03-996bc8564f43-config-data\") pod \"barbican-keystone-listener-7789497c74-pgff5\" (UID: \"8468e621-9a50-49f7-ba03-996bc8564f43\") " pod="openstack/barbican-keystone-listener-7789497c74-pgff5" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.698707 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2793f42e-26b2-4115-9ac5-9b6c1f503218-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-8z8kz\" (UID: \"2793f42e-26b2-4115-9ac5-9b6c1f503218\") " pod="openstack/dnsmasq-dns-848cf88cfc-8z8kz" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.698771 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rwv2\" (UniqueName: \"kubernetes.io/projected/2793f42e-26b2-4115-9ac5-9b6c1f503218-kube-api-access-6rwv2\") pod \"dnsmasq-dns-848cf88cfc-8z8kz\" (UID: \"2793f42e-26b2-4115-9ac5-9b6c1f503218\") " pod="openstack/dnsmasq-dns-848cf88cfc-8z8kz" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.698866 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2793f42e-26b2-4115-9ac5-9b6c1f503218-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-8z8kz\" (UID: \"2793f42e-26b2-4115-9ac5-9b6c1f503218\") " pod="openstack/dnsmasq-dns-848cf88cfc-8z8kz" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.698902 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/353a4cd6-336b-4545-903e-245c9cea97ca-combined-ca-bundle\") pod \"barbican-api-65b6bfc874-zgznk\" (UID: \"353a4cd6-336b-4545-903e-245c9cea97ca\") " pod="openstack/barbican-api-65b6bfc874-zgznk" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.698938 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/353a4cd6-336b-4545-903e-245c9cea97ca-config-data\") pod \"barbican-api-65b6bfc874-zgznk\" (UID: \"353a4cd6-336b-4545-903e-245c9cea97ca\") " pod="openstack/barbican-api-65b6bfc874-zgznk" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.698976 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/353a4cd6-336b-4545-903e-245c9cea97ca-config-data-custom\") pod \"barbican-api-65b6bfc874-zgznk\" (UID: \"353a4cd6-336b-4545-903e-245c9cea97ca\") " pod="openstack/barbican-api-65b6bfc874-zgznk" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.699003 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/353a4cd6-336b-4545-903e-245c9cea97ca-logs\") pod \"barbican-api-65b6bfc874-zgznk\" (UID: \"353a4cd6-336b-4545-903e-245c9cea97ca\") " pod="openstack/barbican-api-65b6bfc874-zgznk" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.699034 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2793f42e-26b2-4115-9ac5-9b6c1f503218-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-8z8kz\" (UID: \"2793f42e-26b2-4115-9ac5-9b6c1f503218\") " pod="openstack/dnsmasq-dns-848cf88cfc-8z8kz" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.699058 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2793f42e-26b2-4115-9ac5-9b6c1f503218-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-8z8kz\" (UID: \"2793f42e-26b2-4115-9ac5-9b6c1f503218\") " pod="openstack/dnsmasq-dns-848cf88cfc-8z8kz" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.699087 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58mxz\" (UniqueName: \"kubernetes.io/projected/353a4cd6-336b-4545-903e-245c9cea97ca-kube-api-access-58mxz\") pod \"barbican-api-65b6bfc874-zgznk\" (UID: \"353a4cd6-336b-4545-903e-245c9cea97ca\") " pod="openstack/barbican-api-65b6bfc874-zgznk" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.699120 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2793f42e-26b2-4115-9ac5-9b6c1f503218-config\") pod \"dnsmasq-dns-848cf88cfc-8z8kz\" (UID: \"2793f42e-26b2-4115-9ac5-9b6c1f503218\") " pod="openstack/dnsmasq-dns-848cf88cfc-8z8kz" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.700064 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2793f42e-26b2-4115-9ac5-9b6c1f503218-config\") pod \"dnsmasq-dns-848cf88cfc-8z8kz\" (UID: \"2793f42e-26b2-4115-9ac5-9b6c1f503218\") " pod="openstack/dnsmasq-dns-848cf88cfc-8z8kz" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.700703 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2793f42e-26b2-4115-9ac5-9b6c1f503218-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-8z8kz\" (UID: \"2793f42e-26b2-4115-9ac5-9b6c1f503218\") " pod="openstack/dnsmasq-dns-848cf88cfc-8z8kz" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.702723 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2793f42e-26b2-4115-9ac5-9b6c1f503218-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-8z8kz\" (UID: \"2793f42e-26b2-4115-9ac5-9b6c1f503218\") " pod="openstack/dnsmasq-dns-848cf88cfc-8z8kz" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.702817 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2793f42e-26b2-4115-9ac5-9b6c1f503218-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-8z8kz\" (UID: \"2793f42e-26b2-4115-9ac5-9b6c1f503218\") " pod="openstack/dnsmasq-dns-848cf88cfc-8z8kz" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.703154 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2793f42e-26b2-4115-9ac5-9b6c1f503218-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-8z8kz\" (UID: \"2793f42e-26b2-4115-9ac5-9b6c1f503218\") " pod="openstack/dnsmasq-dns-848cf88cfc-8z8kz" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.703895 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6ff67b7c6c-9dcnd" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.718141 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rwv2\" (UniqueName: \"kubernetes.io/projected/2793f42e-26b2-4115-9ac5-9b6c1f503218-kube-api-access-6rwv2\") pod \"dnsmasq-dns-848cf88cfc-8z8kz\" (UID: \"2793f42e-26b2-4115-9ac5-9b6c1f503218\") " pod="openstack/dnsmasq-dns-848cf88cfc-8z8kz" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.741987 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7789497c74-pgff5" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.800825 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/353a4cd6-336b-4545-903e-245c9cea97ca-combined-ca-bundle\") pod \"barbican-api-65b6bfc874-zgznk\" (UID: \"353a4cd6-336b-4545-903e-245c9cea97ca\") " pod="openstack/barbican-api-65b6bfc874-zgznk" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.801414 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/353a4cd6-336b-4545-903e-245c9cea97ca-config-data\") pod \"barbican-api-65b6bfc874-zgznk\" (UID: \"353a4cd6-336b-4545-903e-245c9cea97ca\") " pod="openstack/barbican-api-65b6bfc874-zgznk" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.801463 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/353a4cd6-336b-4545-903e-245c9cea97ca-config-data-custom\") pod \"barbican-api-65b6bfc874-zgznk\" (UID: \"353a4cd6-336b-4545-903e-245c9cea97ca\") " pod="openstack/barbican-api-65b6bfc874-zgznk" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.801535 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/353a4cd6-336b-4545-903e-245c9cea97ca-logs\") pod \"barbican-api-65b6bfc874-zgznk\" (UID: \"353a4cd6-336b-4545-903e-245c9cea97ca\") " pod="openstack/barbican-api-65b6bfc874-zgznk" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.801593 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58mxz\" (UniqueName: \"kubernetes.io/projected/353a4cd6-336b-4545-903e-245c9cea97ca-kube-api-access-58mxz\") pod \"barbican-api-65b6bfc874-zgznk\" (UID: \"353a4cd6-336b-4545-903e-245c9cea97ca\") " pod="openstack/barbican-api-65b6bfc874-zgznk" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.804333 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/353a4cd6-336b-4545-903e-245c9cea97ca-combined-ca-bundle\") pod \"barbican-api-65b6bfc874-zgznk\" (UID: \"353a4cd6-336b-4545-903e-245c9cea97ca\") " pod="openstack/barbican-api-65b6bfc874-zgznk" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.804928 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/353a4cd6-336b-4545-903e-245c9cea97ca-logs\") pod \"barbican-api-65b6bfc874-zgznk\" (UID: \"353a4cd6-336b-4545-903e-245c9cea97ca\") " pod="openstack/barbican-api-65b6bfc874-zgznk" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.818898 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/353a4cd6-336b-4545-903e-245c9cea97ca-config-data-custom\") pod \"barbican-api-65b6bfc874-zgznk\" (UID: \"353a4cd6-336b-4545-903e-245c9cea97ca\") " pod="openstack/barbican-api-65b6bfc874-zgznk" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.821880 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58mxz\" (UniqueName: \"kubernetes.io/projected/353a4cd6-336b-4545-903e-245c9cea97ca-kube-api-access-58mxz\") pod \"barbican-api-65b6bfc874-zgznk\" (UID: \"353a4cd6-336b-4545-903e-245c9cea97ca\") " pod="openstack/barbican-api-65b6bfc874-zgznk" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.823748 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/353a4cd6-336b-4545-903e-245c9cea97ca-config-data\") pod \"barbican-api-65b6bfc874-zgznk\" (UID: \"353a4cd6-336b-4545-903e-245c9cea97ca\") " pod="openstack/barbican-api-65b6bfc874-zgznk" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.839833 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-8z8kz" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.989412 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-65b6bfc874-zgznk" Feb 17 13:23:12 crc kubenswrapper[4955]: I0217 13:23:12.996368 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-w6s4s" Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.110024 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1c2d0c5-7647-447f-a1a6-b37d8c09310c-combined-ca-bundle\") pod \"e1c2d0c5-7647-447f-a1a6-b37d8c09310c\" (UID: \"e1c2d0c5-7647-447f-a1a6-b37d8c09310c\") " Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.110393 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1c2d0c5-7647-447f-a1a6-b37d8c09310c-config-data\") pod \"e1c2d0c5-7647-447f-a1a6-b37d8c09310c\" (UID: \"e1c2d0c5-7647-447f-a1a6-b37d8c09310c\") " Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.110528 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6w2b6\" (UniqueName: \"kubernetes.io/projected/e1c2d0c5-7647-447f-a1a6-b37d8c09310c-kube-api-access-6w2b6\") pod \"e1c2d0c5-7647-447f-a1a6-b37d8c09310c\" (UID: \"e1c2d0c5-7647-447f-a1a6-b37d8c09310c\") " Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.110554 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e1c2d0c5-7647-447f-a1a6-b37d8c09310c-etc-machine-id\") pod \"e1c2d0c5-7647-447f-a1a6-b37d8c09310c\" (UID: \"e1c2d0c5-7647-447f-a1a6-b37d8c09310c\") " Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.110601 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e1c2d0c5-7647-447f-a1a6-b37d8c09310c-db-sync-config-data\") pod \"e1c2d0c5-7647-447f-a1a6-b37d8c09310c\" (UID: \"e1c2d0c5-7647-447f-a1a6-b37d8c09310c\") " Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.110649 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1c2d0c5-7647-447f-a1a6-b37d8c09310c-scripts\") pod \"e1c2d0c5-7647-447f-a1a6-b37d8c09310c\" (UID: \"e1c2d0c5-7647-447f-a1a6-b37d8c09310c\") " Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.111597 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e1c2d0c5-7647-447f-a1a6-b37d8c09310c-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "e1c2d0c5-7647-447f-a1a6-b37d8c09310c" (UID: "e1c2d0c5-7647-447f-a1a6-b37d8c09310c"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.115188 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1c2d0c5-7647-447f-a1a6-b37d8c09310c-kube-api-access-6w2b6" (OuterVolumeSpecName: "kube-api-access-6w2b6") pod "e1c2d0c5-7647-447f-a1a6-b37d8c09310c" (UID: "e1c2d0c5-7647-447f-a1a6-b37d8c09310c"). InnerVolumeSpecName "kube-api-access-6w2b6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.115549 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1c2d0c5-7647-447f-a1a6-b37d8c09310c-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "e1c2d0c5-7647-447f-a1a6-b37d8c09310c" (UID: "e1c2d0c5-7647-447f-a1a6-b37d8c09310c"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.118152 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1c2d0c5-7647-447f-a1a6-b37d8c09310c-scripts" (OuterVolumeSpecName: "scripts") pod "e1c2d0c5-7647-447f-a1a6-b37d8c09310c" (UID: "e1c2d0c5-7647-447f-a1a6-b37d8c09310c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.142761 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1c2d0c5-7647-447f-a1a6-b37d8c09310c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e1c2d0c5-7647-447f-a1a6-b37d8c09310c" (UID: "e1c2d0c5-7647-447f-a1a6-b37d8c09310c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.186828 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1c2d0c5-7647-447f-a1a6-b37d8c09310c-config-data" (OuterVolumeSpecName: "config-data") pod "e1c2d0c5-7647-447f-a1a6-b37d8c09310c" (UID: "e1c2d0c5-7647-447f-a1a6-b37d8c09310c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.213017 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6w2b6\" (UniqueName: \"kubernetes.io/projected/e1c2d0c5-7647-447f-a1a6-b37d8c09310c-kube-api-access-6w2b6\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.213047 4955 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e1c2d0c5-7647-447f-a1a6-b37d8c09310c-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.213056 4955 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e1c2d0c5-7647-447f-a1a6-b37d8c09310c-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.213066 4955 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1c2d0c5-7647-447f-a1a6-b37d8c09310c-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.213077 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1c2d0c5-7647-447f-a1a6-b37d8c09310c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.213086 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1c2d0c5-7647-447f-a1a6-b37d8c09310c-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.309006 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-k2p8k" Feb 17 13:23:13 crc kubenswrapper[4955]: E0217 13:23:13.366928 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="01547c5a-5ca7-435b-ba75-52e075800cf3" Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.416424 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8mp9k\" (UniqueName: \"kubernetes.io/projected/04855a3a-d072-45f0-95d1-fb2f93264a52-kube-api-access-8mp9k\") pod \"04855a3a-d072-45f0-95d1-fb2f93264a52\" (UID: \"04855a3a-d072-45f0-95d1-fb2f93264a52\") " Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.416531 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/04855a3a-d072-45f0-95d1-fb2f93264a52-dns-svc\") pod \"04855a3a-d072-45f0-95d1-fb2f93264a52\" (UID: \"04855a3a-d072-45f0-95d1-fb2f93264a52\") " Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.416641 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/04855a3a-d072-45f0-95d1-fb2f93264a52-ovsdbserver-nb\") pod \"04855a3a-d072-45f0-95d1-fb2f93264a52\" (UID: \"04855a3a-d072-45f0-95d1-fb2f93264a52\") " Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.416660 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/04855a3a-d072-45f0-95d1-fb2f93264a52-dns-swift-storage-0\") pod \"04855a3a-d072-45f0-95d1-fb2f93264a52\" (UID: \"04855a3a-d072-45f0-95d1-fb2f93264a52\") " Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.416713 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04855a3a-d072-45f0-95d1-fb2f93264a52-config\") pod \"04855a3a-d072-45f0-95d1-fb2f93264a52\" (UID: \"04855a3a-d072-45f0-95d1-fb2f93264a52\") " Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.416747 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/04855a3a-d072-45f0-95d1-fb2f93264a52-ovsdbserver-sb\") pod \"04855a3a-d072-45f0-95d1-fb2f93264a52\" (UID: \"04855a3a-d072-45f0-95d1-fb2f93264a52\") " Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.435916 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04855a3a-d072-45f0-95d1-fb2f93264a52-kube-api-access-8mp9k" (OuterVolumeSpecName: "kube-api-access-8mp9k") pod "04855a3a-d072-45f0-95d1-fb2f93264a52" (UID: "04855a3a-d072-45f0-95d1-fb2f93264a52"). InnerVolumeSpecName "kube-api-access-8mp9k". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.518653 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8mp9k\" (UniqueName: \"kubernetes.io/projected/04855a3a-d072-45f0-95d1-fb2f93264a52-kube-api-access-8mp9k\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.519556 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04855a3a-d072-45f0-95d1-fb2f93264a52-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "04855a3a-d072-45f0-95d1-fb2f93264a52" (UID: "04855a3a-d072-45f0-95d1-fb2f93264a52"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.550387 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04855a3a-d072-45f0-95d1-fb2f93264a52-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "04855a3a-d072-45f0-95d1-fb2f93264a52" (UID: "04855a3a-d072-45f0-95d1-fb2f93264a52"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.563441 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04855a3a-d072-45f0-95d1-fb2f93264a52-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "04855a3a-d072-45f0-95d1-fb2f93264a52" (UID: "04855a3a-d072-45f0-95d1-fb2f93264a52"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:23:13 crc kubenswrapper[4955]: E0217 13:23:13.591246 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/04855a3a-d072-45f0-95d1-fb2f93264a52-config podName:04855a3a-d072-45f0-95d1-fb2f93264a52 nodeName:}" failed. No retries permitted until 2026-02-17 13:23:14.091216322 +0000 UTC m=+1132.613945865 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config" (UniqueName: "kubernetes.io/configmap/04855a3a-d072-45f0-95d1-fb2f93264a52-config") pod "04855a3a-d072-45f0-95d1-fb2f93264a52" (UID: "04855a3a-d072-45f0-95d1-fb2f93264a52") : error deleting /var/lib/kubelet/pods/04855a3a-d072-45f0-95d1-fb2f93264a52/volume-subpaths: remove /var/lib/kubelet/pods/04855a3a-d072-45f0-95d1-fb2f93264a52/volume-subpaths: no such file or directory Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.591693 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04855a3a-d072-45f0-95d1-fb2f93264a52-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "04855a3a-d072-45f0-95d1-fb2f93264a52" (UID: "04855a3a-d072-45f0-95d1-fb2f93264a52"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.623841 4955 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/04855a3a-d072-45f0-95d1-fb2f93264a52-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.623873 4955 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/04855a3a-d072-45f0-95d1-fb2f93264a52-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.623884 4955 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/04855a3a-d072-45f0-95d1-fb2f93264a52-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.623892 4955 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/04855a3a-d072-45f0-95d1-fb2f93264a52-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:13 crc kubenswrapper[4955]: W0217 13:23:13.680213 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2793f42e_26b2_4115_9ac5_9b6c1f503218.slice/crio-b68c73e9a5be1fcf6f48f84069a92921bfbd3cbc32ed2d9ec6bddb7ff3426d9d WatchSource:0}: Error finding container b68c73e9a5be1fcf6f48f84069a92921bfbd3cbc32ed2d9ec6bddb7ff3426d9d: Status 404 returned error can't find the container with id b68c73e9a5be1fcf6f48f84069a92921bfbd3cbc32ed2d9ec6bddb7ff3426d9d Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.691547 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-8z8kz"] Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.719220 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6ff67b7c6c-9dcnd"] Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.738483 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7789497c74-pgff5"] Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.776360 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-k2p8k" event={"ID":"04855a3a-d072-45f0-95d1-fb2f93264a52","Type":"ContainerDied","Data":"4626844e4ee4838aff936e347470c7f6aa19dd504201e84862aae8029154230c"} Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.776406 4955 scope.go:117] "RemoveContainer" containerID="0627cccc799eb4174c02bd70233cd70409dd7bb8aca5c6794565243148b8f7b4" Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.776514 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-k2p8k" Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.779277 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7789497c74-pgff5" event={"ID":"8468e621-9a50-49f7-ba03-996bc8564f43","Type":"ContainerStarted","Data":"47a8d5d2c388c966efcdbf61d9933547b98390db2cf313e482e466d8733bc52a"} Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.843221 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-8z8kz" event={"ID":"2793f42e-26b2-4115-9ac5-9b6c1f503218","Type":"ContainerStarted","Data":"b68c73e9a5be1fcf6f48f84069a92921bfbd3cbc32ed2d9ec6bddb7ff3426d9d"} Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.850651 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-w6s4s" event={"ID":"e1c2d0c5-7647-447f-a1a6-b37d8c09310c","Type":"ContainerDied","Data":"92880165e23e247c24ba8d154bdc1ba73ca37f070bbe970c913f1f598aeb8b97"} Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.850697 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92880165e23e247c24ba8d154bdc1ba73ca37f070bbe970c913f1f598aeb8b97" Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.850772 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-w6s4s" Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.855073 4955 scope.go:117] "RemoveContainer" containerID="1ccc5c07865713311e4a6ff4900e7ac30db4dd26f40d38d4b3cfffbbe15bb459" Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.862195 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-65b6bfc874-zgznk"] Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.866991 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01547c5a-5ca7-435b-ba75-52e075800cf3","Type":"ContainerStarted","Data":"198e280433fd484d8cb761a7db110f722d42eff7b1ea7c8cf1dbbffef6877c20"} Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.867194 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="01547c5a-5ca7-435b-ba75-52e075800cf3" containerName="ceilometer-notification-agent" containerID="cri-o://6773574f1d1d99cdf57467650dff95afe5982095ed615f3e81c7756f97852f50" gracePeriod=30 Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.867268 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.867307 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="01547c5a-5ca7-435b-ba75-52e075800cf3" containerName="proxy-httpd" containerID="cri-o://198e280433fd484d8cb761a7db110f722d42eff7b1ea7c8cf1dbbffef6877c20" gracePeriod=30 Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.867357 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="01547c5a-5ca7-435b-ba75-52e075800cf3" containerName="sg-core" containerID="cri-o://0576e30010b483c3666500ba4802f893b0075c3a2f67ddc2df4245d7e05db25a" gracePeriod=30 Feb 17 13:23:13 crc kubenswrapper[4955]: I0217 13:23:13.875440 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6ff67b7c6c-9dcnd" event={"ID":"20f0b6cf-c788-45bd-bdb2-cb2e01608a08","Type":"ContainerStarted","Data":"00e0922509406bbaecb9251a2eed9b8831436c20c06e2e544599dacdc3d0c29b"} Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.158908 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04855a3a-d072-45f0-95d1-fb2f93264a52-config\") pod \"04855a3a-d072-45f0-95d1-fb2f93264a52\" (UID: \"04855a3a-d072-45f0-95d1-fb2f93264a52\") " Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.160113 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04855a3a-d072-45f0-95d1-fb2f93264a52-config" (OuterVolumeSpecName: "config") pod "04855a3a-d072-45f0-95d1-fb2f93264a52" (UID: "04855a3a-d072-45f0-95d1-fb2f93264a52"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.262167 4955 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04855a3a-d072-45f0-95d1-fb2f93264a52-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.282904 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Feb 17 13:23:14 crc kubenswrapper[4955]: E0217 13:23:14.283543 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1c2d0c5-7647-447f-a1a6-b37d8c09310c" containerName="cinder-db-sync" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.283566 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1c2d0c5-7647-447f-a1a6-b37d8c09310c" containerName="cinder-db-sync" Feb 17 13:23:14 crc kubenswrapper[4955]: E0217 13:23:14.283616 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04855a3a-d072-45f0-95d1-fb2f93264a52" containerName="dnsmasq-dns" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.283625 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="04855a3a-d072-45f0-95d1-fb2f93264a52" containerName="dnsmasq-dns" Feb 17 13:23:14 crc kubenswrapper[4955]: E0217 13:23:14.283638 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04855a3a-d072-45f0-95d1-fb2f93264a52" containerName="init" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.283646 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="04855a3a-d072-45f0-95d1-fb2f93264a52" containerName="init" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.283885 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="04855a3a-d072-45f0-95d1-fb2f93264a52" containerName="dnsmasq-dns" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.283913 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1c2d0c5-7647-447f-a1a6-b37d8c09310c" containerName="cinder-db-sync" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.285304 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.291174 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.291369 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.291660 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.291829 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-zw4xp" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.310271 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.419302 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-8z8kz"] Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.440935 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-hvvzs"] Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.449809 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-hvvzs" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.454541 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-k2p8k"] Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.469130 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-hvvzs"] Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.470423 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99f00e22-f2b5-4e1f-b00a-9e50a1ed359e-scripts\") pod \"cinder-scheduler-0\" (UID: \"99f00e22-f2b5-4e1f-b00a-9e50a1ed359e\") " pod="openstack/cinder-scheduler-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.470472 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/99f00e22-f2b5-4e1f-b00a-9e50a1ed359e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"99f00e22-f2b5-4e1f-b00a-9e50a1ed359e\") " pod="openstack/cinder-scheduler-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.470505 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/99f00e22-f2b5-4e1f-b00a-9e50a1ed359e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"99f00e22-f2b5-4e1f-b00a-9e50a1ed359e\") " pod="openstack/cinder-scheduler-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.470582 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99f00e22-f2b5-4e1f-b00a-9e50a1ed359e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"99f00e22-f2b5-4e1f-b00a-9e50a1ed359e\") " pod="openstack/cinder-scheduler-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.470604 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99f00e22-f2b5-4e1f-b00a-9e50a1ed359e-config-data\") pod \"cinder-scheduler-0\" (UID: \"99f00e22-f2b5-4e1f-b00a-9e50a1ed359e\") " pod="openstack/cinder-scheduler-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.470726 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7sjc\" (UniqueName: \"kubernetes.io/projected/99f00e22-f2b5-4e1f-b00a-9e50a1ed359e-kube-api-access-x7sjc\") pod \"cinder-scheduler-0\" (UID: \"99f00e22-f2b5-4e1f-b00a-9e50a1ed359e\") " pod="openstack/cinder-scheduler-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.483828 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-k2p8k"] Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.524421 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.525807 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.537118 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.561418 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.572140 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99f00e22-f2b5-4e1f-b00a-9e50a1ed359e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"99f00e22-f2b5-4e1f-b00a-9e50a1ed359e\") " pod="openstack/cinder-scheduler-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.572172 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99f00e22-f2b5-4e1f-b00a-9e50a1ed359e-config-data\") pod \"cinder-scheduler-0\" (UID: \"99f00e22-f2b5-4e1f-b00a-9e50a1ed359e\") " pod="openstack/cinder-scheduler-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.572207 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/29f9dab6-c497-4a2b-a39c-0c58adae2aa1-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-hvvzs\" (UID: \"29f9dab6-c497-4a2b-a39c-0c58adae2aa1\") " pod="openstack/dnsmasq-dns-6578955fd5-hvvzs" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.572230 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29f9dab6-c497-4a2b-a39c-0c58adae2aa1-config\") pod \"dnsmasq-dns-6578955fd5-hvvzs\" (UID: \"29f9dab6-c497-4a2b-a39c-0c58adae2aa1\") " pod="openstack/dnsmasq-dns-6578955fd5-hvvzs" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.572326 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/29f9dab6-c497-4a2b-a39c-0c58adae2aa1-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-hvvzs\" (UID: \"29f9dab6-c497-4a2b-a39c-0c58adae2aa1\") " pod="openstack/dnsmasq-dns-6578955fd5-hvvzs" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.572347 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7sjc\" (UniqueName: \"kubernetes.io/projected/99f00e22-f2b5-4e1f-b00a-9e50a1ed359e-kube-api-access-x7sjc\") pod \"cinder-scheduler-0\" (UID: \"99f00e22-f2b5-4e1f-b00a-9e50a1ed359e\") " pod="openstack/cinder-scheduler-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.572374 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99f00e22-f2b5-4e1f-b00a-9e50a1ed359e-scripts\") pod \"cinder-scheduler-0\" (UID: \"99f00e22-f2b5-4e1f-b00a-9e50a1ed359e\") " pod="openstack/cinder-scheduler-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.572398 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/99f00e22-f2b5-4e1f-b00a-9e50a1ed359e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"99f00e22-f2b5-4e1f-b00a-9e50a1ed359e\") " pod="openstack/cinder-scheduler-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.572431 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/99f00e22-f2b5-4e1f-b00a-9e50a1ed359e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"99f00e22-f2b5-4e1f-b00a-9e50a1ed359e\") " pod="openstack/cinder-scheduler-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.572453 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfgpf\" (UniqueName: \"kubernetes.io/projected/29f9dab6-c497-4a2b-a39c-0c58adae2aa1-kube-api-access-rfgpf\") pod \"dnsmasq-dns-6578955fd5-hvvzs\" (UID: \"29f9dab6-c497-4a2b-a39c-0c58adae2aa1\") " pod="openstack/dnsmasq-dns-6578955fd5-hvvzs" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.572494 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/29f9dab6-c497-4a2b-a39c-0c58adae2aa1-dns-svc\") pod \"dnsmasq-dns-6578955fd5-hvvzs\" (UID: \"29f9dab6-c497-4a2b-a39c-0c58adae2aa1\") " pod="openstack/dnsmasq-dns-6578955fd5-hvvzs" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.572510 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/29f9dab6-c497-4a2b-a39c-0c58adae2aa1-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-hvvzs\" (UID: \"29f9dab6-c497-4a2b-a39c-0c58adae2aa1\") " pod="openstack/dnsmasq-dns-6578955fd5-hvvzs" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.573704 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/99f00e22-f2b5-4e1f-b00a-9e50a1ed359e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"99f00e22-f2b5-4e1f-b00a-9e50a1ed359e\") " pod="openstack/cinder-scheduler-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.578202 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99f00e22-f2b5-4e1f-b00a-9e50a1ed359e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"99f00e22-f2b5-4e1f-b00a-9e50a1ed359e\") " pod="openstack/cinder-scheduler-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.579357 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99f00e22-f2b5-4e1f-b00a-9e50a1ed359e-config-data\") pod \"cinder-scheduler-0\" (UID: \"99f00e22-f2b5-4e1f-b00a-9e50a1ed359e\") " pod="openstack/cinder-scheduler-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.581270 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99f00e22-f2b5-4e1f-b00a-9e50a1ed359e-scripts\") pod \"cinder-scheduler-0\" (UID: \"99f00e22-f2b5-4e1f-b00a-9e50a1ed359e\") " pod="openstack/cinder-scheduler-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.584250 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/99f00e22-f2b5-4e1f-b00a-9e50a1ed359e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"99f00e22-f2b5-4e1f-b00a-9e50a1ed359e\") " pod="openstack/cinder-scheduler-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.599355 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7sjc\" (UniqueName: \"kubernetes.io/projected/99f00e22-f2b5-4e1f-b00a-9e50a1ed359e-kube-api-access-x7sjc\") pod \"cinder-scheduler-0\" (UID: \"99f00e22-f2b5-4e1f-b00a-9e50a1ed359e\") " pod="openstack/cinder-scheduler-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.609447 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.673632 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/29f9dab6-c497-4a2b-a39c-0c58adae2aa1-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-hvvzs\" (UID: \"29f9dab6-c497-4a2b-a39c-0c58adae2aa1\") " pod="openstack/dnsmasq-dns-6578955fd5-hvvzs" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.673680 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29f9dab6-c497-4a2b-a39c-0c58adae2aa1-config\") pod \"dnsmasq-dns-6578955fd5-hvvzs\" (UID: \"29f9dab6-c497-4a2b-a39c-0c58adae2aa1\") " pod="openstack/dnsmasq-dns-6578955fd5-hvvzs" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.673706 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kz9th\" (UniqueName: \"kubernetes.io/projected/96771c04-ac0c-4eb8-8aa9-d59b754617d4-kube-api-access-kz9th\") pod \"cinder-api-0\" (UID: \"96771c04-ac0c-4eb8-8aa9-d59b754617d4\") " pod="openstack/cinder-api-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.673765 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96771c04-ac0c-4eb8-8aa9-d59b754617d4-config-data\") pod \"cinder-api-0\" (UID: \"96771c04-ac0c-4eb8-8aa9-d59b754617d4\") " pod="openstack/cinder-api-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.673804 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96771c04-ac0c-4eb8-8aa9-d59b754617d4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"96771c04-ac0c-4eb8-8aa9-d59b754617d4\") " pod="openstack/cinder-api-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.673828 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96771c04-ac0c-4eb8-8aa9-d59b754617d4-logs\") pod \"cinder-api-0\" (UID: \"96771c04-ac0c-4eb8-8aa9-d59b754617d4\") " pod="openstack/cinder-api-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.673855 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/29f9dab6-c497-4a2b-a39c-0c58adae2aa1-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-hvvzs\" (UID: \"29f9dab6-c497-4a2b-a39c-0c58adae2aa1\") " pod="openstack/dnsmasq-dns-6578955fd5-hvvzs" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.673909 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfgpf\" (UniqueName: \"kubernetes.io/projected/29f9dab6-c497-4a2b-a39c-0c58adae2aa1-kube-api-access-rfgpf\") pod \"dnsmasq-dns-6578955fd5-hvvzs\" (UID: \"29f9dab6-c497-4a2b-a39c-0c58adae2aa1\") " pod="openstack/dnsmasq-dns-6578955fd5-hvvzs" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.673932 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/96771c04-ac0c-4eb8-8aa9-d59b754617d4-config-data-custom\") pod \"cinder-api-0\" (UID: \"96771c04-ac0c-4eb8-8aa9-d59b754617d4\") " pod="openstack/cinder-api-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.673965 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96771c04-ac0c-4eb8-8aa9-d59b754617d4-scripts\") pod \"cinder-api-0\" (UID: \"96771c04-ac0c-4eb8-8aa9-d59b754617d4\") " pod="openstack/cinder-api-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.673983 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/29f9dab6-c497-4a2b-a39c-0c58adae2aa1-dns-svc\") pod \"dnsmasq-dns-6578955fd5-hvvzs\" (UID: \"29f9dab6-c497-4a2b-a39c-0c58adae2aa1\") " pod="openstack/dnsmasq-dns-6578955fd5-hvvzs" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.673999 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/29f9dab6-c497-4a2b-a39c-0c58adae2aa1-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-hvvzs\" (UID: \"29f9dab6-c497-4a2b-a39c-0c58adae2aa1\") " pod="openstack/dnsmasq-dns-6578955fd5-hvvzs" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.674016 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/96771c04-ac0c-4eb8-8aa9-d59b754617d4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"96771c04-ac0c-4eb8-8aa9-d59b754617d4\") " pod="openstack/cinder-api-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.675021 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/29f9dab6-c497-4a2b-a39c-0c58adae2aa1-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-hvvzs\" (UID: \"29f9dab6-c497-4a2b-a39c-0c58adae2aa1\") " pod="openstack/dnsmasq-dns-6578955fd5-hvvzs" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.675568 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29f9dab6-c497-4a2b-a39c-0c58adae2aa1-config\") pod \"dnsmasq-dns-6578955fd5-hvvzs\" (UID: \"29f9dab6-c497-4a2b-a39c-0c58adae2aa1\") " pod="openstack/dnsmasq-dns-6578955fd5-hvvzs" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.676695 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/29f9dab6-c497-4a2b-a39c-0c58adae2aa1-dns-svc\") pod \"dnsmasq-dns-6578955fd5-hvvzs\" (UID: \"29f9dab6-c497-4a2b-a39c-0c58adae2aa1\") " pod="openstack/dnsmasq-dns-6578955fd5-hvvzs" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.677142 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/29f9dab6-c497-4a2b-a39c-0c58adae2aa1-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-hvvzs\" (UID: \"29f9dab6-c497-4a2b-a39c-0c58adae2aa1\") " pod="openstack/dnsmasq-dns-6578955fd5-hvvzs" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.678342 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/29f9dab6-c497-4a2b-a39c-0c58adae2aa1-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-hvvzs\" (UID: \"29f9dab6-c497-4a2b-a39c-0c58adae2aa1\") " pod="openstack/dnsmasq-dns-6578955fd5-hvvzs" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.704624 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfgpf\" (UniqueName: \"kubernetes.io/projected/29f9dab6-c497-4a2b-a39c-0c58adae2aa1-kube-api-access-rfgpf\") pod \"dnsmasq-dns-6578955fd5-hvvzs\" (UID: \"29f9dab6-c497-4a2b-a39c-0c58adae2aa1\") " pod="openstack/dnsmasq-dns-6578955fd5-hvvzs" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.782231 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96771c04-ac0c-4eb8-8aa9-d59b754617d4-scripts\") pod \"cinder-api-0\" (UID: \"96771c04-ac0c-4eb8-8aa9-d59b754617d4\") " pod="openstack/cinder-api-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.782277 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/96771c04-ac0c-4eb8-8aa9-d59b754617d4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"96771c04-ac0c-4eb8-8aa9-d59b754617d4\") " pod="openstack/cinder-api-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.782322 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kz9th\" (UniqueName: \"kubernetes.io/projected/96771c04-ac0c-4eb8-8aa9-d59b754617d4-kube-api-access-kz9th\") pod \"cinder-api-0\" (UID: \"96771c04-ac0c-4eb8-8aa9-d59b754617d4\") " pod="openstack/cinder-api-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.782408 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96771c04-ac0c-4eb8-8aa9-d59b754617d4-config-data\") pod \"cinder-api-0\" (UID: \"96771c04-ac0c-4eb8-8aa9-d59b754617d4\") " pod="openstack/cinder-api-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.782445 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96771c04-ac0c-4eb8-8aa9-d59b754617d4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"96771c04-ac0c-4eb8-8aa9-d59b754617d4\") " pod="openstack/cinder-api-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.782473 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96771c04-ac0c-4eb8-8aa9-d59b754617d4-logs\") pod \"cinder-api-0\" (UID: \"96771c04-ac0c-4eb8-8aa9-d59b754617d4\") " pod="openstack/cinder-api-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.782544 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/96771c04-ac0c-4eb8-8aa9-d59b754617d4-config-data-custom\") pod \"cinder-api-0\" (UID: \"96771c04-ac0c-4eb8-8aa9-d59b754617d4\") " pod="openstack/cinder-api-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.782970 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/96771c04-ac0c-4eb8-8aa9-d59b754617d4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"96771c04-ac0c-4eb8-8aa9-d59b754617d4\") " pod="openstack/cinder-api-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.783369 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96771c04-ac0c-4eb8-8aa9-d59b754617d4-logs\") pod \"cinder-api-0\" (UID: \"96771c04-ac0c-4eb8-8aa9-d59b754617d4\") " pod="openstack/cinder-api-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.788747 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96771c04-ac0c-4eb8-8aa9-d59b754617d4-config-data\") pod \"cinder-api-0\" (UID: \"96771c04-ac0c-4eb8-8aa9-d59b754617d4\") " pod="openstack/cinder-api-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.798218 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/96771c04-ac0c-4eb8-8aa9-d59b754617d4-config-data-custom\") pod \"cinder-api-0\" (UID: \"96771c04-ac0c-4eb8-8aa9-d59b754617d4\") " pod="openstack/cinder-api-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.798497 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-hvvzs" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.805501 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96771c04-ac0c-4eb8-8aa9-d59b754617d4-scripts\") pod \"cinder-api-0\" (UID: \"96771c04-ac0c-4eb8-8aa9-d59b754617d4\") " pod="openstack/cinder-api-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.809000 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96771c04-ac0c-4eb8-8aa9-d59b754617d4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"96771c04-ac0c-4eb8-8aa9-d59b754617d4\") " pod="openstack/cinder-api-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.810661 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kz9th\" (UniqueName: \"kubernetes.io/projected/96771c04-ac0c-4eb8-8aa9-d59b754617d4-kube-api-access-kz9th\") pod \"cinder-api-0\" (UID: \"96771c04-ac0c-4eb8-8aa9-d59b754617d4\") " pod="openstack/cinder-api-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.884993 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.911459 4955 generic.go:334] "Generic (PLEG): container finished" podID="01547c5a-5ca7-435b-ba75-52e075800cf3" containerID="198e280433fd484d8cb761a7db110f722d42eff7b1ea7c8cf1dbbffef6877c20" exitCode=0 Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.911484 4955 generic.go:334] "Generic (PLEG): container finished" podID="01547c5a-5ca7-435b-ba75-52e075800cf3" containerID="0576e30010b483c3666500ba4802f893b0075c3a2f67ddc2df4245d7e05db25a" exitCode=2 Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.911548 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01547c5a-5ca7-435b-ba75-52e075800cf3","Type":"ContainerDied","Data":"198e280433fd484d8cb761a7db110f722d42eff7b1ea7c8cf1dbbffef6877c20"} Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.911601 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01547c5a-5ca7-435b-ba75-52e075800cf3","Type":"ContainerDied","Data":"0576e30010b483c3666500ba4802f893b0075c3a2f67ddc2df4245d7e05db25a"} Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.917493 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65b6bfc874-zgznk" event={"ID":"353a4cd6-336b-4545-903e-245c9cea97ca","Type":"ContainerStarted","Data":"9e1cde53c1148029ca5f31661b223a8898cf091c0dc94ec858633087db1437ec"} Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.917537 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65b6bfc874-zgznk" event={"ID":"353a4cd6-336b-4545-903e-245c9cea97ca","Type":"ContainerStarted","Data":"7e9f2e7d38f46f5983cb4dfffe1c3cdfe6f350832d7f2c71a0ef1d278ec760f6"} Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.917549 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65b6bfc874-zgznk" event={"ID":"353a4cd6-336b-4545-903e-245c9cea97ca","Type":"ContainerStarted","Data":"2499fcc3732e6ff01b91c42700f4df6b62b3a18843343f3bd5e83c56479e2b5b"} Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.920121 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-65b6bfc874-zgznk" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.920337 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-65b6bfc874-zgznk" Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.936520 4955 generic.go:334] "Generic (PLEG): container finished" podID="2793f42e-26b2-4115-9ac5-9b6c1f503218" containerID="949d4ded065f983a7b4f0c9a7c9118fc2891f670cc064eeab87d63e19100e60f" exitCode=0 Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.936573 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-8z8kz" event={"ID":"2793f42e-26b2-4115-9ac5-9b6c1f503218","Type":"ContainerDied","Data":"949d4ded065f983a7b4f0c9a7c9118fc2891f670cc064eeab87d63e19100e60f"} Feb 17 13:23:14 crc kubenswrapper[4955]: I0217 13:23:14.951442 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-65b6bfc874-zgznk" podStartSLOduration=2.951422621 podStartE2EDuration="2.951422621s" podCreationTimestamp="2026-02-17 13:23:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:23:14.937936736 +0000 UTC m=+1133.460666289" watchObservedRunningTime="2026-02-17 13:23:14.951422621 +0000 UTC m=+1133.474152164" Feb 17 13:23:15 crc kubenswrapper[4955]: I0217 13:23:15.274887 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 17 13:23:15 crc kubenswrapper[4955]: E0217 13:23:15.298600 4955 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Feb 17 13:23:15 crc kubenswrapper[4955]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/2793f42e-26b2-4115-9ac5-9b6c1f503218/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Feb 17 13:23:15 crc kubenswrapper[4955]: > podSandboxID="b68c73e9a5be1fcf6f48f84069a92921bfbd3cbc32ed2d9ec6bddb7ff3426d9d" Feb 17 13:23:15 crc kubenswrapper[4955]: E0217 13:23:15.298792 4955 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 17 13:23:15 crc kubenswrapper[4955]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n99h8bhd9h696h649h588h5c6h658h5b4h57fh65h89h5f5h56h696h5dh8h57h597h68ch568h58dh66hf4h675h598h588h67dhb5h69h5dh6bq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-swift-storage-0,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-swift-storage-0,SubPath:dns-swift-storage-0,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-nb,SubPath:ovsdbserver-nb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-sb,SubPath:ovsdbserver-sb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6rwv2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-848cf88cfc-8z8kz_openstack(2793f42e-26b2-4115-9ac5-9b6c1f503218): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/2793f42e-26b2-4115-9ac5-9b6c1f503218/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Feb 17 13:23:15 crc kubenswrapper[4955]: > logger="UnhandledError" Feb 17 13:23:15 crc kubenswrapper[4955]: E0217 13:23:15.302369 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/2793f42e-26b2-4115-9ac5-9b6c1f503218/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-848cf88cfc-8z8kz" podUID="2793f42e-26b2-4115-9ac5-9b6c1f503218" Feb 17 13:23:15 crc kubenswrapper[4955]: I0217 13:23:15.514467 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-hvvzs"] Feb 17 13:23:15 crc kubenswrapper[4955]: I0217 13:23:15.665393 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 17 13:23:15 crc kubenswrapper[4955]: I0217 13:23:15.945546 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"99f00e22-f2b5-4e1f-b00a-9e50a1ed359e","Type":"ContainerStarted","Data":"398adef26b9f13332ecfddd83d219068aad43f3c63ecfa90d85c0275922e7634"} Feb 17 13:23:15 crc kubenswrapper[4955]: I0217 13:23:15.948482 4955 generic.go:334] "Generic (PLEG): container finished" podID="01547c5a-5ca7-435b-ba75-52e075800cf3" containerID="6773574f1d1d99cdf57467650dff95afe5982095ed615f3e81c7756f97852f50" exitCode=0 Feb 17 13:23:15 crc kubenswrapper[4955]: I0217 13:23:15.948580 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01547c5a-5ca7-435b-ba75-52e075800cf3","Type":"ContainerDied","Data":"6773574f1d1d99cdf57467650dff95afe5982095ed615f3e81c7756f97852f50"} Feb 17 13:23:16 crc kubenswrapper[4955]: I0217 13:23:16.244199 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04855a3a-d072-45f0-95d1-fb2f93264a52" path="/var/lib/kubelet/pods/04855a3a-d072-45f0-95d1-fb2f93264a52/volumes" Feb 17 13:23:16 crc kubenswrapper[4955]: I0217 13:23:16.245353 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-5788cc4448-245bs" Feb 17 13:23:16 crc kubenswrapper[4955]: I0217 13:23:16.400485 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7fffbcf7cb-wr74b" Feb 17 13:23:16 crc kubenswrapper[4955]: I0217 13:23:16.730332 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-8z8kz" Feb 17 13:23:16 crc kubenswrapper[4955]: I0217 13:23:16.745158 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 17 13:23:16 crc kubenswrapper[4955]: I0217 13:23:16.784146 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rwv2\" (UniqueName: \"kubernetes.io/projected/2793f42e-26b2-4115-9ac5-9b6c1f503218-kube-api-access-6rwv2\") pod \"2793f42e-26b2-4115-9ac5-9b6c1f503218\" (UID: \"2793f42e-26b2-4115-9ac5-9b6c1f503218\") " Feb 17 13:23:16 crc kubenswrapper[4955]: I0217 13:23:16.784329 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2793f42e-26b2-4115-9ac5-9b6c1f503218-ovsdbserver-nb\") pod \"2793f42e-26b2-4115-9ac5-9b6c1f503218\" (UID: \"2793f42e-26b2-4115-9ac5-9b6c1f503218\") " Feb 17 13:23:16 crc kubenswrapper[4955]: I0217 13:23:16.784401 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2793f42e-26b2-4115-9ac5-9b6c1f503218-ovsdbserver-sb\") pod \"2793f42e-26b2-4115-9ac5-9b6c1f503218\" (UID: \"2793f42e-26b2-4115-9ac5-9b6c1f503218\") " Feb 17 13:23:16 crc kubenswrapper[4955]: I0217 13:23:16.784428 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2793f42e-26b2-4115-9ac5-9b6c1f503218-dns-svc\") pod \"2793f42e-26b2-4115-9ac5-9b6c1f503218\" (UID: \"2793f42e-26b2-4115-9ac5-9b6c1f503218\") " Feb 17 13:23:16 crc kubenswrapper[4955]: I0217 13:23:16.784501 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2793f42e-26b2-4115-9ac5-9b6c1f503218-dns-swift-storage-0\") pod \"2793f42e-26b2-4115-9ac5-9b6c1f503218\" (UID: \"2793f42e-26b2-4115-9ac5-9b6c1f503218\") " Feb 17 13:23:16 crc kubenswrapper[4955]: I0217 13:23:16.784679 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2793f42e-26b2-4115-9ac5-9b6c1f503218-config\") pod \"2793f42e-26b2-4115-9ac5-9b6c1f503218\" (UID: \"2793f42e-26b2-4115-9ac5-9b6c1f503218\") " Feb 17 13:23:16 crc kubenswrapper[4955]: I0217 13:23:16.797440 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2793f42e-26b2-4115-9ac5-9b6c1f503218-kube-api-access-6rwv2" (OuterVolumeSpecName: "kube-api-access-6rwv2") pod "2793f42e-26b2-4115-9ac5-9b6c1f503218" (UID: "2793f42e-26b2-4115-9ac5-9b6c1f503218"). InnerVolumeSpecName "kube-api-access-6rwv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:23:16 crc kubenswrapper[4955]: I0217 13:23:16.853922 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2793f42e-26b2-4115-9ac5-9b6c1f503218-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2793f42e-26b2-4115-9ac5-9b6c1f503218" (UID: "2793f42e-26b2-4115-9ac5-9b6c1f503218"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:23:16 crc kubenswrapper[4955]: I0217 13:23:16.858338 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2793f42e-26b2-4115-9ac5-9b6c1f503218-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2793f42e-26b2-4115-9ac5-9b6c1f503218" (UID: "2793f42e-26b2-4115-9ac5-9b6c1f503218"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:23:16 crc kubenswrapper[4955]: I0217 13:23:16.868169 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2793f42e-26b2-4115-9ac5-9b6c1f503218-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2793f42e-26b2-4115-9ac5-9b6c1f503218" (UID: "2793f42e-26b2-4115-9ac5-9b6c1f503218"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:23:16 crc kubenswrapper[4955]: I0217 13:23:16.868177 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2793f42e-26b2-4115-9ac5-9b6c1f503218-config" (OuterVolumeSpecName: "config") pod "2793f42e-26b2-4115-9ac5-9b6c1f503218" (UID: "2793f42e-26b2-4115-9ac5-9b6c1f503218"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:23:16 crc kubenswrapper[4955]: I0217 13:23:16.877736 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2793f42e-26b2-4115-9ac5-9b6c1f503218-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2793f42e-26b2-4115-9ac5-9b6c1f503218" (UID: "2793f42e-26b2-4115-9ac5-9b6c1f503218"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:23:16 crc kubenswrapper[4955]: I0217 13:23:16.886955 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01547c5a-5ca7-435b-ba75-52e075800cf3-config-data\") pod \"01547c5a-5ca7-435b-ba75-52e075800cf3\" (UID: \"01547c5a-5ca7-435b-ba75-52e075800cf3\") " Feb 17 13:23:16 crc kubenswrapper[4955]: I0217 13:23:16.887001 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01547c5a-5ca7-435b-ba75-52e075800cf3-scripts\") pod \"01547c5a-5ca7-435b-ba75-52e075800cf3\" (UID: \"01547c5a-5ca7-435b-ba75-52e075800cf3\") " Feb 17 13:23:16 crc kubenswrapper[4955]: I0217 13:23:16.887073 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01547c5a-5ca7-435b-ba75-52e075800cf3-log-httpd\") pod \"01547c5a-5ca7-435b-ba75-52e075800cf3\" (UID: \"01547c5a-5ca7-435b-ba75-52e075800cf3\") " Feb 17 13:23:16 crc kubenswrapper[4955]: I0217 13:23:16.887127 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xcff\" (UniqueName: \"kubernetes.io/projected/01547c5a-5ca7-435b-ba75-52e075800cf3-kube-api-access-4xcff\") pod \"01547c5a-5ca7-435b-ba75-52e075800cf3\" (UID: \"01547c5a-5ca7-435b-ba75-52e075800cf3\") " Feb 17 13:23:16 crc kubenswrapper[4955]: I0217 13:23:16.887163 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01547c5a-5ca7-435b-ba75-52e075800cf3-run-httpd\") pod \"01547c5a-5ca7-435b-ba75-52e075800cf3\" (UID: \"01547c5a-5ca7-435b-ba75-52e075800cf3\") " Feb 17 13:23:16 crc kubenswrapper[4955]: I0217 13:23:16.887208 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/01547c5a-5ca7-435b-ba75-52e075800cf3-sg-core-conf-yaml\") pod \"01547c5a-5ca7-435b-ba75-52e075800cf3\" (UID: \"01547c5a-5ca7-435b-ba75-52e075800cf3\") " Feb 17 13:23:16 crc kubenswrapper[4955]: I0217 13:23:16.887354 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01547c5a-5ca7-435b-ba75-52e075800cf3-combined-ca-bundle\") pod \"01547c5a-5ca7-435b-ba75-52e075800cf3\" (UID: \"01547c5a-5ca7-435b-ba75-52e075800cf3\") " Feb 17 13:23:16 crc kubenswrapper[4955]: I0217 13:23:16.887689 4955 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2793f42e-26b2-4115-9ac5-9b6c1f503218-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:16 crc kubenswrapper[4955]: I0217 13:23:16.887705 4955 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2793f42e-26b2-4115-9ac5-9b6c1f503218-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:16 crc kubenswrapper[4955]: I0217 13:23:16.887714 4955 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2793f42e-26b2-4115-9ac5-9b6c1f503218-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:16 crc kubenswrapper[4955]: I0217 13:23:16.887723 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6rwv2\" (UniqueName: \"kubernetes.io/projected/2793f42e-26b2-4115-9ac5-9b6c1f503218-kube-api-access-6rwv2\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:16 crc kubenswrapper[4955]: I0217 13:23:16.887732 4955 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2793f42e-26b2-4115-9ac5-9b6c1f503218-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:16 crc kubenswrapper[4955]: I0217 13:23:16.887740 4955 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2793f42e-26b2-4115-9ac5-9b6c1f503218-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:16 crc kubenswrapper[4955]: I0217 13:23:16.888270 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01547c5a-5ca7-435b-ba75-52e075800cf3-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "01547c5a-5ca7-435b-ba75-52e075800cf3" (UID: "01547c5a-5ca7-435b-ba75-52e075800cf3"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:23:16 crc kubenswrapper[4955]: I0217 13:23:16.889110 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01547c5a-5ca7-435b-ba75-52e075800cf3-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "01547c5a-5ca7-435b-ba75-52e075800cf3" (UID: "01547c5a-5ca7-435b-ba75-52e075800cf3"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:23:16 crc kubenswrapper[4955]: I0217 13:23:16.899273 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01547c5a-5ca7-435b-ba75-52e075800cf3-scripts" (OuterVolumeSpecName: "scripts") pod "01547c5a-5ca7-435b-ba75-52e075800cf3" (UID: "01547c5a-5ca7-435b-ba75-52e075800cf3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:16 crc kubenswrapper[4955]: I0217 13:23:16.899324 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01547c5a-5ca7-435b-ba75-52e075800cf3-kube-api-access-4xcff" (OuterVolumeSpecName: "kube-api-access-4xcff") pod "01547c5a-5ca7-435b-ba75-52e075800cf3" (UID: "01547c5a-5ca7-435b-ba75-52e075800cf3"). InnerVolumeSpecName "kube-api-access-4xcff". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:23:16 crc kubenswrapper[4955]: I0217 13:23:16.949771 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01547c5a-5ca7-435b-ba75-52e075800cf3-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "01547c5a-5ca7-435b-ba75-52e075800cf3" (UID: "01547c5a-5ca7-435b-ba75-52e075800cf3"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:16 crc kubenswrapper[4955]: I0217 13:23:16.989467 4955 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01547c5a-5ca7-435b-ba75-52e075800cf3-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:16 crc kubenswrapper[4955]: I0217 13:23:16.989498 4955 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01547c5a-5ca7-435b-ba75-52e075800cf3-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:16 crc kubenswrapper[4955]: I0217 13:23:16.989530 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xcff\" (UniqueName: \"kubernetes.io/projected/01547c5a-5ca7-435b-ba75-52e075800cf3-kube-api-access-4xcff\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:16 crc kubenswrapper[4955]: I0217 13:23:16.989545 4955 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01547c5a-5ca7-435b-ba75-52e075800cf3-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:16 crc kubenswrapper[4955]: I0217 13:23:16.989556 4955 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/01547c5a-5ca7-435b-ba75-52e075800cf3-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.000947 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-8z8kz" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.001068 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-8z8kz" event={"ID":"2793f42e-26b2-4115-9ac5-9b6c1f503218","Type":"ContainerDied","Data":"b68c73e9a5be1fcf6f48f84069a92921bfbd3cbc32ed2d9ec6bddb7ff3426d9d"} Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.001124 4955 scope.go:117] "RemoveContainer" containerID="949d4ded065f983a7b4f0c9a7c9118fc2891f670cc064eeab87d63e19100e60f" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.006808 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.006864 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01547c5a-5ca7-435b-ba75-52e075800cf3","Type":"ContainerDied","Data":"c1577238904bde655336171e1f902b973ec8981862e0b663c6e6dacfeb580d1d"} Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.010187 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-hvvzs" event={"ID":"29f9dab6-c497-4a2b-a39c-0c58adae2aa1","Type":"ContainerStarted","Data":"f90ddae94e8e106c7edb0893caf8923a733f2b267bba1313ad0a500fee111ac1"} Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.012826 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"96771c04-ac0c-4eb8-8aa9-d59b754617d4","Type":"ContainerStarted","Data":"61e20a02fc68bc5fb0b2d7985ccdd92bf0c4b8326b146fc59411e1856624f36c"} Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.030874 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01547c5a-5ca7-435b-ba75-52e075800cf3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "01547c5a-5ca7-435b-ba75-52e075800cf3" (UID: "01547c5a-5ca7-435b-ba75-52e075800cf3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.055803 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01547c5a-5ca7-435b-ba75-52e075800cf3-config-data" (OuterVolumeSpecName: "config-data") pod "01547c5a-5ca7-435b-ba75-52e075800cf3" (UID: "01547c5a-5ca7-435b-ba75-52e075800cf3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.091360 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01547c5a-5ca7-435b-ba75-52e075800cf3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.091396 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01547c5a-5ca7-435b-ba75-52e075800cf3-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.147828 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-8z8kz"] Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.158933 4955 scope.go:117] "RemoveContainer" containerID="198e280433fd484d8cb761a7db110f722d42eff7b1ea7c8cf1dbbffef6877c20" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.165490 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-8z8kz"] Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.281217 4955 scope.go:117] "RemoveContainer" containerID="0576e30010b483c3666500ba4802f893b0075c3a2f67ddc2df4245d7e05db25a" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.311674 4955 scope.go:117] "RemoveContainer" containerID="6773574f1d1d99cdf57467650dff95afe5982095ed615f3e81c7756f97852f50" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.407486 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.440363 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.460055 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:23:17 crc kubenswrapper[4955]: E0217 13:23:17.460573 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2793f42e-26b2-4115-9ac5-9b6c1f503218" containerName="init" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.460602 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="2793f42e-26b2-4115-9ac5-9b6c1f503218" containerName="init" Feb 17 13:23:17 crc kubenswrapper[4955]: E0217 13:23:17.460623 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01547c5a-5ca7-435b-ba75-52e075800cf3" containerName="ceilometer-notification-agent" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.460631 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="01547c5a-5ca7-435b-ba75-52e075800cf3" containerName="ceilometer-notification-agent" Feb 17 13:23:17 crc kubenswrapper[4955]: E0217 13:23:17.460652 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01547c5a-5ca7-435b-ba75-52e075800cf3" containerName="proxy-httpd" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.460659 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="01547c5a-5ca7-435b-ba75-52e075800cf3" containerName="proxy-httpd" Feb 17 13:23:17 crc kubenswrapper[4955]: E0217 13:23:17.460680 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01547c5a-5ca7-435b-ba75-52e075800cf3" containerName="sg-core" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.460686 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="01547c5a-5ca7-435b-ba75-52e075800cf3" containerName="sg-core" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.460878 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="01547c5a-5ca7-435b-ba75-52e075800cf3" containerName="proxy-httpd" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.460899 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="2793f42e-26b2-4115-9ac5-9b6c1f503218" containerName="init" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.460912 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="01547c5a-5ca7-435b-ba75-52e075800cf3" containerName="sg-core" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.460923 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="01547c5a-5ca7-435b-ba75-52e075800cf3" containerName="ceilometer-notification-agent" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.466556 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.471317 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.471384 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.473414 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.519558 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6cf59496f7-6ntrd" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.599240 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d1b9c8f8-6829-4eaf-b5bf-48121eaaf818-horizon-secret-key\") pod \"d1b9c8f8-6829-4eaf-b5bf-48121eaaf818\" (UID: \"d1b9c8f8-6829-4eaf-b5bf-48121eaaf818\") " Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.599295 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bv2g2\" (UniqueName: \"kubernetes.io/projected/d1b9c8f8-6829-4eaf-b5bf-48121eaaf818-kube-api-access-bv2g2\") pod \"d1b9c8f8-6829-4eaf-b5bf-48121eaaf818\" (UID: \"d1b9c8f8-6829-4eaf-b5bf-48121eaaf818\") " Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.599360 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d1b9c8f8-6829-4eaf-b5bf-48121eaaf818-logs\") pod \"d1b9c8f8-6829-4eaf-b5bf-48121eaaf818\" (UID: \"d1b9c8f8-6829-4eaf-b5bf-48121eaaf818\") " Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.599427 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d1b9c8f8-6829-4eaf-b5bf-48121eaaf818-scripts\") pod \"d1b9c8f8-6829-4eaf-b5bf-48121eaaf818\" (UID: \"d1b9c8f8-6829-4eaf-b5bf-48121eaaf818\") " Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.599457 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d1b9c8f8-6829-4eaf-b5bf-48121eaaf818-config-data\") pod \"d1b9c8f8-6829-4eaf-b5bf-48121eaaf818\" (UID: \"d1b9c8f8-6829-4eaf-b5bf-48121eaaf818\") " Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.599673 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d20f3280-ab79-434f-a89a-11d6fbfc0edf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d20f3280-ab79-434f-a89a-11d6fbfc0edf\") " pod="openstack/ceilometer-0" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.599733 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rslhr\" (UniqueName: \"kubernetes.io/projected/d20f3280-ab79-434f-a89a-11d6fbfc0edf-kube-api-access-rslhr\") pod \"ceilometer-0\" (UID: \"d20f3280-ab79-434f-a89a-11d6fbfc0edf\") " pod="openstack/ceilometer-0" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.599863 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d20f3280-ab79-434f-a89a-11d6fbfc0edf-scripts\") pod \"ceilometer-0\" (UID: \"d20f3280-ab79-434f-a89a-11d6fbfc0edf\") " pod="openstack/ceilometer-0" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.599914 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d20f3280-ab79-434f-a89a-11d6fbfc0edf-run-httpd\") pod \"ceilometer-0\" (UID: \"d20f3280-ab79-434f-a89a-11d6fbfc0edf\") " pod="openstack/ceilometer-0" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.599938 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d20f3280-ab79-434f-a89a-11d6fbfc0edf-log-httpd\") pod \"ceilometer-0\" (UID: \"d20f3280-ab79-434f-a89a-11d6fbfc0edf\") " pod="openstack/ceilometer-0" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.599961 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d20f3280-ab79-434f-a89a-11d6fbfc0edf-config-data\") pod \"ceilometer-0\" (UID: \"d20f3280-ab79-434f-a89a-11d6fbfc0edf\") " pod="openstack/ceilometer-0" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.599998 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d20f3280-ab79-434f-a89a-11d6fbfc0edf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d20f3280-ab79-434f-a89a-11d6fbfc0edf\") " pod="openstack/ceilometer-0" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.600957 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1b9c8f8-6829-4eaf-b5bf-48121eaaf818-logs" (OuterVolumeSpecName: "logs") pod "d1b9c8f8-6829-4eaf-b5bf-48121eaaf818" (UID: "d1b9c8f8-6829-4eaf-b5bf-48121eaaf818"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.612172 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1b9c8f8-6829-4eaf-b5bf-48121eaaf818-kube-api-access-bv2g2" (OuterVolumeSpecName: "kube-api-access-bv2g2") pod "d1b9c8f8-6829-4eaf-b5bf-48121eaaf818" (UID: "d1b9c8f8-6829-4eaf-b5bf-48121eaaf818"). InnerVolumeSpecName "kube-api-access-bv2g2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.615877 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1b9c8f8-6829-4eaf-b5bf-48121eaaf818-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "d1b9c8f8-6829-4eaf-b5bf-48121eaaf818" (UID: "d1b9c8f8-6829-4eaf-b5bf-48121eaaf818"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.627538 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1b9c8f8-6829-4eaf-b5bf-48121eaaf818-config-data" (OuterVolumeSpecName: "config-data") pod "d1b9c8f8-6829-4eaf-b5bf-48121eaaf818" (UID: "d1b9c8f8-6829-4eaf-b5bf-48121eaaf818"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.639468 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1b9c8f8-6829-4eaf-b5bf-48121eaaf818-scripts" (OuterVolumeSpecName: "scripts") pod "d1b9c8f8-6829-4eaf-b5bf-48121eaaf818" (UID: "d1b9c8f8-6829-4eaf-b5bf-48121eaaf818"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.701305 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d20f3280-ab79-434f-a89a-11d6fbfc0edf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d20f3280-ab79-434f-a89a-11d6fbfc0edf\") " pod="openstack/ceilometer-0" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.701391 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rslhr\" (UniqueName: \"kubernetes.io/projected/d20f3280-ab79-434f-a89a-11d6fbfc0edf-kube-api-access-rslhr\") pod \"ceilometer-0\" (UID: \"d20f3280-ab79-434f-a89a-11d6fbfc0edf\") " pod="openstack/ceilometer-0" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.701438 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d20f3280-ab79-434f-a89a-11d6fbfc0edf-scripts\") pod \"ceilometer-0\" (UID: \"d20f3280-ab79-434f-a89a-11d6fbfc0edf\") " pod="openstack/ceilometer-0" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.701482 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d20f3280-ab79-434f-a89a-11d6fbfc0edf-run-httpd\") pod \"ceilometer-0\" (UID: \"d20f3280-ab79-434f-a89a-11d6fbfc0edf\") " pod="openstack/ceilometer-0" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.701502 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d20f3280-ab79-434f-a89a-11d6fbfc0edf-log-httpd\") pod \"ceilometer-0\" (UID: \"d20f3280-ab79-434f-a89a-11d6fbfc0edf\") " pod="openstack/ceilometer-0" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.701523 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d20f3280-ab79-434f-a89a-11d6fbfc0edf-config-data\") pod \"ceilometer-0\" (UID: \"d20f3280-ab79-434f-a89a-11d6fbfc0edf\") " pod="openstack/ceilometer-0" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.701564 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d20f3280-ab79-434f-a89a-11d6fbfc0edf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d20f3280-ab79-434f-a89a-11d6fbfc0edf\") " pod="openstack/ceilometer-0" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.701629 4955 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d1b9c8f8-6829-4eaf-b5bf-48121eaaf818-logs\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.701647 4955 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d1b9c8f8-6829-4eaf-b5bf-48121eaaf818-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.701658 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d1b9c8f8-6829-4eaf-b5bf-48121eaaf818-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.701670 4955 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d1b9c8f8-6829-4eaf-b5bf-48121eaaf818-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.701682 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bv2g2\" (UniqueName: \"kubernetes.io/projected/d1b9c8f8-6829-4eaf-b5bf-48121eaaf818-kube-api-access-bv2g2\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.704921 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d20f3280-ab79-434f-a89a-11d6fbfc0edf-log-httpd\") pod \"ceilometer-0\" (UID: \"d20f3280-ab79-434f-a89a-11d6fbfc0edf\") " pod="openstack/ceilometer-0" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.705894 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d20f3280-ab79-434f-a89a-11d6fbfc0edf-run-httpd\") pod \"ceilometer-0\" (UID: \"d20f3280-ab79-434f-a89a-11d6fbfc0edf\") " pod="openstack/ceilometer-0" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.708305 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d20f3280-ab79-434f-a89a-11d6fbfc0edf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d20f3280-ab79-434f-a89a-11d6fbfc0edf\") " pod="openstack/ceilometer-0" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.709154 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d20f3280-ab79-434f-a89a-11d6fbfc0edf-scripts\") pod \"ceilometer-0\" (UID: \"d20f3280-ab79-434f-a89a-11d6fbfc0edf\") " pod="openstack/ceilometer-0" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.710063 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d20f3280-ab79-434f-a89a-11d6fbfc0edf-config-data\") pod \"ceilometer-0\" (UID: \"d20f3280-ab79-434f-a89a-11d6fbfc0edf\") " pod="openstack/ceilometer-0" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.715424 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d20f3280-ab79-434f-a89a-11d6fbfc0edf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d20f3280-ab79-434f-a89a-11d6fbfc0edf\") " pod="openstack/ceilometer-0" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.725243 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rslhr\" (UniqueName: \"kubernetes.io/projected/d20f3280-ab79-434f-a89a-11d6fbfc0edf-kube-api-access-rslhr\") pod \"ceilometer-0\" (UID: \"d20f3280-ab79-434f-a89a-11d6fbfc0edf\") " pod="openstack/ceilometer-0" Feb 17 13:23:17 crc kubenswrapper[4955]: I0217 13:23:17.809139 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.035851 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6ff67b7c6c-9dcnd" event={"ID":"20f0b6cf-c788-45bd-bdb2-cb2e01608a08","Type":"ContainerStarted","Data":"852f3359993dfd206d30a7eeea08cbbfd9138853631873fb21cd2248fc13118d"} Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.036173 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6ff67b7c6c-9dcnd" event={"ID":"20f0b6cf-c788-45bd-bdb2-cb2e01608a08","Type":"ContainerStarted","Data":"6a7832feef0cb8cda35be53351e69f6ae678562aff21df589533d46c532eb1e4"} Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.044048 4955 generic.go:334] "Generic (PLEG): container finished" podID="29f9dab6-c497-4a2b-a39c-0c58adae2aa1" containerID="800728be94187d7cbb16633c7f612b8d818340e1f95cd868d2a7088601c9ea38" exitCode=0 Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.044106 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-hvvzs" event={"ID":"29f9dab6-c497-4a2b-a39c-0c58adae2aa1","Type":"ContainerDied","Data":"800728be94187d7cbb16633c7f612b8d818340e1f95cd868d2a7088601c9ea38"} Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.046624 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"96771c04-ac0c-4eb8-8aa9-d59b754617d4","Type":"ContainerStarted","Data":"7daf52da12552e9ff871d56fe0c83c8dfdafb7b7c497ab0251922a2dbc6dab46"} Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.050855 4955 generic.go:334] "Generic (PLEG): container finished" podID="d1b9c8f8-6829-4eaf-b5bf-48121eaaf818" containerID="3f4ac0f91d02f64d61272851b2a1ac47fab311746345b16f0c2110dbf11b51b0" exitCode=137 Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.050915 4955 generic.go:334] "Generic (PLEG): container finished" podID="d1b9c8f8-6829-4eaf-b5bf-48121eaaf818" containerID="9fc523848c08f25598fae2f5e3fad537b2578bef89b29e3368d51587809c9183" exitCode=137 Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.050970 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6cf59496f7-6ntrd" Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.051051 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6cf59496f7-6ntrd" event={"ID":"d1b9c8f8-6829-4eaf-b5bf-48121eaaf818","Type":"ContainerDied","Data":"3f4ac0f91d02f64d61272851b2a1ac47fab311746345b16f0c2110dbf11b51b0"} Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.051087 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6cf59496f7-6ntrd" event={"ID":"d1b9c8f8-6829-4eaf-b5bf-48121eaaf818","Type":"ContainerDied","Data":"9fc523848c08f25598fae2f5e3fad537b2578bef89b29e3368d51587809c9183"} Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.051100 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6cf59496f7-6ntrd" event={"ID":"d1b9c8f8-6829-4eaf-b5bf-48121eaaf818","Type":"ContainerDied","Data":"53baeb1e2441390fb1afebdf726ab4dd0cf54f830b271eb797ad7942b592192d"} Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.051141 4955 scope.go:117] "RemoveContainer" containerID="3f4ac0f91d02f64d61272851b2a1ac47fab311746345b16f0c2110dbf11b51b0" Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.077697 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7789497c74-pgff5" event={"ID":"8468e621-9a50-49f7-ba03-996bc8564f43","Type":"ContainerStarted","Data":"df3ca47c5c6611a7cb4f68b2c0325863889b343e0a032d42c1d14fe5444a2a6c"} Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.077741 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7789497c74-pgff5" event={"ID":"8468e621-9a50-49f7-ba03-996bc8564f43","Type":"ContainerStarted","Data":"23b7dd6d99878f1b11989a1799dce9d6e52cda10e61bd7f2f4aa862505a93254"} Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.080052 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-6ff67b7c6c-9dcnd" podStartSLOduration=3.194590528 podStartE2EDuration="6.080029763s" podCreationTimestamp="2026-02-17 13:23:12 +0000 UTC" firstStartedPulling="2026-02-17 13:23:13.738912555 +0000 UTC m=+1132.261642098" lastFinishedPulling="2026-02-17 13:23:16.62435179 +0000 UTC m=+1135.147081333" observedRunningTime="2026-02-17 13:23:18.05959522 +0000 UTC m=+1136.582324763" watchObservedRunningTime="2026-02-17 13:23:18.080029763 +0000 UTC m=+1136.602759306" Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.097020 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"99f00e22-f2b5-4e1f-b00a-9e50a1ed359e","Type":"ContainerStarted","Data":"668d903f87d617139f441377187fff1074088f120f5b14ef8c8dc4dc0c9c2fac"} Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.147434 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-7789497c74-pgff5" podStartSLOduration=3.292512701 podStartE2EDuration="6.147321682s" podCreationTimestamp="2026-02-17 13:23:12 +0000 UTC" firstStartedPulling="2026-02-17 13:23:13.761567801 +0000 UTC m=+1132.284297344" lastFinishedPulling="2026-02-17 13:23:16.616376782 +0000 UTC m=+1135.139106325" observedRunningTime="2026-02-17 13:23:18.122166955 +0000 UTC m=+1136.644896498" watchObservedRunningTime="2026-02-17 13:23:18.147321682 +0000 UTC m=+1136.670051225" Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.198877 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6cf59496f7-6ntrd"] Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.208184 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6cf59496f7-6ntrd"] Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.239283 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01547c5a-5ca7-435b-ba75-52e075800cf3" path="/var/lib/kubelet/pods/01547c5a-5ca7-435b-ba75-52e075800cf3/volumes" Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.240261 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2793f42e-26b2-4115-9ac5-9b6c1f503218" path="/var/lib/kubelet/pods/2793f42e-26b2-4115-9ac5-9b6c1f503218/volumes" Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.241272 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1b9c8f8-6829-4eaf-b5bf-48121eaaf818" path="/var/lib/kubelet/pods/d1b9c8f8-6829-4eaf-b5bf-48121eaaf818/volumes" Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.336046 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.348050 4955 scope.go:117] "RemoveContainer" containerID="9fc523848c08f25598fae2f5e3fad537b2578bef89b29e3368d51587809c9183" Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.415372 4955 scope.go:117] "RemoveContainer" containerID="3f4ac0f91d02f64d61272851b2a1ac47fab311746345b16f0c2110dbf11b51b0" Feb 17 13:23:18 crc kubenswrapper[4955]: E0217 13:23:18.417126 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f4ac0f91d02f64d61272851b2a1ac47fab311746345b16f0c2110dbf11b51b0\": container with ID starting with 3f4ac0f91d02f64d61272851b2a1ac47fab311746345b16f0c2110dbf11b51b0 not found: ID does not exist" containerID="3f4ac0f91d02f64d61272851b2a1ac47fab311746345b16f0c2110dbf11b51b0" Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.417164 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f4ac0f91d02f64d61272851b2a1ac47fab311746345b16f0c2110dbf11b51b0"} err="failed to get container status \"3f4ac0f91d02f64d61272851b2a1ac47fab311746345b16f0c2110dbf11b51b0\": rpc error: code = NotFound desc = could not find container \"3f4ac0f91d02f64d61272851b2a1ac47fab311746345b16f0c2110dbf11b51b0\": container with ID starting with 3f4ac0f91d02f64d61272851b2a1ac47fab311746345b16f0c2110dbf11b51b0 not found: ID does not exist" Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.417207 4955 scope.go:117] "RemoveContainer" containerID="9fc523848c08f25598fae2f5e3fad537b2578bef89b29e3368d51587809c9183" Feb 17 13:23:18 crc kubenswrapper[4955]: E0217 13:23:18.433360 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fc523848c08f25598fae2f5e3fad537b2578bef89b29e3368d51587809c9183\": container with ID starting with 9fc523848c08f25598fae2f5e3fad537b2578bef89b29e3368d51587809c9183 not found: ID does not exist" containerID="9fc523848c08f25598fae2f5e3fad537b2578bef89b29e3368d51587809c9183" Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.433425 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fc523848c08f25598fae2f5e3fad537b2578bef89b29e3368d51587809c9183"} err="failed to get container status \"9fc523848c08f25598fae2f5e3fad537b2578bef89b29e3368d51587809c9183\": rpc error: code = NotFound desc = could not find container \"9fc523848c08f25598fae2f5e3fad537b2578bef89b29e3368d51587809c9183\": container with ID starting with 9fc523848c08f25598fae2f5e3fad537b2578bef89b29e3368d51587809c9183 not found: ID does not exist" Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.433473 4955 scope.go:117] "RemoveContainer" containerID="3f4ac0f91d02f64d61272851b2a1ac47fab311746345b16f0c2110dbf11b51b0" Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.434427 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f4ac0f91d02f64d61272851b2a1ac47fab311746345b16f0c2110dbf11b51b0"} err="failed to get container status \"3f4ac0f91d02f64d61272851b2a1ac47fab311746345b16f0c2110dbf11b51b0\": rpc error: code = NotFound desc = could not find container \"3f4ac0f91d02f64d61272851b2a1ac47fab311746345b16f0c2110dbf11b51b0\": container with ID starting with 3f4ac0f91d02f64d61272851b2a1ac47fab311746345b16f0c2110dbf11b51b0 not found: ID does not exist" Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.434443 4955 scope.go:117] "RemoveContainer" containerID="9fc523848c08f25598fae2f5e3fad537b2578bef89b29e3368d51587809c9183" Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.434853 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fc523848c08f25598fae2f5e3fad537b2578bef89b29e3368d51587809c9183"} err="failed to get container status \"9fc523848c08f25598fae2f5e3fad537b2578bef89b29e3368d51587809c9183\": rpc error: code = NotFound desc = could not find container \"9fc523848c08f25598fae2f5e3fad537b2578bef89b29e3368d51587809c9183\": container with ID starting with 9fc523848c08f25598fae2f5e3fad537b2578bef89b29e3368d51587809c9183 not found: ID does not exist" Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.563872 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7fffbcf7cb-wr74b" Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.680816 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-5788cc4448-245bs" Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.685612 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5788cc4448-245bs"] Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.793972 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.896490 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-76cdc74c45-lss64" Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.947684 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd349c9d-f6b6-48b5-a22f-e2c1d81f6711-scripts\") pod \"dd349c9d-f6b6-48b5-a22f-e2c1d81f6711\" (UID: \"dd349c9d-f6b6-48b5-a22f-e2c1d81f6711\") " Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.947805 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dd349c9d-f6b6-48b5-a22f-e2c1d81f6711-config-data\") pod \"dd349c9d-f6b6-48b5-a22f-e2c1d81f6711\" (UID: \"dd349c9d-f6b6-48b5-a22f-e2c1d81f6711\") " Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.947837 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/dd349c9d-f6b6-48b5-a22f-e2c1d81f6711-horizon-secret-key\") pod \"dd349c9d-f6b6-48b5-a22f-e2c1d81f6711\" (UID: \"dd349c9d-f6b6-48b5-a22f-e2c1d81f6711\") " Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.947940 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2k65n\" (UniqueName: \"kubernetes.io/projected/dd349c9d-f6b6-48b5-a22f-e2c1d81f6711-kube-api-access-2k65n\") pod \"dd349c9d-f6b6-48b5-a22f-e2c1d81f6711\" (UID: \"dd349c9d-f6b6-48b5-a22f-e2c1d81f6711\") " Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.948096 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd349c9d-f6b6-48b5-a22f-e2c1d81f6711-logs\") pod \"dd349c9d-f6b6-48b5-a22f-e2c1d81f6711\" (UID: \"dd349c9d-f6b6-48b5-a22f-e2c1d81f6711\") " Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.948968 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd349c9d-f6b6-48b5-a22f-e2c1d81f6711-logs" (OuterVolumeSpecName: "logs") pod "dd349c9d-f6b6-48b5-a22f-e2c1d81f6711" (UID: "dd349c9d-f6b6-48b5-a22f-e2c1d81f6711"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.956705 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd349c9d-f6b6-48b5-a22f-e2c1d81f6711-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "dd349c9d-f6b6-48b5-a22f-e2c1d81f6711" (UID: "dd349c9d-f6b6-48b5-a22f-e2c1d81f6711"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.959008 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd349c9d-f6b6-48b5-a22f-e2c1d81f6711-kube-api-access-2k65n" (OuterVolumeSpecName: "kube-api-access-2k65n") pod "dd349c9d-f6b6-48b5-a22f-e2c1d81f6711" (UID: "dd349c9d-f6b6-48b5-a22f-e2c1d81f6711"). InnerVolumeSpecName "kube-api-access-2k65n". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:23:18 crc kubenswrapper[4955]: I0217 13:23:18.991957 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd349c9d-f6b6-48b5-a22f-e2c1d81f6711-config-data" (OuterVolumeSpecName: "config-data") pod "dd349c9d-f6b6-48b5-a22f-e2c1d81f6711" (UID: "dd349c9d-f6b6-48b5-a22f-e2c1d81f6711"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.020372 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd349c9d-f6b6-48b5-a22f-e2c1d81f6711-scripts" (OuterVolumeSpecName: "scripts") pod "dd349c9d-f6b6-48b5-a22f-e2c1d81f6711" (UID: "dd349c9d-f6b6-48b5-a22f-e2c1d81f6711"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.049530 4955 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/dd349c9d-f6b6-48b5-a22f-e2c1d81f6711-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.049563 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2k65n\" (UniqueName: \"kubernetes.io/projected/dd349c9d-f6b6-48b5-a22f-e2c1d81f6711-kube-api-access-2k65n\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.049579 4955 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd349c9d-f6b6-48b5-a22f-e2c1d81f6711-logs\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.049593 4955 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd349c9d-f6b6-48b5-a22f-e2c1d81f6711-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.049604 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dd349c9d-f6b6-48b5-a22f-e2c1d81f6711-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.193500 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-hvvzs" event={"ID":"29f9dab6-c497-4a2b-a39c-0c58adae2aa1","Type":"ContainerStarted","Data":"ed9ae269161b452111acb8fe7e3cac0476b8ea07308d6f3e4c8d0072aacdca7e"} Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.193853 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6578955fd5-hvvzs" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.209911 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"96771c04-ac0c-4eb8-8aa9-d59b754617d4","Type":"ContainerStarted","Data":"924535e4f8c48b4e7872940dd0a8e1ca12e8b1f69d0012cb24a3e0fb3ae89d14"} Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.214986 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.223799 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6578955fd5-hvvzs" podStartSLOduration=5.223763397 podStartE2EDuration="5.223763397s" podCreationTimestamp="2026-02-17 13:23:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:23:19.21227682 +0000 UTC m=+1137.735006363" watchObservedRunningTime="2026-02-17 13:23:19.223763397 +0000 UTC m=+1137.746492940" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.249575 4955 generic.go:334] "Generic (PLEG): container finished" podID="dd349c9d-f6b6-48b5-a22f-e2c1d81f6711" containerID="be4213dde918fe6799bdc14d9565d677ae84d633e730632a7b5c4918dff901ca" exitCode=137 Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.249884 4955 generic.go:334] "Generic (PLEG): container finished" podID="dd349c9d-f6b6-48b5-a22f-e2c1d81f6711" containerID="2bd1a1e5ef749430ee242be08db7aa848aff639e569b11a60f2d47e7252bc145" exitCode=137 Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.249795 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-76cdc74c45-lss64" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.251270 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.251243691 podStartE2EDuration="5.251243691s" podCreationTimestamp="2026-02-17 13:23:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:23:19.238615491 +0000 UTC m=+1137.761345034" watchObservedRunningTime="2026-02-17 13:23:19.251243691 +0000 UTC m=+1137.773973234" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.249812 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-76cdc74c45-lss64" event={"ID":"dd349c9d-f6b6-48b5-a22f-e2c1d81f6711","Type":"ContainerDied","Data":"be4213dde918fe6799bdc14d9565d677ae84d633e730632a7b5c4918dff901ca"} Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.252808 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-76cdc74c45-lss64" event={"ID":"dd349c9d-f6b6-48b5-a22f-e2c1d81f6711","Type":"ContainerDied","Data":"2bd1a1e5ef749430ee242be08db7aa848aff639e569b11a60f2d47e7252bc145"} Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.252837 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-76cdc74c45-lss64" event={"ID":"dd349c9d-f6b6-48b5-a22f-e2c1d81f6711","Type":"ContainerDied","Data":"25358c9f4c2ef72c3ffa95ec983ed5222c638d8f3bc40656d219b1179d9404cf"} Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.252872 4955 scope.go:117] "RemoveContainer" containerID="be4213dde918fe6799bdc14d9565d677ae84d633e730632a7b5c4918dff901ca" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.276203 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d20f3280-ab79-434f-a89a-11d6fbfc0edf","Type":"ContainerStarted","Data":"74359edeb30c4e7e7464c3a8d030d1e01cfdbd7f32b7d2c458c9d18b55ebb128"} Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.276686 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5788cc4448-245bs" podUID="1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf" containerName="horizon-log" containerID="cri-o://32e18fff4a97bffae1a71c95b67ec8d234531f3a2c24d149ec79a131c6c4a75a" gracePeriod=30 Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.277404 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5788cc4448-245bs" podUID="1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf" containerName="horizon" containerID="cri-o://52cdaba09384ae938ed1a6b17e232b9095c1468ccdd7386221ad6a01d6be8ff4" gracePeriod=30 Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.383285 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-76cdc74c45-lss64"] Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.405397 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-76cdc74c45-lss64"] Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.430834 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-57f5fb6b66-7vp2j"] Feb 17 13:23:19 crc kubenswrapper[4955]: E0217 13:23:19.436457 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1b9c8f8-6829-4eaf-b5bf-48121eaaf818" containerName="horizon-log" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.436492 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1b9c8f8-6829-4eaf-b5bf-48121eaaf818" containerName="horizon-log" Feb 17 13:23:19 crc kubenswrapper[4955]: E0217 13:23:19.436507 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1b9c8f8-6829-4eaf-b5bf-48121eaaf818" containerName="horizon" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.436513 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1b9c8f8-6829-4eaf-b5bf-48121eaaf818" containerName="horizon" Feb 17 13:23:19 crc kubenswrapper[4955]: E0217 13:23:19.436530 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd349c9d-f6b6-48b5-a22f-e2c1d81f6711" containerName="horizon-log" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.436536 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd349c9d-f6b6-48b5-a22f-e2c1d81f6711" containerName="horizon-log" Feb 17 13:23:19 crc kubenswrapper[4955]: E0217 13:23:19.436554 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd349c9d-f6b6-48b5-a22f-e2c1d81f6711" containerName="horizon" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.436560 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd349c9d-f6b6-48b5-a22f-e2c1d81f6711" containerName="horizon" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.439646 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1b9c8f8-6829-4eaf-b5bf-48121eaaf818" containerName="horizon" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.439814 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd349c9d-f6b6-48b5-a22f-e2c1d81f6711" containerName="horizon-log" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.439835 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1b9c8f8-6829-4eaf-b5bf-48121eaaf818" containerName="horizon-log" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.439853 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd349c9d-f6b6-48b5-a22f-e2c1d81f6711" containerName="horizon" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.440903 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-57f5fb6b66-7vp2j"] Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.441003 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-57f5fb6b66-7vp2j" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.446252 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.446489 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.465674 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87c5d914-d03a-41e9-9b1c-8c34a5e0758b-combined-ca-bundle\") pod \"barbican-api-57f5fb6b66-7vp2j\" (UID: \"87c5d914-d03a-41e9-9b1c-8c34a5e0758b\") " pod="openstack/barbican-api-57f5fb6b66-7vp2j" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.465717 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdvsk\" (UniqueName: \"kubernetes.io/projected/87c5d914-d03a-41e9-9b1c-8c34a5e0758b-kube-api-access-kdvsk\") pod \"barbican-api-57f5fb6b66-7vp2j\" (UID: \"87c5d914-d03a-41e9-9b1c-8c34a5e0758b\") " pod="openstack/barbican-api-57f5fb6b66-7vp2j" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.465804 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87c5d914-d03a-41e9-9b1c-8c34a5e0758b-config-data\") pod \"barbican-api-57f5fb6b66-7vp2j\" (UID: \"87c5d914-d03a-41e9-9b1c-8c34a5e0758b\") " pod="openstack/barbican-api-57f5fb6b66-7vp2j" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.465821 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/87c5d914-d03a-41e9-9b1c-8c34a5e0758b-public-tls-certs\") pod \"barbican-api-57f5fb6b66-7vp2j\" (UID: \"87c5d914-d03a-41e9-9b1c-8c34a5e0758b\") " pod="openstack/barbican-api-57f5fb6b66-7vp2j" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.465855 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87c5d914-d03a-41e9-9b1c-8c34a5e0758b-logs\") pod \"barbican-api-57f5fb6b66-7vp2j\" (UID: \"87c5d914-d03a-41e9-9b1c-8c34a5e0758b\") " pod="openstack/barbican-api-57f5fb6b66-7vp2j" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.465899 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87c5d914-d03a-41e9-9b1c-8c34a5e0758b-internal-tls-certs\") pod \"barbican-api-57f5fb6b66-7vp2j\" (UID: \"87c5d914-d03a-41e9-9b1c-8c34a5e0758b\") " pod="openstack/barbican-api-57f5fb6b66-7vp2j" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.465939 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/87c5d914-d03a-41e9-9b1c-8c34a5e0758b-config-data-custom\") pod \"barbican-api-57f5fb6b66-7vp2j\" (UID: \"87c5d914-d03a-41e9-9b1c-8c34a5e0758b\") " pod="openstack/barbican-api-57f5fb6b66-7vp2j" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.536506 4955 scope.go:117] "RemoveContainer" containerID="2bd1a1e5ef749430ee242be08db7aa848aff639e569b11a60f2d47e7252bc145" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.562276 4955 scope.go:117] "RemoveContainer" containerID="be4213dde918fe6799bdc14d9565d677ae84d633e730632a7b5c4918dff901ca" Feb 17 13:23:19 crc kubenswrapper[4955]: E0217 13:23:19.563332 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be4213dde918fe6799bdc14d9565d677ae84d633e730632a7b5c4918dff901ca\": container with ID starting with be4213dde918fe6799bdc14d9565d677ae84d633e730632a7b5c4918dff901ca not found: ID does not exist" containerID="be4213dde918fe6799bdc14d9565d677ae84d633e730632a7b5c4918dff901ca" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.563369 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be4213dde918fe6799bdc14d9565d677ae84d633e730632a7b5c4918dff901ca"} err="failed to get container status \"be4213dde918fe6799bdc14d9565d677ae84d633e730632a7b5c4918dff901ca\": rpc error: code = NotFound desc = could not find container \"be4213dde918fe6799bdc14d9565d677ae84d633e730632a7b5c4918dff901ca\": container with ID starting with be4213dde918fe6799bdc14d9565d677ae84d633e730632a7b5c4918dff901ca not found: ID does not exist" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.563410 4955 scope.go:117] "RemoveContainer" containerID="2bd1a1e5ef749430ee242be08db7aa848aff639e569b11a60f2d47e7252bc145" Feb 17 13:23:19 crc kubenswrapper[4955]: E0217 13:23:19.563676 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2bd1a1e5ef749430ee242be08db7aa848aff639e569b11a60f2d47e7252bc145\": container with ID starting with 2bd1a1e5ef749430ee242be08db7aa848aff639e569b11a60f2d47e7252bc145 not found: ID does not exist" containerID="2bd1a1e5ef749430ee242be08db7aa848aff639e569b11a60f2d47e7252bc145" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.563705 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2bd1a1e5ef749430ee242be08db7aa848aff639e569b11a60f2d47e7252bc145"} err="failed to get container status \"2bd1a1e5ef749430ee242be08db7aa848aff639e569b11a60f2d47e7252bc145\": rpc error: code = NotFound desc = could not find container \"2bd1a1e5ef749430ee242be08db7aa848aff639e569b11a60f2d47e7252bc145\": container with ID starting with 2bd1a1e5ef749430ee242be08db7aa848aff639e569b11a60f2d47e7252bc145 not found: ID does not exist" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.563724 4955 scope.go:117] "RemoveContainer" containerID="be4213dde918fe6799bdc14d9565d677ae84d633e730632a7b5c4918dff901ca" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.564162 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be4213dde918fe6799bdc14d9565d677ae84d633e730632a7b5c4918dff901ca"} err="failed to get container status \"be4213dde918fe6799bdc14d9565d677ae84d633e730632a7b5c4918dff901ca\": rpc error: code = NotFound desc = could not find container \"be4213dde918fe6799bdc14d9565d677ae84d633e730632a7b5c4918dff901ca\": container with ID starting with be4213dde918fe6799bdc14d9565d677ae84d633e730632a7b5c4918dff901ca not found: ID does not exist" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.564204 4955 scope.go:117] "RemoveContainer" containerID="2bd1a1e5ef749430ee242be08db7aa848aff639e569b11a60f2d47e7252bc145" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.564466 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2bd1a1e5ef749430ee242be08db7aa848aff639e569b11a60f2d47e7252bc145"} err="failed to get container status \"2bd1a1e5ef749430ee242be08db7aa848aff639e569b11a60f2d47e7252bc145\": rpc error: code = NotFound desc = could not find container \"2bd1a1e5ef749430ee242be08db7aa848aff639e569b11a60f2d47e7252bc145\": container with ID starting with 2bd1a1e5ef749430ee242be08db7aa848aff639e569b11a60f2d47e7252bc145 not found: ID does not exist" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.567673 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87c5d914-d03a-41e9-9b1c-8c34a5e0758b-config-data\") pod \"barbican-api-57f5fb6b66-7vp2j\" (UID: \"87c5d914-d03a-41e9-9b1c-8c34a5e0758b\") " pod="openstack/barbican-api-57f5fb6b66-7vp2j" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.567717 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/87c5d914-d03a-41e9-9b1c-8c34a5e0758b-public-tls-certs\") pod \"barbican-api-57f5fb6b66-7vp2j\" (UID: \"87c5d914-d03a-41e9-9b1c-8c34a5e0758b\") " pod="openstack/barbican-api-57f5fb6b66-7vp2j" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.567750 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87c5d914-d03a-41e9-9b1c-8c34a5e0758b-logs\") pod \"barbican-api-57f5fb6b66-7vp2j\" (UID: \"87c5d914-d03a-41e9-9b1c-8c34a5e0758b\") " pod="openstack/barbican-api-57f5fb6b66-7vp2j" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.567807 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87c5d914-d03a-41e9-9b1c-8c34a5e0758b-internal-tls-certs\") pod \"barbican-api-57f5fb6b66-7vp2j\" (UID: \"87c5d914-d03a-41e9-9b1c-8c34a5e0758b\") " pod="openstack/barbican-api-57f5fb6b66-7vp2j" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.567848 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/87c5d914-d03a-41e9-9b1c-8c34a5e0758b-config-data-custom\") pod \"barbican-api-57f5fb6b66-7vp2j\" (UID: \"87c5d914-d03a-41e9-9b1c-8c34a5e0758b\") " pod="openstack/barbican-api-57f5fb6b66-7vp2j" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.567928 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87c5d914-d03a-41e9-9b1c-8c34a5e0758b-combined-ca-bundle\") pod \"barbican-api-57f5fb6b66-7vp2j\" (UID: \"87c5d914-d03a-41e9-9b1c-8c34a5e0758b\") " pod="openstack/barbican-api-57f5fb6b66-7vp2j" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.567952 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdvsk\" (UniqueName: \"kubernetes.io/projected/87c5d914-d03a-41e9-9b1c-8c34a5e0758b-kube-api-access-kdvsk\") pod \"barbican-api-57f5fb6b66-7vp2j\" (UID: \"87c5d914-d03a-41e9-9b1c-8c34a5e0758b\") " pod="openstack/barbican-api-57f5fb6b66-7vp2j" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.568575 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87c5d914-d03a-41e9-9b1c-8c34a5e0758b-logs\") pod \"barbican-api-57f5fb6b66-7vp2j\" (UID: \"87c5d914-d03a-41e9-9b1c-8c34a5e0758b\") " pod="openstack/barbican-api-57f5fb6b66-7vp2j" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.572609 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/87c5d914-d03a-41e9-9b1c-8c34a5e0758b-config-data-custom\") pod \"barbican-api-57f5fb6b66-7vp2j\" (UID: \"87c5d914-d03a-41e9-9b1c-8c34a5e0758b\") " pod="openstack/barbican-api-57f5fb6b66-7vp2j" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.575578 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87c5d914-d03a-41e9-9b1c-8c34a5e0758b-config-data\") pod \"barbican-api-57f5fb6b66-7vp2j\" (UID: \"87c5d914-d03a-41e9-9b1c-8c34a5e0758b\") " pod="openstack/barbican-api-57f5fb6b66-7vp2j" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.576513 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/87c5d914-d03a-41e9-9b1c-8c34a5e0758b-public-tls-certs\") pod \"barbican-api-57f5fb6b66-7vp2j\" (UID: \"87c5d914-d03a-41e9-9b1c-8c34a5e0758b\") " pod="openstack/barbican-api-57f5fb6b66-7vp2j" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.578040 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87c5d914-d03a-41e9-9b1c-8c34a5e0758b-combined-ca-bundle\") pod \"barbican-api-57f5fb6b66-7vp2j\" (UID: \"87c5d914-d03a-41e9-9b1c-8c34a5e0758b\") " pod="openstack/barbican-api-57f5fb6b66-7vp2j" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.582463 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87c5d914-d03a-41e9-9b1c-8c34a5e0758b-internal-tls-certs\") pod \"barbican-api-57f5fb6b66-7vp2j\" (UID: \"87c5d914-d03a-41e9-9b1c-8c34a5e0758b\") " pod="openstack/barbican-api-57f5fb6b66-7vp2j" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.585253 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdvsk\" (UniqueName: \"kubernetes.io/projected/87c5d914-d03a-41e9-9b1c-8c34a5e0758b-kube-api-access-kdvsk\") pod \"barbican-api-57f5fb6b66-7vp2j\" (UID: \"87c5d914-d03a-41e9-9b1c-8c34a5e0758b\") " pod="openstack/barbican-api-57f5fb6b66-7vp2j" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.676931 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-746fcf4988-4q7tw" Feb 17 13:23:19 crc kubenswrapper[4955]: I0217 13:23:19.785395 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-57f5fb6b66-7vp2j" Feb 17 13:23:20 crc kubenswrapper[4955]: I0217 13:23:20.022977 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-784f78c585-7lvx4"] Feb 17 13:23:20 crc kubenswrapper[4955]: I0217 13:23:20.023498 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-784f78c585-7lvx4" podUID="e08d2b73-a720-4a28-bdc0-d8e8008417af" containerName="neutron-api" containerID="cri-o://4c25f2067734abe7fc2878ff092cd406523ac1cd49c75130c410877ed3c8a5bb" gracePeriod=30 Feb 17 13:23:20 crc kubenswrapper[4955]: I0217 13:23:20.024152 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-784f78c585-7lvx4" podUID="e08d2b73-a720-4a28-bdc0-d8e8008417af" containerName="neutron-httpd" containerID="cri-o://edfe665cc4bfb8b12efc35a155058e0faa141827cfcb336ebc28576eb4699582" gracePeriod=30 Feb 17 13:23:20 crc kubenswrapper[4955]: I0217 13:23:20.057219 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7c5ddc6895-552ts"] Feb 17 13:23:20 crc kubenswrapper[4955]: I0217 13:23:20.058913 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7c5ddc6895-552ts" Feb 17 13:23:20 crc kubenswrapper[4955]: I0217 13:23:20.077260 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25c07d2a-4f61-48fc-88e3-0ed671ad9cff-combined-ca-bundle\") pod \"neutron-7c5ddc6895-552ts\" (UID: \"25c07d2a-4f61-48fc-88e3-0ed671ad9cff\") " pod="openstack/neutron-7c5ddc6895-552ts" Feb 17 13:23:20 crc kubenswrapper[4955]: I0217 13:23:20.077357 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/25c07d2a-4f61-48fc-88e3-0ed671ad9cff-httpd-config\") pod \"neutron-7c5ddc6895-552ts\" (UID: \"25c07d2a-4f61-48fc-88e3-0ed671ad9cff\") " pod="openstack/neutron-7c5ddc6895-552ts" Feb 17 13:23:20 crc kubenswrapper[4955]: I0217 13:23:20.077381 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/25c07d2a-4f61-48fc-88e3-0ed671ad9cff-public-tls-certs\") pod \"neutron-7c5ddc6895-552ts\" (UID: \"25c07d2a-4f61-48fc-88e3-0ed671ad9cff\") " pod="openstack/neutron-7c5ddc6895-552ts" Feb 17 13:23:20 crc kubenswrapper[4955]: I0217 13:23:20.077414 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/25c07d2a-4f61-48fc-88e3-0ed671ad9cff-internal-tls-certs\") pod \"neutron-7c5ddc6895-552ts\" (UID: \"25c07d2a-4f61-48fc-88e3-0ed671ad9cff\") " pod="openstack/neutron-7c5ddc6895-552ts" Feb 17 13:23:20 crc kubenswrapper[4955]: I0217 13:23:20.081637 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/25c07d2a-4f61-48fc-88e3-0ed671ad9cff-config\") pod \"neutron-7c5ddc6895-552ts\" (UID: \"25c07d2a-4f61-48fc-88e3-0ed671ad9cff\") " pod="openstack/neutron-7c5ddc6895-552ts" Feb 17 13:23:20 crc kubenswrapper[4955]: I0217 13:23:20.081693 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjgbk\" (UniqueName: \"kubernetes.io/projected/25c07d2a-4f61-48fc-88e3-0ed671ad9cff-kube-api-access-jjgbk\") pod \"neutron-7c5ddc6895-552ts\" (UID: \"25c07d2a-4f61-48fc-88e3-0ed671ad9cff\") " pod="openstack/neutron-7c5ddc6895-552ts" Feb 17 13:23:20 crc kubenswrapper[4955]: I0217 13:23:20.081852 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/25c07d2a-4f61-48fc-88e3-0ed671ad9cff-ovndb-tls-certs\") pod \"neutron-7c5ddc6895-552ts\" (UID: \"25c07d2a-4f61-48fc-88e3-0ed671ad9cff\") " pod="openstack/neutron-7c5ddc6895-552ts" Feb 17 13:23:20 crc kubenswrapper[4955]: I0217 13:23:20.089731 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7c5ddc6895-552ts"] Feb 17 13:23:20 crc kubenswrapper[4955]: I0217 13:23:20.147028 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-784f78c585-7lvx4" podUID="e08d2b73-a720-4a28-bdc0-d8e8008417af" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.154:9696/\": read tcp 10.217.0.2:51764->10.217.0.154:9696: read: connection reset by peer" Feb 17 13:23:20 crc kubenswrapper[4955]: I0217 13:23:20.183240 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25c07d2a-4f61-48fc-88e3-0ed671ad9cff-combined-ca-bundle\") pod \"neutron-7c5ddc6895-552ts\" (UID: \"25c07d2a-4f61-48fc-88e3-0ed671ad9cff\") " pod="openstack/neutron-7c5ddc6895-552ts" Feb 17 13:23:20 crc kubenswrapper[4955]: I0217 13:23:20.183536 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/25c07d2a-4f61-48fc-88e3-0ed671ad9cff-httpd-config\") pod \"neutron-7c5ddc6895-552ts\" (UID: \"25c07d2a-4f61-48fc-88e3-0ed671ad9cff\") " pod="openstack/neutron-7c5ddc6895-552ts" Feb 17 13:23:20 crc kubenswrapper[4955]: I0217 13:23:20.183670 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/25c07d2a-4f61-48fc-88e3-0ed671ad9cff-public-tls-certs\") pod \"neutron-7c5ddc6895-552ts\" (UID: \"25c07d2a-4f61-48fc-88e3-0ed671ad9cff\") " pod="openstack/neutron-7c5ddc6895-552ts" Feb 17 13:23:20 crc kubenswrapper[4955]: I0217 13:23:20.183803 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/25c07d2a-4f61-48fc-88e3-0ed671ad9cff-internal-tls-certs\") pod \"neutron-7c5ddc6895-552ts\" (UID: \"25c07d2a-4f61-48fc-88e3-0ed671ad9cff\") " pod="openstack/neutron-7c5ddc6895-552ts" Feb 17 13:23:20 crc kubenswrapper[4955]: I0217 13:23:20.183950 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/25c07d2a-4f61-48fc-88e3-0ed671ad9cff-config\") pod \"neutron-7c5ddc6895-552ts\" (UID: \"25c07d2a-4f61-48fc-88e3-0ed671ad9cff\") " pod="openstack/neutron-7c5ddc6895-552ts" Feb 17 13:23:20 crc kubenswrapper[4955]: I0217 13:23:20.184039 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjgbk\" (UniqueName: \"kubernetes.io/projected/25c07d2a-4f61-48fc-88e3-0ed671ad9cff-kube-api-access-jjgbk\") pod \"neutron-7c5ddc6895-552ts\" (UID: \"25c07d2a-4f61-48fc-88e3-0ed671ad9cff\") " pod="openstack/neutron-7c5ddc6895-552ts" Feb 17 13:23:20 crc kubenswrapper[4955]: I0217 13:23:20.184187 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/25c07d2a-4f61-48fc-88e3-0ed671ad9cff-ovndb-tls-certs\") pod \"neutron-7c5ddc6895-552ts\" (UID: \"25c07d2a-4f61-48fc-88e3-0ed671ad9cff\") " pod="openstack/neutron-7c5ddc6895-552ts" Feb 17 13:23:20 crc kubenswrapper[4955]: I0217 13:23:20.191723 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/25c07d2a-4f61-48fc-88e3-0ed671ad9cff-internal-tls-certs\") pod \"neutron-7c5ddc6895-552ts\" (UID: \"25c07d2a-4f61-48fc-88e3-0ed671ad9cff\") " pod="openstack/neutron-7c5ddc6895-552ts" Feb 17 13:23:20 crc kubenswrapper[4955]: I0217 13:23:20.191774 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/25c07d2a-4f61-48fc-88e3-0ed671ad9cff-httpd-config\") pod \"neutron-7c5ddc6895-552ts\" (UID: \"25c07d2a-4f61-48fc-88e3-0ed671ad9cff\") " pod="openstack/neutron-7c5ddc6895-552ts" Feb 17 13:23:20 crc kubenswrapper[4955]: I0217 13:23:20.192034 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/25c07d2a-4f61-48fc-88e3-0ed671ad9cff-config\") pod \"neutron-7c5ddc6895-552ts\" (UID: \"25c07d2a-4f61-48fc-88e3-0ed671ad9cff\") " pod="openstack/neutron-7c5ddc6895-552ts" Feb 17 13:23:20 crc kubenswrapper[4955]: I0217 13:23:20.192135 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25c07d2a-4f61-48fc-88e3-0ed671ad9cff-combined-ca-bundle\") pod \"neutron-7c5ddc6895-552ts\" (UID: \"25c07d2a-4f61-48fc-88e3-0ed671ad9cff\") " pod="openstack/neutron-7c5ddc6895-552ts" Feb 17 13:23:20 crc kubenswrapper[4955]: I0217 13:23:20.192508 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/25c07d2a-4f61-48fc-88e3-0ed671ad9cff-ovndb-tls-certs\") pod \"neutron-7c5ddc6895-552ts\" (UID: \"25c07d2a-4f61-48fc-88e3-0ed671ad9cff\") " pod="openstack/neutron-7c5ddc6895-552ts" Feb 17 13:23:20 crc kubenswrapper[4955]: I0217 13:23:20.194422 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/25c07d2a-4f61-48fc-88e3-0ed671ad9cff-public-tls-certs\") pod \"neutron-7c5ddc6895-552ts\" (UID: \"25c07d2a-4f61-48fc-88e3-0ed671ad9cff\") " pod="openstack/neutron-7c5ddc6895-552ts" Feb 17 13:23:20 crc kubenswrapper[4955]: I0217 13:23:20.208221 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjgbk\" (UniqueName: \"kubernetes.io/projected/25c07d2a-4f61-48fc-88e3-0ed671ad9cff-kube-api-access-jjgbk\") pod \"neutron-7c5ddc6895-552ts\" (UID: \"25c07d2a-4f61-48fc-88e3-0ed671ad9cff\") " pod="openstack/neutron-7c5ddc6895-552ts" Feb 17 13:23:20 crc kubenswrapper[4955]: I0217 13:23:20.244135 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd349c9d-f6b6-48b5-a22f-e2c1d81f6711" path="/var/lib/kubelet/pods/dd349c9d-f6b6-48b5-a22f-e2c1d81f6711/volumes" Feb 17 13:23:20 crc kubenswrapper[4955]: I0217 13:23:20.284389 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"99f00e22-f2b5-4e1f-b00a-9e50a1ed359e","Type":"ContainerStarted","Data":"555092eb6d9a00102f0ef85ead43250fb502025f9d327672684751329d350428"} Feb 17 13:23:20 crc kubenswrapper[4955]: I0217 13:23:20.286736 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d20f3280-ab79-434f-a89a-11d6fbfc0edf","Type":"ContainerStarted","Data":"812e7e1ffda69c6175b2a09a089bce147530345353d05c3617de2af0ebb6f577"} Feb 17 13:23:20 crc kubenswrapper[4955]: I0217 13:23:20.286773 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d20f3280-ab79-434f-a89a-11d6fbfc0edf","Type":"ContainerStarted","Data":"e0798da9a93b55d7c2bbe07fbb48b498acc48e833fefe3eff9186170ec9f0299"} Feb 17 13:23:20 crc kubenswrapper[4955]: I0217 13:23:20.293189 4955 generic.go:334] "Generic (PLEG): container finished" podID="e08d2b73-a720-4a28-bdc0-d8e8008417af" containerID="edfe665cc4bfb8b12efc35a155058e0faa141827cfcb336ebc28576eb4699582" exitCode=0 Feb 17 13:23:20 crc kubenswrapper[4955]: I0217 13:23:20.293595 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="96771c04-ac0c-4eb8-8aa9-d59b754617d4" containerName="cinder-api-log" containerID="cri-o://7daf52da12552e9ff871d56fe0c83c8dfdafb7b7c497ab0251922a2dbc6dab46" gracePeriod=30 Feb 17 13:23:20 crc kubenswrapper[4955]: I0217 13:23:20.293908 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-784f78c585-7lvx4" event={"ID":"e08d2b73-a720-4a28-bdc0-d8e8008417af","Type":"ContainerDied","Data":"edfe665cc4bfb8b12efc35a155058e0faa141827cfcb336ebc28576eb4699582"} Feb 17 13:23:20 crc kubenswrapper[4955]: I0217 13:23:20.294893 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="96771c04-ac0c-4eb8-8aa9-d59b754617d4" containerName="cinder-api" containerID="cri-o://924535e4f8c48b4e7872940dd0a8e1ca12e8b1f69d0012cb24a3e0fb3ae89d14" gracePeriod=30 Feb 17 13:23:20 crc kubenswrapper[4955]: I0217 13:23:20.310093 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.765358592 podStartE2EDuration="6.310075123s" podCreationTimestamp="2026-02-17 13:23:14 +0000 UTC" firstStartedPulling="2026-02-17 13:23:15.335711973 +0000 UTC m=+1133.858441516" lastFinishedPulling="2026-02-17 13:23:16.880428504 +0000 UTC m=+1135.403158047" observedRunningTime="2026-02-17 13:23:20.30365892 +0000 UTC m=+1138.826388473" watchObservedRunningTime="2026-02-17 13:23:20.310075123 +0000 UTC m=+1138.832804656" Feb 17 13:23:20 crc kubenswrapper[4955]: I0217 13:23:20.341039 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-57f5fb6b66-7vp2j"] Feb 17 13:23:20 crc kubenswrapper[4955]: W0217 13:23:20.353438 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod87c5d914_d03a_41e9_9b1c_8c34a5e0758b.slice/crio-2313d767c6048d9e97aa8ce67e0e42c8f2ba73a15ec1d4b7c89f72f76f478140 WatchSource:0}: Error finding container 2313d767c6048d9e97aa8ce67e0e42c8f2ba73a15ec1d4b7c89f72f76f478140: Status 404 returned error can't find the container with id 2313d767c6048d9e97aa8ce67e0e42c8f2ba73a15ec1d4b7c89f72f76f478140 Feb 17 13:23:20 crc kubenswrapper[4955]: I0217 13:23:20.398291 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7c5ddc6895-552ts" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.069830 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7c5ddc6895-552ts"] Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.079854 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.099297 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96771c04-ac0c-4eb8-8aa9-d59b754617d4-config-data\") pod \"96771c04-ac0c-4eb8-8aa9-d59b754617d4\" (UID: \"96771c04-ac0c-4eb8-8aa9-d59b754617d4\") " Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.099383 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kz9th\" (UniqueName: \"kubernetes.io/projected/96771c04-ac0c-4eb8-8aa9-d59b754617d4-kube-api-access-kz9th\") pod \"96771c04-ac0c-4eb8-8aa9-d59b754617d4\" (UID: \"96771c04-ac0c-4eb8-8aa9-d59b754617d4\") " Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.099439 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96771c04-ac0c-4eb8-8aa9-d59b754617d4-logs\") pod \"96771c04-ac0c-4eb8-8aa9-d59b754617d4\" (UID: \"96771c04-ac0c-4eb8-8aa9-d59b754617d4\") " Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.099465 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/96771c04-ac0c-4eb8-8aa9-d59b754617d4-etc-machine-id\") pod \"96771c04-ac0c-4eb8-8aa9-d59b754617d4\" (UID: \"96771c04-ac0c-4eb8-8aa9-d59b754617d4\") " Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.099491 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/96771c04-ac0c-4eb8-8aa9-d59b754617d4-config-data-custom\") pod \"96771c04-ac0c-4eb8-8aa9-d59b754617d4\" (UID: \"96771c04-ac0c-4eb8-8aa9-d59b754617d4\") " Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.099509 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96771c04-ac0c-4eb8-8aa9-d59b754617d4-combined-ca-bundle\") pod \"96771c04-ac0c-4eb8-8aa9-d59b754617d4\" (UID: \"96771c04-ac0c-4eb8-8aa9-d59b754617d4\") " Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.099593 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96771c04-ac0c-4eb8-8aa9-d59b754617d4-scripts\") pod \"96771c04-ac0c-4eb8-8aa9-d59b754617d4\" (UID: \"96771c04-ac0c-4eb8-8aa9-d59b754617d4\") " Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.101871 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/96771c04-ac0c-4eb8-8aa9-d59b754617d4-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "96771c04-ac0c-4eb8-8aa9-d59b754617d4" (UID: "96771c04-ac0c-4eb8-8aa9-d59b754617d4"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.104797 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96771c04-ac0c-4eb8-8aa9-d59b754617d4-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "96771c04-ac0c-4eb8-8aa9-d59b754617d4" (UID: "96771c04-ac0c-4eb8-8aa9-d59b754617d4"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.105939 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96771c04-ac0c-4eb8-8aa9-d59b754617d4-logs" (OuterVolumeSpecName: "logs") pod "96771c04-ac0c-4eb8-8aa9-d59b754617d4" (UID: "96771c04-ac0c-4eb8-8aa9-d59b754617d4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.107907 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96771c04-ac0c-4eb8-8aa9-d59b754617d4-kube-api-access-kz9th" (OuterVolumeSpecName: "kube-api-access-kz9th") pod "96771c04-ac0c-4eb8-8aa9-d59b754617d4" (UID: "96771c04-ac0c-4eb8-8aa9-d59b754617d4"). InnerVolumeSpecName "kube-api-access-kz9th". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.128418 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96771c04-ac0c-4eb8-8aa9-d59b754617d4-scripts" (OuterVolumeSpecName: "scripts") pod "96771c04-ac0c-4eb8-8aa9-d59b754617d4" (UID: "96771c04-ac0c-4eb8-8aa9-d59b754617d4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.149556 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96771c04-ac0c-4eb8-8aa9-d59b754617d4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "96771c04-ac0c-4eb8-8aa9-d59b754617d4" (UID: "96771c04-ac0c-4eb8-8aa9-d59b754617d4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.169373 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96771c04-ac0c-4eb8-8aa9-d59b754617d4-config-data" (OuterVolumeSpecName: "config-data") pod "96771c04-ac0c-4eb8-8aa9-d59b754617d4" (UID: "96771c04-ac0c-4eb8-8aa9-d59b754617d4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.202021 4955 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96771c04-ac0c-4eb8-8aa9-d59b754617d4-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.202046 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96771c04-ac0c-4eb8-8aa9-d59b754617d4-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.202056 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kz9th\" (UniqueName: \"kubernetes.io/projected/96771c04-ac0c-4eb8-8aa9-d59b754617d4-kube-api-access-kz9th\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.202066 4955 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96771c04-ac0c-4eb8-8aa9-d59b754617d4-logs\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.202075 4955 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/96771c04-ac0c-4eb8-8aa9-d59b754617d4-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.202083 4955 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/96771c04-ac0c-4eb8-8aa9-d59b754617d4-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.202092 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96771c04-ac0c-4eb8-8aa9-d59b754617d4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.307144 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d20f3280-ab79-434f-a89a-11d6fbfc0edf","Type":"ContainerStarted","Data":"867f11c8b8d3d4d6f0679b281725a532c1d72c6f1cbca9ed91bfbb6f366fd041"} Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.309201 4955 generic.go:334] "Generic (PLEG): container finished" podID="96771c04-ac0c-4eb8-8aa9-d59b754617d4" containerID="924535e4f8c48b4e7872940dd0a8e1ca12e8b1f69d0012cb24a3e0fb3ae89d14" exitCode=0 Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.309246 4955 generic.go:334] "Generic (PLEG): container finished" podID="96771c04-ac0c-4eb8-8aa9-d59b754617d4" containerID="7daf52da12552e9ff871d56fe0c83c8dfdafb7b7c497ab0251922a2dbc6dab46" exitCode=143 Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.309376 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"96771c04-ac0c-4eb8-8aa9-d59b754617d4","Type":"ContainerDied","Data":"924535e4f8c48b4e7872940dd0a8e1ca12e8b1f69d0012cb24a3e0fb3ae89d14"} Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.309561 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"96771c04-ac0c-4eb8-8aa9-d59b754617d4","Type":"ContainerDied","Data":"7daf52da12552e9ff871d56fe0c83c8dfdafb7b7c497ab0251922a2dbc6dab46"} Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.309648 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"96771c04-ac0c-4eb8-8aa9-d59b754617d4","Type":"ContainerDied","Data":"61e20a02fc68bc5fb0b2d7985ccdd92bf0c4b8326b146fc59411e1856624f36c"} Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.309672 4955 scope.go:117] "RemoveContainer" containerID="924535e4f8c48b4e7872940dd0a8e1ca12e8b1f69d0012cb24a3e0fb3ae89d14" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.309901 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.344066 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7c5ddc6895-552ts" event={"ID":"25c07d2a-4f61-48fc-88e3-0ed671ad9cff","Type":"ContainerStarted","Data":"1a6866687d59a5f5dc8fe4c04a298d468795c97cfd8d1b75cf6bf1059db47d36"} Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.358228 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57f5fb6b66-7vp2j" event={"ID":"87c5d914-d03a-41e9-9b1c-8c34a5e0758b","Type":"ContainerStarted","Data":"609dbc65a3eff005e8d226c73e13022e3072a0394956e090e2f13ae60e80e4ed"} Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.358288 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57f5fb6b66-7vp2j" event={"ID":"87c5d914-d03a-41e9-9b1c-8c34a5e0758b","Type":"ContainerStarted","Data":"4af4ac2a97a640486a47385904abeea7100537d51e52fd57f5bc06274f6ad61a"} Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.358303 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57f5fb6b66-7vp2j" event={"ID":"87c5d914-d03a-41e9-9b1c-8c34a5e0758b","Type":"ContainerStarted","Data":"2313d767c6048d9e97aa8ce67e0e42c8f2ba73a15ec1d4b7c89f72f76f478140"} Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.370538 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.377511 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.431009 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Feb 17 13:23:21 crc kubenswrapper[4955]: E0217 13:23:21.431676 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96771c04-ac0c-4eb8-8aa9-d59b754617d4" containerName="cinder-api" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.431716 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="96771c04-ac0c-4eb8-8aa9-d59b754617d4" containerName="cinder-api" Feb 17 13:23:21 crc kubenswrapper[4955]: E0217 13:23:21.431798 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96771c04-ac0c-4eb8-8aa9-d59b754617d4" containerName="cinder-api-log" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.431807 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="96771c04-ac0c-4eb8-8aa9-d59b754617d4" containerName="cinder-api-log" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.432142 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="96771c04-ac0c-4eb8-8aa9-d59b754617d4" containerName="cinder-api-log" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.432170 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="96771c04-ac0c-4eb8-8aa9-d59b754617d4" containerName="cinder-api" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.433585 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.437192 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.437385 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.441661 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-57f5fb6b66-7vp2j" podStartSLOduration=2.441646911 podStartE2EDuration="2.441646911s" podCreationTimestamp="2026-02-17 13:23:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:23:21.394828826 +0000 UTC m=+1139.917558389" watchObservedRunningTime="2026-02-17 13:23:21.441646911 +0000 UTC m=+1139.964376454" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.442481 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.470364 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.621077 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1148d9f-bf24-449c-83b4-9a64d0572df0-public-tls-certs\") pod \"cinder-api-0\" (UID: \"b1148d9f-bf24-449c-83b4-9a64d0572df0\") " pod="openstack/cinder-api-0" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.621167 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1148d9f-bf24-449c-83b4-9a64d0572df0-config-data\") pod \"cinder-api-0\" (UID: \"b1148d9f-bf24-449c-83b4-9a64d0572df0\") " pod="openstack/cinder-api-0" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.621199 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1148d9f-bf24-449c-83b4-9a64d0572df0-logs\") pod \"cinder-api-0\" (UID: \"b1148d9f-bf24-449c-83b4-9a64d0572df0\") " pod="openstack/cinder-api-0" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.621233 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b1148d9f-bf24-449c-83b4-9a64d0572df0-config-data-custom\") pod \"cinder-api-0\" (UID: \"b1148d9f-bf24-449c-83b4-9a64d0572df0\") " pod="openstack/cinder-api-0" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.621330 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b1148d9f-bf24-449c-83b4-9a64d0572df0-etc-machine-id\") pod \"cinder-api-0\" (UID: \"b1148d9f-bf24-449c-83b4-9a64d0572df0\") " pod="openstack/cinder-api-0" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.621362 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1148d9f-bf24-449c-83b4-9a64d0572df0-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"b1148d9f-bf24-449c-83b4-9a64d0572df0\") " pod="openstack/cinder-api-0" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.621402 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1148d9f-bf24-449c-83b4-9a64d0572df0-scripts\") pod \"cinder-api-0\" (UID: \"b1148d9f-bf24-449c-83b4-9a64d0572df0\") " pod="openstack/cinder-api-0" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.621543 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnqkk\" (UniqueName: \"kubernetes.io/projected/b1148d9f-bf24-449c-83b4-9a64d0572df0-kube-api-access-rnqkk\") pod \"cinder-api-0\" (UID: \"b1148d9f-bf24-449c-83b4-9a64d0572df0\") " pod="openstack/cinder-api-0" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.621702 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1148d9f-bf24-449c-83b4-9a64d0572df0-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"b1148d9f-bf24-449c-83b4-9a64d0572df0\") " pod="openstack/cinder-api-0" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.723692 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1148d9f-bf24-449c-83b4-9a64d0572df0-public-tls-certs\") pod \"cinder-api-0\" (UID: \"b1148d9f-bf24-449c-83b4-9a64d0572df0\") " pod="openstack/cinder-api-0" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.723816 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1148d9f-bf24-449c-83b4-9a64d0572df0-config-data\") pod \"cinder-api-0\" (UID: \"b1148d9f-bf24-449c-83b4-9a64d0572df0\") " pod="openstack/cinder-api-0" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.723859 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1148d9f-bf24-449c-83b4-9a64d0572df0-logs\") pod \"cinder-api-0\" (UID: \"b1148d9f-bf24-449c-83b4-9a64d0572df0\") " pod="openstack/cinder-api-0" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.723888 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b1148d9f-bf24-449c-83b4-9a64d0572df0-config-data-custom\") pod \"cinder-api-0\" (UID: \"b1148d9f-bf24-449c-83b4-9a64d0572df0\") " pod="openstack/cinder-api-0" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.723962 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b1148d9f-bf24-449c-83b4-9a64d0572df0-etc-machine-id\") pod \"cinder-api-0\" (UID: \"b1148d9f-bf24-449c-83b4-9a64d0572df0\") " pod="openstack/cinder-api-0" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.723991 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1148d9f-bf24-449c-83b4-9a64d0572df0-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"b1148d9f-bf24-449c-83b4-9a64d0572df0\") " pod="openstack/cinder-api-0" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.724040 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1148d9f-bf24-449c-83b4-9a64d0572df0-scripts\") pod \"cinder-api-0\" (UID: \"b1148d9f-bf24-449c-83b4-9a64d0572df0\") " pod="openstack/cinder-api-0" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.724085 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnqkk\" (UniqueName: \"kubernetes.io/projected/b1148d9f-bf24-449c-83b4-9a64d0572df0-kube-api-access-rnqkk\") pod \"cinder-api-0\" (UID: \"b1148d9f-bf24-449c-83b4-9a64d0572df0\") " pod="openstack/cinder-api-0" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.724134 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1148d9f-bf24-449c-83b4-9a64d0572df0-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"b1148d9f-bf24-449c-83b4-9a64d0572df0\") " pod="openstack/cinder-api-0" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.724386 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1148d9f-bf24-449c-83b4-9a64d0572df0-logs\") pod \"cinder-api-0\" (UID: \"b1148d9f-bf24-449c-83b4-9a64d0572df0\") " pod="openstack/cinder-api-0" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.724741 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b1148d9f-bf24-449c-83b4-9a64d0572df0-etc-machine-id\") pod \"cinder-api-0\" (UID: \"b1148d9f-bf24-449c-83b4-9a64d0572df0\") " pod="openstack/cinder-api-0" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.728817 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1148d9f-bf24-449c-83b4-9a64d0572df0-public-tls-certs\") pod \"cinder-api-0\" (UID: \"b1148d9f-bf24-449c-83b4-9a64d0572df0\") " pod="openstack/cinder-api-0" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.729154 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1148d9f-bf24-449c-83b4-9a64d0572df0-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"b1148d9f-bf24-449c-83b4-9a64d0572df0\") " pod="openstack/cinder-api-0" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.729278 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1148d9f-bf24-449c-83b4-9a64d0572df0-scripts\") pod \"cinder-api-0\" (UID: \"b1148d9f-bf24-449c-83b4-9a64d0572df0\") " pod="openstack/cinder-api-0" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.729601 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b1148d9f-bf24-449c-83b4-9a64d0572df0-config-data-custom\") pod \"cinder-api-0\" (UID: \"b1148d9f-bf24-449c-83b4-9a64d0572df0\") " pod="openstack/cinder-api-0" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.732472 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1148d9f-bf24-449c-83b4-9a64d0572df0-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"b1148d9f-bf24-449c-83b4-9a64d0572df0\") " pod="openstack/cinder-api-0" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.738274 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1148d9f-bf24-449c-83b4-9a64d0572df0-config-data\") pod \"cinder-api-0\" (UID: \"b1148d9f-bf24-449c-83b4-9a64d0572df0\") " pod="openstack/cinder-api-0" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.742392 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnqkk\" (UniqueName: \"kubernetes.io/projected/b1148d9f-bf24-449c-83b4-9a64d0572df0-kube-api-access-rnqkk\") pod \"cinder-api-0\" (UID: \"b1148d9f-bf24-449c-83b4-9a64d0572df0\") " pod="openstack/cinder-api-0" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.782003 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 17 13:23:21 crc kubenswrapper[4955]: I0217 13:23:21.901977 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-784f78c585-7lvx4" podUID="e08d2b73-a720-4a28-bdc0-d8e8008417af" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.154:9696/\": dial tcp 10.217.0.154:9696: connect: connection refused" Feb 17 13:23:22 crc kubenswrapper[4955]: I0217 13:23:22.247210 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96771c04-ac0c-4eb8-8aa9-d59b754617d4" path="/var/lib/kubelet/pods/96771c04-ac0c-4eb8-8aa9-d59b754617d4/volumes" Feb 17 13:23:22 crc kubenswrapper[4955]: I0217 13:23:22.364621 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-57f5fb6b66-7vp2j" Feb 17 13:23:22 crc kubenswrapper[4955]: I0217 13:23:22.364662 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-57f5fb6b66-7vp2j" Feb 17 13:23:23 crc kubenswrapper[4955]: I0217 13:23:23.397216 4955 generic.go:334] "Generic (PLEG): container finished" podID="1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf" containerID="52cdaba09384ae938ed1a6b17e232b9095c1468ccdd7386221ad6a01d6be8ff4" exitCode=0 Feb 17 13:23:23 crc kubenswrapper[4955]: I0217 13:23:23.397307 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5788cc4448-245bs" event={"ID":"1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf","Type":"ContainerDied","Data":"52cdaba09384ae938ed1a6b17e232b9095c1468ccdd7386221ad6a01d6be8ff4"} Feb 17 13:23:23 crc kubenswrapper[4955]: I0217 13:23:23.649732 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5788cc4448-245bs" podUID="1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Feb 17 13:23:23 crc kubenswrapper[4955]: I0217 13:23:23.882266 4955 scope.go:117] "RemoveContainer" containerID="7daf52da12552e9ff871d56fe0c83c8dfdafb7b7c497ab0251922a2dbc6dab46" Feb 17 13:23:24 crc kubenswrapper[4955]: I0217 13:23:24.058339 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-65b6bfc874-zgznk" podUID="353a4cd6-336b-4545-903e-245c9cea97ca" containerName="barbican-api-log" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 17 13:23:24 crc kubenswrapper[4955]: I0217 13:23:24.099758 4955 scope.go:117] "RemoveContainer" containerID="924535e4f8c48b4e7872940dd0a8e1ca12e8b1f69d0012cb24a3e0fb3ae89d14" Feb 17 13:23:24 crc kubenswrapper[4955]: E0217 13:23:24.101303 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"924535e4f8c48b4e7872940dd0a8e1ca12e8b1f69d0012cb24a3e0fb3ae89d14\": container with ID starting with 924535e4f8c48b4e7872940dd0a8e1ca12e8b1f69d0012cb24a3e0fb3ae89d14 not found: ID does not exist" containerID="924535e4f8c48b4e7872940dd0a8e1ca12e8b1f69d0012cb24a3e0fb3ae89d14" Feb 17 13:23:24 crc kubenswrapper[4955]: I0217 13:23:24.101346 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"924535e4f8c48b4e7872940dd0a8e1ca12e8b1f69d0012cb24a3e0fb3ae89d14"} err="failed to get container status \"924535e4f8c48b4e7872940dd0a8e1ca12e8b1f69d0012cb24a3e0fb3ae89d14\": rpc error: code = NotFound desc = could not find container \"924535e4f8c48b4e7872940dd0a8e1ca12e8b1f69d0012cb24a3e0fb3ae89d14\": container with ID starting with 924535e4f8c48b4e7872940dd0a8e1ca12e8b1f69d0012cb24a3e0fb3ae89d14 not found: ID does not exist" Feb 17 13:23:24 crc kubenswrapper[4955]: I0217 13:23:24.101372 4955 scope.go:117] "RemoveContainer" containerID="7daf52da12552e9ff871d56fe0c83c8dfdafb7b7c497ab0251922a2dbc6dab46" Feb 17 13:23:24 crc kubenswrapper[4955]: E0217 13:23:24.102389 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7daf52da12552e9ff871d56fe0c83c8dfdafb7b7c497ab0251922a2dbc6dab46\": container with ID starting with 7daf52da12552e9ff871d56fe0c83c8dfdafb7b7c497ab0251922a2dbc6dab46 not found: ID does not exist" containerID="7daf52da12552e9ff871d56fe0c83c8dfdafb7b7c497ab0251922a2dbc6dab46" Feb 17 13:23:24 crc kubenswrapper[4955]: I0217 13:23:24.102419 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7daf52da12552e9ff871d56fe0c83c8dfdafb7b7c497ab0251922a2dbc6dab46"} err="failed to get container status \"7daf52da12552e9ff871d56fe0c83c8dfdafb7b7c497ab0251922a2dbc6dab46\": rpc error: code = NotFound desc = could not find container \"7daf52da12552e9ff871d56fe0c83c8dfdafb7b7c497ab0251922a2dbc6dab46\": container with ID starting with 7daf52da12552e9ff871d56fe0c83c8dfdafb7b7c497ab0251922a2dbc6dab46 not found: ID does not exist" Feb 17 13:23:24 crc kubenswrapper[4955]: I0217 13:23:24.102436 4955 scope.go:117] "RemoveContainer" containerID="924535e4f8c48b4e7872940dd0a8e1ca12e8b1f69d0012cb24a3e0fb3ae89d14" Feb 17 13:23:24 crc kubenswrapper[4955]: I0217 13:23:24.107152 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"924535e4f8c48b4e7872940dd0a8e1ca12e8b1f69d0012cb24a3e0fb3ae89d14"} err="failed to get container status \"924535e4f8c48b4e7872940dd0a8e1ca12e8b1f69d0012cb24a3e0fb3ae89d14\": rpc error: code = NotFound desc = could not find container \"924535e4f8c48b4e7872940dd0a8e1ca12e8b1f69d0012cb24a3e0fb3ae89d14\": container with ID starting with 924535e4f8c48b4e7872940dd0a8e1ca12e8b1f69d0012cb24a3e0fb3ae89d14 not found: ID does not exist" Feb 17 13:23:24 crc kubenswrapper[4955]: I0217 13:23:24.107189 4955 scope.go:117] "RemoveContainer" containerID="7daf52da12552e9ff871d56fe0c83c8dfdafb7b7c497ab0251922a2dbc6dab46" Feb 17 13:23:24 crc kubenswrapper[4955]: I0217 13:23:24.109462 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7daf52da12552e9ff871d56fe0c83c8dfdafb7b7c497ab0251922a2dbc6dab46"} err="failed to get container status \"7daf52da12552e9ff871d56fe0c83c8dfdafb7b7c497ab0251922a2dbc6dab46\": rpc error: code = NotFound desc = could not find container \"7daf52da12552e9ff871d56fe0c83c8dfdafb7b7c497ab0251922a2dbc6dab46\": container with ID starting with 7daf52da12552e9ff871d56fe0c83c8dfdafb7b7c497ab0251922a2dbc6dab46 not found: ID does not exist" Feb 17 13:23:24 crc kubenswrapper[4955]: I0217 13:23:24.462025 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7c5ddc6895-552ts" event={"ID":"25c07d2a-4f61-48fc-88e3-0ed671ad9cff","Type":"ContainerStarted","Data":"aea686f49f2b704f3fc24780ef183b09ce7f70fb69968abeb587b177d361fbc7"} Feb 17 13:23:24 crc kubenswrapper[4955]: I0217 13:23:24.611246 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Feb 17 13:23:24 crc kubenswrapper[4955]: I0217 13:23:24.669624 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 17 13:23:24 crc kubenswrapper[4955]: I0217 13:23:24.801022 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6578955fd5-hvvzs" Feb 17 13:23:24 crc kubenswrapper[4955]: I0217 13:23:24.863945 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-hkwx7"] Feb 17 13:23:24 crc kubenswrapper[4955]: I0217 13:23:24.864176 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b7b667979-hkwx7" podUID="cca48ad8-a612-4d37-b39d-116707dbef56" containerName="dnsmasq-dns" containerID="cri-o://70afb0d3b38304267ebdd70a3406513bfd74b19c6d94549763933370586c836c" gracePeriod=10 Feb 17 13:23:25 crc kubenswrapper[4955]: I0217 13:23:25.079178 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Feb 17 13:23:25 crc kubenswrapper[4955]: I0217 13:23:25.228455 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-65b6bfc874-zgznk" Feb 17 13:23:25 crc kubenswrapper[4955]: I0217 13:23:25.274645 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-65b6bfc874-zgznk" Feb 17 13:23:25 crc kubenswrapper[4955]: I0217 13:23:25.542004 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7c5ddc6895-552ts" event={"ID":"25c07d2a-4f61-48fc-88e3-0ed671ad9cff","Type":"ContainerStarted","Data":"cbd8bafad267be2efe0669a328460f8e8d6c13285d7fce1645736b599804dff5"} Feb 17 13:23:25 crc kubenswrapper[4955]: I0217 13:23:25.542499 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7c5ddc6895-552ts" Feb 17 13:23:25 crc kubenswrapper[4955]: I0217 13:23:25.582796 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7c5ddc6895-552ts" podStartSLOduration=5.582763044 podStartE2EDuration="5.582763044s" podCreationTimestamp="2026-02-17 13:23:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:23:25.570399672 +0000 UTC m=+1144.093129225" watchObservedRunningTime="2026-02-17 13:23:25.582763044 +0000 UTC m=+1144.105492587" Feb 17 13:23:25 crc kubenswrapper[4955]: I0217 13:23:25.643392 4955 generic.go:334] "Generic (PLEG): container finished" podID="cca48ad8-a612-4d37-b39d-116707dbef56" containerID="70afb0d3b38304267ebdd70a3406513bfd74b19c6d94549763933370586c836c" exitCode=0 Feb 17 13:23:25 crc kubenswrapper[4955]: I0217 13:23:25.643481 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-hkwx7" event={"ID":"cca48ad8-a612-4d37-b39d-116707dbef56","Type":"ContainerDied","Data":"70afb0d3b38304267ebdd70a3406513bfd74b19c6d94549763933370586c836c"} Feb 17 13:23:25 crc kubenswrapper[4955]: I0217 13:23:25.651017 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b1148d9f-bf24-449c-83b4-9a64d0572df0","Type":"ContainerStarted","Data":"d095db80f5b6d92b20db3c3e684e9d3ec1f957a075fac4b333a6ec4320143eae"} Feb 17 13:23:25 crc kubenswrapper[4955]: I0217 13:23:25.703081 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d20f3280-ab79-434f-a89a-11d6fbfc0edf","Type":"ContainerStarted","Data":"4cf2d0a3c593da686ace3c484b767ecbf30fa77a09a675c837b3c46000475432"} Feb 17 13:23:25 crc kubenswrapper[4955]: I0217 13:23:25.703224 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 17 13:23:25 crc kubenswrapper[4955]: I0217 13:23:25.733082 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.91766384 podStartE2EDuration="8.733062502s" podCreationTimestamp="2026-02-17 13:23:17 +0000 UTC" firstStartedPulling="2026-02-17 13:23:18.415556783 +0000 UTC m=+1136.938286326" lastFinishedPulling="2026-02-17 13:23:24.230955445 +0000 UTC m=+1142.753684988" observedRunningTime="2026-02-17 13:23:25.732605059 +0000 UTC m=+1144.255334612" watchObservedRunningTime="2026-02-17 13:23:25.733062502 +0000 UTC m=+1144.255792045" Feb 17 13:23:25 crc kubenswrapper[4955]: I0217 13:23:25.821146 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-hkwx7" Feb 17 13:23:25 crc kubenswrapper[4955]: I0217 13:23:25.836259 4955 generic.go:334] "Generic (PLEG): container finished" podID="e08d2b73-a720-4a28-bdc0-d8e8008417af" containerID="4c25f2067734abe7fc2878ff092cd406523ac1cd49c75130c410877ed3c8a5bb" exitCode=0 Feb 17 13:23:25 crc kubenswrapper[4955]: I0217 13:23:25.837927 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-784f78c585-7lvx4" event={"ID":"e08d2b73-a720-4a28-bdc0-d8e8008417af","Type":"ContainerDied","Data":"4c25f2067734abe7fc2878ff092cd406523ac1cd49c75130c410877ed3c8a5bb"} Feb 17 13:23:25 crc kubenswrapper[4955]: I0217 13:23:25.911583 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cca48ad8-a612-4d37-b39d-116707dbef56-dns-swift-storage-0\") pod \"cca48ad8-a612-4d37-b39d-116707dbef56\" (UID: \"cca48ad8-a612-4d37-b39d-116707dbef56\") " Feb 17 13:23:25 crc kubenswrapper[4955]: I0217 13:23:25.911732 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8kghk\" (UniqueName: \"kubernetes.io/projected/cca48ad8-a612-4d37-b39d-116707dbef56-kube-api-access-8kghk\") pod \"cca48ad8-a612-4d37-b39d-116707dbef56\" (UID: \"cca48ad8-a612-4d37-b39d-116707dbef56\") " Feb 17 13:23:25 crc kubenswrapper[4955]: I0217 13:23:25.911773 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cca48ad8-a612-4d37-b39d-116707dbef56-ovsdbserver-nb\") pod \"cca48ad8-a612-4d37-b39d-116707dbef56\" (UID: \"cca48ad8-a612-4d37-b39d-116707dbef56\") " Feb 17 13:23:25 crc kubenswrapper[4955]: I0217 13:23:25.919236 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cca48ad8-a612-4d37-b39d-116707dbef56-dns-svc\") pod \"cca48ad8-a612-4d37-b39d-116707dbef56\" (UID: \"cca48ad8-a612-4d37-b39d-116707dbef56\") " Feb 17 13:23:25 crc kubenswrapper[4955]: I0217 13:23:25.919285 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cca48ad8-a612-4d37-b39d-116707dbef56-config\") pod \"cca48ad8-a612-4d37-b39d-116707dbef56\" (UID: \"cca48ad8-a612-4d37-b39d-116707dbef56\") " Feb 17 13:23:25 crc kubenswrapper[4955]: I0217 13:23:25.919498 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cca48ad8-a612-4d37-b39d-116707dbef56-ovsdbserver-sb\") pod \"cca48ad8-a612-4d37-b39d-116707dbef56\" (UID: \"cca48ad8-a612-4d37-b39d-116707dbef56\") " Feb 17 13:23:25 crc kubenswrapper[4955]: I0217 13:23:25.942846 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 17 13:23:25 crc kubenswrapper[4955]: I0217 13:23:25.950544 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-784f78c585-7lvx4" Feb 17 13:23:25 crc kubenswrapper[4955]: I0217 13:23:25.966003 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cca48ad8-a612-4d37-b39d-116707dbef56-kube-api-access-8kghk" (OuterVolumeSpecName: "kube-api-access-8kghk") pod "cca48ad8-a612-4d37-b39d-116707dbef56" (UID: "cca48ad8-a612-4d37-b39d-116707dbef56"). InnerVolumeSpecName "kube-api-access-8kghk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.060331 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxpf4\" (UniqueName: \"kubernetes.io/projected/e08d2b73-a720-4a28-bdc0-d8e8008417af-kube-api-access-fxpf4\") pod \"e08d2b73-a720-4a28-bdc0-d8e8008417af\" (UID: \"e08d2b73-a720-4a28-bdc0-d8e8008417af\") " Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.087814 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e08d2b73-a720-4a28-bdc0-d8e8008417af-kube-api-access-fxpf4" (OuterVolumeSpecName: "kube-api-access-fxpf4") pod "e08d2b73-a720-4a28-bdc0-d8e8008417af" (UID: "e08d2b73-a720-4a28-bdc0-d8e8008417af"). InnerVolumeSpecName "kube-api-access-fxpf4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.088249 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e08d2b73-a720-4a28-bdc0-d8e8008417af-internal-tls-certs\") pod \"e08d2b73-a720-4a28-bdc0-d8e8008417af\" (UID: \"e08d2b73-a720-4a28-bdc0-d8e8008417af\") " Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.088341 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e08d2b73-a720-4a28-bdc0-d8e8008417af-config\") pod \"e08d2b73-a720-4a28-bdc0-d8e8008417af\" (UID: \"e08d2b73-a720-4a28-bdc0-d8e8008417af\") " Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.088423 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e08d2b73-a720-4a28-bdc0-d8e8008417af-ovndb-tls-certs\") pod \"e08d2b73-a720-4a28-bdc0-d8e8008417af\" (UID: \"e08d2b73-a720-4a28-bdc0-d8e8008417af\") " Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.088604 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e08d2b73-a720-4a28-bdc0-d8e8008417af-public-tls-certs\") pod \"e08d2b73-a720-4a28-bdc0-d8e8008417af\" (UID: \"e08d2b73-a720-4a28-bdc0-d8e8008417af\") " Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.088710 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e08d2b73-a720-4a28-bdc0-d8e8008417af-httpd-config\") pod \"e08d2b73-a720-4a28-bdc0-d8e8008417af\" (UID: \"e08d2b73-a720-4a28-bdc0-d8e8008417af\") " Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.088823 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e08d2b73-a720-4a28-bdc0-d8e8008417af-combined-ca-bundle\") pod \"e08d2b73-a720-4a28-bdc0-d8e8008417af\" (UID: \"e08d2b73-a720-4a28-bdc0-d8e8008417af\") " Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.089502 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxpf4\" (UniqueName: \"kubernetes.io/projected/e08d2b73-a720-4a28-bdc0-d8e8008417af-kube-api-access-fxpf4\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.089577 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8kghk\" (UniqueName: \"kubernetes.io/projected/cca48ad8-a612-4d37-b39d-116707dbef56-kube-api-access-8kghk\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.129964 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cca48ad8-a612-4d37-b39d-116707dbef56-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "cca48ad8-a612-4d37-b39d-116707dbef56" (UID: "cca48ad8-a612-4d37-b39d-116707dbef56"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.130715 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cca48ad8-a612-4d37-b39d-116707dbef56-config" (OuterVolumeSpecName: "config") pod "cca48ad8-a612-4d37-b39d-116707dbef56" (UID: "cca48ad8-a612-4d37-b39d-116707dbef56"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.144397 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cca48ad8-a612-4d37-b39d-116707dbef56-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "cca48ad8-a612-4d37-b39d-116707dbef56" (UID: "cca48ad8-a612-4d37-b39d-116707dbef56"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.144508 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e08d2b73-a720-4a28-bdc0-d8e8008417af-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "e08d2b73-a720-4a28-bdc0-d8e8008417af" (UID: "e08d2b73-a720-4a28-bdc0-d8e8008417af"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.188245 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cca48ad8-a612-4d37-b39d-116707dbef56-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "cca48ad8-a612-4d37-b39d-116707dbef56" (UID: "cca48ad8-a612-4d37-b39d-116707dbef56"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.191473 4955 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cca48ad8-a612-4d37-b39d-116707dbef56-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.191503 4955 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cca48ad8-a612-4d37-b39d-116707dbef56-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.191514 4955 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cca48ad8-a612-4d37-b39d-116707dbef56-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.191522 4955 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cca48ad8-a612-4d37-b39d-116707dbef56-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.191532 4955 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e08d2b73-a720-4a28-bdc0-d8e8008417af-httpd-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.220439 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cca48ad8-a612-4d37-b39d-116707dbef56-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cca48ad8-a612-4d37-b39d-116707dbef56" (UID: "cca48ad8-a612-4d37-b39d-116707dbef56"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.236121 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e08d2b73-a720-4a28-bdc0-d8e8008417af-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e08d2b73-a720-4a28-bdc0-d8e8008417af" (UID: "e08d2b73-a720-4a28-bdc0-d8e8008417af"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.283263 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-845ffbbf5d-ntk67" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.285764 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e08d2b73-a720-4a28-bdc0-d8e8008417af-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "e08d2b73-a720-4a28-bdc0-d8e8008417af" (UID: "e08d2b73-a720-4a28-bdc0-d8e8008417af"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.290928 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e08d2b73-a720-4a28-bdc0-d8e8008417af-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e08d2b73-a720-4a28-bdc0-d8e8008417af" (UID: "e08d2b73-a720-4a28-bdc0-d8e8008417af"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.291386 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-845ffbbf5d-ntk67" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.292794 4955 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e08d2b73-a720-4a28-bdc0-d8e8008417af-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.292811 4955 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e08d2b73-a720-4a28-bdc0-d8e8008417af-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.292820 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e08d2b73-a720-4a28-bdc0-d8e8008417af-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.292828 4955 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cca48ad8-a612-4d37-b39d-116707dbef56-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.298005 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e08d2b73-a720-4a28-bdc0-d8e8008417af-config" (OuterVolumeSpecName: "config") pod "e08d2b73-a720-4a28-bdc0-d8e8008417af" (UID: "e08d2b73-a720-4a28-bdc0-d8e8008417af"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.339827 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e08d2b73-a720-4a28-bdc0-d8e8008417af-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "e08d2b73-a720-4a28-bdc0-d8e8008417af" (UID: "e08d2b73-a720-4a28-bdc0-d8e8008417af"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.397647 4955 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/e08d2b73-a720-4a28-bdc0-d8e8008417af-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.397688 4955 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e08d2b73-a720-4a28-bdc0-d8e8008417af-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.553823 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-5d56c994-tx7w7"] Feb 17 13:23:26 crc kubenswrapper[4955]: E0217 13:23:26.554882 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e08d2b73-a720-4a28-bdc0-d8e8008417af" containerName="neutron-api" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.554904 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="e08d2b73-a720-4a28-bdc0-d8e8008417af" containerName="neutron-api" Feb 17 13:23:26 crc kubenswrapper[4955]: E0217 13:23:26.554916 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cca48ad8-a612-4d37-b39d-116707dbef56" containerName="init" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.554923 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="cca48ad8-a612-4d37-b39d-116707dbef56" containerName="init" Feb 17 13:23:26 crc kubenswrapper[4955]: E0217 13:23:26.554958 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cca48ad8-a612-4d37-b39d-116707dbef56" containerName="dnsmasq-dns" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.554965 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="cca48ad8-a612-4d37-b39d-116707dbef56" containerName="dnsmasq-dns" Feb 17 13:23:26 crc kubenswrapper[4955]: E0217 13:23:26.554987 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e08d2b73-a720-4a28-bdc0-d8e8008417af" containerName="neutron-httpd" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.554993 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="e08d2b73-a720-4a28-bdc0-d8e8008417af" containerName="neutron-httpd" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.555348 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="e08d2b73-a720-4a28-bdc0-d8e8008417af" containerName="neutron-httpd" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.555370 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="e08d2b73-a720-4a28-bdc0-d8e8008417af" containerName="neutron-api" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.555389 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="cca48ad8-a612-4d37-b39d-116707dbef56" containerName="dnsmasq-dns" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.557046 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5d56c994-tx7w7" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.602191 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54f84d50-4d2b-4d71-91e0-e6a7a00f568e-config-data\") pod \"placement-5d56c994-tx7w7\" (UID: \"54f84d50-4d2b-4d71-91e0-e6a7a00f568e\") " pod="openstack/placement-5d56c994-tx7w7" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.602297 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54f84d50-4d2b-4d71-91e0-e6a7a00f568e-combined-ca-bundle\") pod \"placement-5d56c994-tx7w7\" (UID: \"54f84d50-4d2b-4d71-91e0-e6a7a00f568e\") " pod="openstack/placement-5d56c994-tx7w7" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.602448 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54f84d50-4d2b-4d71-91e0-e6a7a00f568e-scripts\") pod \"placement-5d56c994-tx7w7\" (UID: \"54f84d50-4d2b-4d71-91e0-e6a7a00f568e\") " pod="openstack/placement-5d56c994-tx7w7" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.602497 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/54f84d50-4d2b-4d71-91e0-e6a7a00f568e-public-tls-certs\") pod \"placement-5d56c994-tx7w7\" (UID: \"54f84d50-4d2b-4d71-91e0-e6a7a00f568e\") " pod="openstack/placement-5d56c994-tx7w7" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.602520 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mggs5\" (UniqueName: \"kubernetes.io/projected/54f84d50-4d2b-4d71-91e0-e6a7a00f568e-kube-api-access-mggs5\") pod \"placement-5d56c994-tx7w7\" (UID: \"54f84d50-4d2b-4d71-91e0-e6a7a00f568e\") " pod="openstack/placement-5d56c994-tx7w7" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.602654 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/54f84d50-4d2b-4d71-91e0-e6a7a00f568e-internal-tls-certs\") pod \"placement-5d56c994-tx7w7\" (UID: \"54f84d50-4d2b-4d71-91e0-e6a7a00f568e\") " pod="openstack/placement-5d56c994-tx7w7" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.602727 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/54f84d50-4d2b-4d71-91e0-e6a7a00f568e-logs\") pod \"placement-5d56c994-tx7w7\" (UID: \"54f84d50-4d2b-4d71-91e0-e6a7a00f568e\") " pod="openstack/placement-5d56c994-tx7w7" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.603224 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5d56c994-tx7w7"] Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.704460 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54f84d50-4d2b-4d71-91e0-e6a7a00f568e-config-data\") pod \"placement-5d56c994-tx7w7\" (UID: \"54f84d50-4d2b-4d71-91e0-e6a7a00f568e\") " pod="openstack/placement-5d56c994-tx7w7" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.704523 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54f84d50-4d2b-4d71-91e0-e6a7a00f568e-combined-ca-bundle\") pod \"placement-5d56c994-tx7w7\" (UID: \"54f84d50-4d2b-4d71-91e0-e6a7a00f568e\") " pod="openstack/placement-5d56c994-tx7w7" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.704572 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54f84d50-4d2b-4d71-91e0-e6a7a00f568e-scripts\") pod \"placement-5d56c994-tx7w7\" (UID: \"54f84d50-4d2b-4d71-91e0-e6a7a00f568e\") " pod="openstack/placement-5d56c994-tx7w7" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.704598 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/54f84d50-4d2b-4d71-91e0-e6a7a00f568e-public-tls-certs\") pod \"placement-5d56c994-tx7w7\" (UID: \"54f84d50-4d2b-4d71-91e0-e6a7a00f568e\") " pod="openstack/placement-5d56c994-tx7w7" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.704620 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mggs5\" (UniqueName: \"kubernetes.io/projected/54f84d50-4d2b-4d71-91e0-e6a7a00f568e-kube-api-access-mggs5\") pod \"placement-5d56c994-tx7w7\" (UID: \"54f84d50-4d2b-4d71-91e0-e6a7a00f568e\") " pod="openstack/placement-5d56c994-tx7w7" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.704675 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/54f84d50-4d2b-4d71-91e0-e6a7a00f568e-internal-tls-certs\") pod \"placement-5d56c994-tx7w7\" (UID: \"54f84d50-4d2b-4d71-91e0-e6a7a00f568e\") " pod="openstack/placement-5d56c994-tx7w7" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.704696 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/54f84d50-4d2b-4d71-91e0-e6a7a00f568e-logs\") pod \"placement-5d56c994-tx7w7\" (UID: \"54f84d50-4d2b-4d71-91e0-e6a7a00f568e\") " pod="openstack/placement-5d56c994-tx7w7" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.707265 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/54f84d50-4d2b-4d71-91e0-e6a7a00f568e-logs\") pod \"placement-5d56c994-tx7w7\" (UID: \"54f84d50-4d2b-4d71-91e0-e6a7a00f568e\") " pod="openstack/placement-5d56c994-tx7w7" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.716361 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/54f84d50-4d2b-4d71-91e0-e6a7a00f568e-internal-tls-certs\") pod \"placement-5d56c994-tx7w7\" (UID: \"54f84d50-4d2b-4d71-91e0-e6a7a00f568e\") " pod="openstack/placement-5d56c994-tx7w7" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.716914 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/54f84d50-4d2b-4d71-91e0-e6a7a00f568e-public-tls-certs\") pod \"placement-5d56c994-tx7w7\" (UID: \"54f84d50-4d2b-4d71-91e0-e6a7a00f568e\") " pod="openstack/placement-5d56c994-tx7w7" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.717437 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54f84d50-4d2b-4d71-91e0-e6a7a00f568e-config-data\") pod \"placement-5d56c994-tx7w7\" (UID: \"54f84d50-4d2b-4d71-91e0-e6a7a00f568e\") " pod="openstack/placement-5d56c994-tx7w7" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.722252 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54f84d50-4d2b-4d71-91e0-e6a7a00f568e-scripts\") pod \"placement-5d56c994-tx7w7\" (UID: \"54f84d50-4d2b-4d71-91e0-e6a7a00f568e\") " pod="openstack/placement-5d56c994-tx7w7" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.722363 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54f84d50-4d2b-4d71-91e0-e6a7a00f568e-combined-ca-bundle\") pod \"placement-5d56c994-tx7w7\" (UID: \"54f84d50-4d2b-4d71-91e0-e6a7a00f568e\") " pod="openstack/placement-5d56c994-tx7w7" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.730542 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mggs5\" (UniqueName: \"kubernetes.io/projected/54f84d50-4d2b-4d71-91e0-e6a7a00f568e-kube-api-access-mggs5\") pod \"placement-5d56c994-tx7w7\" (UID: \"54f84d50-4d2b-4d71-91e0-e6a7a00f568e\") " pod="openstack/placement-5d56c994-tx7w7" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.878387 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-hkwx7" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.879069 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-hkwx7" event={"ID":"cca48ad8-a612-4d37-b39d-116707dbef56","Type":"ContainerDied","Data":"91cbb2a12ff82366fed197af6c7e2e6f2aca666921be26f3407ebb01be77dc5d"} Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.879366 4955 scope.go:117] "RemoveContainer" containerID="70afb0d3b38304267ebdd70a3406513bfd74b19c6d94549763933370586c836c" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.894142 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b1148d9f-bf24-449c-83b4-9a64d0572df0","Type":"ContainerStarted","Data":"881087f67405009d4826166656c2eb47a8bfc9ba067856b9ffb86073469857e2"} Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.907255 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5d56c994-tx7w7" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.931823 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-784f78c585-7lvx4" Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.931877 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-784f78c585-7lvx4" event={"ID":"e08d2b73-a720-4a28-bdc0-d8e8008417af","Type":"ContainerDied","Data":"efd2421edee17423c62f62becce62b64eca1b74343b0d02d67e8b07d9939a824"} Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.932743 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-hkwx7"] Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.933155 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="99f00e22-f2b5-4e1f-b00a-9e50a1ed359e" containerName="cinder-scheduler" containerID="cri-o://668d903f87d617139f441377187fff1074088f120f5b14ef8c8dc4dc0c9c2fac" gracePeriod=30 Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.933249 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="99f00e22-f2b5-4e1f-b00a-9e50a1ed359e" containerName="probe" containerID="cri-o://555092eb6d9a00102f0ef85ead43250fb502025f9d327672684751329d350428" gracePeriod=30 Feb 17 13:23:26 crc kubenswrapper[4955]: I0217 13:23:26.974001 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-hkwx7"] Feb 17 13:23:27 crc kubenswrapper[4955]: I0217 13:23:27.003114 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-784f78c585-7lvx4"] Feb 17 13:23:27 crc kubenswrapper[4955]: I0217 13:23:27.014952 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-784f78c585-7lvx4"] Feb 17 13:23:27 crc kubenswrapper[4955]: I0217 13:23:27.017976 4955 scope.go:117] "RemoveContainer" containerID="bc35cbd2f0bc63599385fc63a8fe9deeaaa338b17281a5bec3563abf75ade0cd" Feb 17 13:23:27 crc kubenswrapper[4955]: I0217 13:23:27.176172 4955 scope.go:117] "RemoveContainer" containerID="edfe665cc4bfb8b12efc35a155058e0faa141827cfcb336ebc28576eb4699582" Feb 17 13:23:27 crc kubenswrapper[4955]: I0217 13:23:27.185803 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 17 13:23:27 crc kubenswrapper[4955]: I0217 13:23:27.198079 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 17 13:23:27 crc kubenswrapper[4955]: I0217 13:23:27.284686 4955 scope.go:117] "RemoveContainer" containerID="4c25f2067734abe7fc2878ff092cd406523ac1cd49c75130c410877ed3c8a5bb" Feb 17 13:23:27 crc kubenswrapper[4955]: I0217 13:23:27.629893 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5d56c994-tx7w7"] Feb 17 13:23:27 crc kubenswrapper[4955]: I0217 13:23:27.965168 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b1148d9f-bf24-449c-83b4-9a64d0572df0","Type":"ContainerStarted","Data":"d6f2401572467f00fbc6f037e0f7b9c84f83f60f41a82bfcb2173c03d4f715f4"} Feb 17 13:23:27 crc kubenswrapper[4955]: I0217 13:23:27.965827 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Feb 17 13:23:27 crc kubenswrapper[4955]: I0217 13:23:27.975335 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 17 13:23:27 crc kubenswrapper[4955]: I0217 13:23:27.993435 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=6.993418687 podStartE2EDuration="6.993418687s" podCreationTimestamp="2026-02-17 13:23:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:23:27.987639562 +0000 UTC m=+1146.510369105" watchObservedRunningTime="2026-02-17 13:23:27.993418687 +0000 UTC m=+1146.516148230" Feb 17 13:23:28 crc kubenswrapper[4955]: I0217 13:23:28.010062 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5d56c994-tx7w7" event={"ID":"54f84d50-4d2b-4d71-91e0-e6a7a00f568e","Type":"ContainerStarted","Data":"c51f1ff562680b633e891e190a7633ae9236a0284945422e98b97ad21256148c"} Feb 17 13:23:28 crc kubenswrapper[4955]: I0217 13:23:28.010287 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5d56c994-tx7w7" event={"ID":"54f84d50-4d2b-4d71-91e0-e6a7a00f568e","Type":"ContainerStarted","Data":"55b7f99f186d45275ded30d7568df4ca66599884e04250c00493f5a708df8221"} Feb 17 13:23:28 crc kubenswrapper[4955]: I0217 13:23:28.027928 4955 generic.go:334] "Generic (PLEG): container finished" podID="99f00e22-f2b5-4e1f-b00a-9e50a1ed359e" containerID="555092eb6d9a00102f0ef85ead43250fb502025f9d327672684751329d350428" exitCode=0 Feb 17 13:23:28 crc kubenswrapper[4955]: I0217 13:23:28.028205 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"99f00e22-f2b5-4e1f-b00a-9e50a1ed359e","Type":"ContainerDied","Data":"555092eb6d9a00102f0ef85ead43250fb502025f9d327672684751329d350428"} Feb 17 13:23:28 crc kubenswrapper[4955]: I0217 13:23:28.233806 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cca48ad8-a612-4d37-b39d-116707dbef56" path="/var/lib/kubelet/pods/cca48ad8-a612-4d37-b39d-116707dbef56/volumes" Feb 17 13:23:28 crc kubenswrapper[4955]: I0217 13:23:28.234556 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e08d2b73-a720-4a28-bdc0-d8e8008417af" path="/var/lib/kubelet/pods/e08d2b73-a720-4a28-bdc0-d8e8008417af/volumes" Feb 17 13:23:28 crc kubenswrapper[4955]: I0217 13:23:28.480382 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 17 13:23:29 crc kubenswrapper[4955]: I0217 13:23:29.048227 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5d56c994-tx7w7" event={"ID":"54f84d50-4d2b-4d71-91e0-e6a7a00f568e","Type":"ContainerStarted","Data":"6ebae3ea4927ce8704eac70c960ae3f0aa3aa1843af9069b717841906941803a"} Feb 17 13:23:29 crc kubenswrapper[4955]: I0217 13:23:29.049652 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5d56c994-tx7w7" Feb 17 13:23:29 crc kubenswrapper[4955]: I0217 13:23:29.049701 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5d56c994-tx7w7" Feb 17 13:23:29 crc kubenswrapper[4955]: I0217 13:23:29.067485 4955 generic.go:334] "Generic (PLEG): container finished" podID="99f00e22-f2b5-4e1f-b00a-9e50a1ed359e" containerID="668d903f87d617139f441377187fff1074088f120f5b14ef8c8dc4dc0c9c2fac" exitCode=0 Feb 17 13:23:29 crc kubenswrapper[4955]: I0217 13:23:29.067579 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"99f00e22-f2b5-4e1f-b00a-9e50a1ed359e","Type":"ContainerDied","Data":"668d903f87d617139f441377187fff1074088f120f5b14ef8c8dc4dc0c9c2fac"} Feb 17 13:23:29 crc kubenswrapper[4955]: I0217 13:23:29.090873 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-5d56c994-tx7w7" podStartSLOduration=3.09084745 podStartE2EDuration="3.09084745s" podCreationTimestamp="2026-02-17 13:23:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:23:29.077423518 +0000 UTC m=+1147.600153061" watchObservedRunningTime="2026-02-17 13:23:29.09084745 +0000 UTC m=+1147.613577003" Feb 17 13:23:29 crc kubenswrapper[4955]: I0217 13:23:29.204615 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 17 13:23:29 crc kubenswrapper[4955]: I0217 13:23:29.263637 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99f00e22-f2b5-4e1f-b00a-9e50a1ed359e-combined-ca-bundle\") pod \"99f00e22-f2b5-4e1f-b00a-9e50a1ed359e\" (UID: \"99f00e22-f2b5-4e1f-b00a-9e50a1ed359e\") " Feb 17 13:23:29 crc kubenswrapper[4955]: I0217 13:23:29.263842 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99f00e22-f2b5-4e1f-b00a-9e50a1ed359e-config-data\") pod \"99f00e22-f2b5-4e1f-b00a-9e50a1ed359e\" (UID: \"99f00e22-f2b5-4e1f-b00a-9e50a1ed359e\") " Feb 17 13:23:29 crc kubenswrapper[4955]: I0217 13:23:29.263919 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/99f00e22-f2b5-4e1f-b00a-9e50a1ed359e-config-data-custom\") pod \"99f00e22-f2b5-4e1f-b00a-9e50a1ed359e\" (UID: \"99f00e22-f2b5-4e1f-b00a-9e50a1ed359e\") " Feb 17 13:23:29 crc kubenswrapper[4955]: I0217 13:23:29.264127 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7sjc\" (UniqueName: \"kubernetes.io/projected/99f00e22-f2b5-4e1f-b00a-9e50a1ed359e-kube-api-access-x7sjc\") pod \"99f00e22-f2b5-4e1f-b00a-9e50a1ed359e\" (UID: \"99f00e22-f2b5-4e1f-b00a-9e50a1ed359e\") " Feb 17 13:23:29 crc kubenswrapper[4955]: I0217 13:23:29.264190 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99f00e22-f2b5-4e1f-b00a-9e50a1ed359e-scripts\") pod \"99f00e22-f2b5-4e1f-b00a-9e50a1ed359e\" (UID: \"99f00e22-f2b5-4e1f-b00a-9e50a1ed359e\") " Feb 17 13:23:29 crc kubenswrapper[4955]: I0217 13:23:29.264236 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/99f00e22-f2b5-4e1f-b00a-9e50a1ed359e-etc-machine-id\") pod \"99f00e22-f2b5-4e1f-b00a-9e50a1ed359e\" (UID: \"99f00e22-f2b5-4e1f-b00a-9e50a1ed359e\") " Feb 17 13:23:29 crc kubenswrapper[4955]: I0217 13:23:29.271415 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/99f00e22-f2b5-4e1f-b00a-9e50a1ed359e-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "99f00e22-f2b5-4e1f-b00a-9e50a1ed359e" (UID: "99f00e22-f2b5-4e1f-b00a-9e50a1ed359e"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 13:23:29 crc kubenswrapper[4955]: I0217 13:23:29.274463 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99f00e22-f2b5-4e1f-b00a-9e50a1ed359e-kube-api-access-x7sjc" (OuterVolumeSpecName: "kube-api-access-x7sjc") pod "99f00e22-f2b5-4e1f-b00a-9e50a1ed359e" (UID: "99f00e22-f2b5-4e1f-b00a-9e50a1ed359e"). InnerVolumeSpecName "kube-api-access-x7sjc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:23:29 crc kubenswrapper[4955]: I0217 13:23:29.284987 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99f00e22-f2b5-4e1f-b00a-9e50a1ed359e-scripts" (OuterVolumeSpecName: "scripts") pod "99f00e22-f2b5-4e1f-b00a-9e50a1ed359e" (UID: "99f00e22-f2b5-4e1f-b00a-9e50a1ed359e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:29 crc kubenswrapper[4955]: I0217 13:23:29.285916 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99f00e22-f2b5-4e1f-b00a-9e50a1ed359e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "99f00e22-f2b5-4e1f-b00a-9e50a1ed359e" (UID: "99f00e22-f2b5-4e1f-b00a-9e50a1ed359e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:29 crc kubenswrapper[4955]: I0217 13:23:29.356949 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99f00e22-f2b5-4e1f-b00a-9e50a1ed359e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "99f00e22-f2b5-4e1f-b00a-9e50a1ed359e" (UID: "99f00e22-f2b5-4e1f-b00a-9e50a1ed359e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:29 crc kubenswrapper[4955]: I0217 13:23:29.366163 4955 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/99f00e22-f2b5-4e1f-b00a-9e50a1ed359e-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:29 crc kubenswrapper[4955]: I0217 13:23:29.366236 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7sjc\" (UniqueName: \"kubernetes.io/projected/99f00e22-f2b5-4e1f-b00a-9e50a1ed359e-kube-api-access-x7sjc\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:29 crc kubenswrapper[4955]: I0217 13:23:29.366251 4955 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99f00e22-f2b5-4e1f-b00a-9e50a1ed359e-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:29 crc kubenswrapper[4955]: I0217 13:23:29.366260 4955 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/99f00e22-f2b5-4e1f-b00a-9e50a1ed359e-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:29 crc kubenswrapper[4955]: I0217 13:23:29.366269 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99f00e22-f2b5-4e1f-b00a-9e50a1ed359e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:29 crc kubenswrapper[4955]: I0217 13:23:29.494194 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99f00e22-f2b5-4e1f-b00a-9e50a1ed359e-config-data" (OuterVolumeSpecName: "config-data") pod "99f00e22-f2b5-4e1f-b00a-9e50a1ed359e" (UID: "99f00e22-f2b5-4e1f-b00a-9e50a1ed359e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:29 crc kubenswrapper[4955]: I0217 13:23:29.570951 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99f00e22-f2b5-4e1f-b00a-9e50a1ed359e-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:30 crc kubenswrapper[4955]: I0217 13:23:30.095196 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"99f00e22-f2b5-4e1f-b00a-9e50a1ed359e","Type":"ContainerDied","Data":"398adef26b9f13332ecfddd83d219068aad43f3c63ecfa90d85c0275922e7634"} Feb 17 13:23:30 crc kubenswrapper[4955]: I0217 13:23:30.095585 4955 scope.go:117] "RemoveContainer" containerID="555092eb6d9a00102f0ef85ead43250fb502025f9d327672684751329d350428" Feb 17 13:23:30 crc kubenswrapper[4955]: I0217 13:23:30.095235 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 17 13:23:30 crc kubenswrapper[4955]: I0217 13:23:30.171025 4955 scope.go:117] "RemoveContainer" containerID="668d903f87d617139f441377187fff1074088f120f5b14ef8c8dc4dc0c9c2fac" Feb 17 13:23:30 crc kubenswrapper[4955]: I0217 13:23:30.174027 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 17 13:23:30 crc kubenswrapper[4955]: I0217 13:23:30.199227 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 17 13:23:30 crc kubenswrapper[4955]: I0217 13:23:30.214479 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Feb 17 13:23:30 crc kubenswrapper[4955]: E0217 13:23:30.214967 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99f00e22-f2b5-4e1f-b00a-9e50a1ed359e" containerName="probe" Feb 17 13:23:30 crc kubenswrapper[4955]: I0217 13:23:30.214986 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="99f00e22-f2b5-4e1f-b00a-9e50a1ed359e" containerName="probe" Feb 17 13:23:30 crc kubenswrapper[4955]: E0217 13:23:30.215015 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99f00e22-f2b5-4e1f-b00a-9e50a1ed359e" containerName="cinder-scheduler" Feb 17 13:23:30 crc kubenswrapper[4955]: I0217 13:23:30.215024 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="99f00e22-f2b5-4e1f-b00a-9e50a1ed359e" containerName="cinder-scheduler" Feb 17 13:23:30 crc kubenswrapper[4955]: I0217 13:23:30.215257 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="99f00e22-f2b5-4e1f-b00a-9e50a1ed359e" containerName="probe" Feb 17 13:23:30 crc kubenswrapper[4955]: I0217 13:23:30.215278 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="99f00e22-f2b5-4e1f-b00a-9e50a1ed359e" containerName="cinder-scheduler" Feb 17 13:23:30 crc kubenswrapper[4955]: I0217 13:23:30.216648 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 17 13:23:30 crc kubenswrapper[4955]: I0217 13:23:30.222836 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 17 13:23:30 crc kubenswrapper[4955]: I0217 13:23:30.226609 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Feb 17 13:23:30 crc kubenswrapper[4955]: I0217 13:23:30.234471 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99f00e22-f2b5-4e1f-b00a-9e50a1ed359e" path="/var/lib/kubelet/pods/99f00e22-f2b5-4e1f-b00a-9e50a1ed359e/volumes" Feb 17 13:23:30 crc kubenswrapper[4955]: I0217 13:23:30.283933 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7\") " pod="openstack/cinder-scheduler-0" Feb 17 13:23:30 crc kubenswrapper[4955]: I0217 13:23:30.283989 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7-config-data\") pod \"cinder-scheduler-0\" (UID: \"2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7\") " pod="openstack/cinder-scheduler-0" Feb 17 13:23:30 crc kubenswrapper[4955]: I0217 13:23:30.284020 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7\") " pod="openstack/cinder-scheduler-0" Feb 17 13:23:30 crc kubenswrapper[4955]: I0217 13:23:30.284052 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6pth\" (UniqueName: \"kubernetes.io/projected/2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7-kube-api-access-m6pth\") pod \"cinder-scheduler-0\" (UID: \"2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7\") " pod="openstack/cinder-scheduler-0" Feb 17 13:23:30 crc kubenswrapper[4955]: I0217 13:23:30.284087 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7\") " pod="openstack/cinder-scheduler-0" Feb 17 13:23:30 crc kubenswrapper[4955]: I0217 13:23:30.284123 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7-scripts\") pod \"cinder-scheduler-0\" (UID: \"2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7\") " pod="openstack/cinder-scheduler-0" Feb 17 13:23:30 crc kubenswrapper[4955]: I0217 13:23:30.385212 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7\") " pod="openstack/cinder-scheduler-0" Feb 17 13:23:30 crc kubenswrapper[4955]: I0217 13:23:30.385271 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6pth\" (UniqueName: \"kubernetes.io/projected/2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7-kube-api-access-m6pth\") pod \"cinder-scheduler-0\" (UID: \"2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7\") " pod="openstack/cinder-scheduler-0" Feb 17 13:23:30 crc kubenswrapper[4955]: I0217 13:23:30.385325 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7\") " pod="openstack/cinder-scheduler-0" Feb 17 13:23:30 crc kubenswrapper[4955]: I0217 13:23:30.385378 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7\") " pod="openstack/cinder-scheduler-0" Feb 17 13:23:30 crc kubenswrapper[4955]: I0217 13:23:30.385703 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7-scripts\") pod \"cinder-scheduler-0\" (UID: \"2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7\") " pod="openstack/cinder-scheduler-0" Feb 17 13:23:30 crc kubenswrapper[4955]: I0217 13:23:30.386238 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7\") " pod="openstack/cinder-scheduler-0" Feb 17 13:23:30 crc kubenswrapper[4955]: I0217 13:23:30.386274 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7-config-data\") pod \"cinder-scheduler-0\" (UID: \"2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7\") " pod="openstack/cinder-scheduler-0" Feb 17 13:23:30 crc kubenswrapper[4955]: I0217 13:23:30.390046 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7-scripts\") pod \"cinder-scheduler-0\" (UID: \"2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7\") " pod="openstack/cinder-scheduler-0" Feb 17 13:23:30 crc kubenswrapper[4955]: I0217 13:23:30.390230 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7\") " pod="openstack/cinder-scheduler-0" Feb 17 13:23:30 crc kubenswrapper[4955]: I0217 13:23:30.392047 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7-config-data\") pod \"cinder-scheduler-0\" (UID: \"2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7\") " pod="openstack/cinder-scheduler-0" Feb 17 13:23:30 crc kubenswrapper[4955]: I0217 13:23:30.392258 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7\") " pod="openstack/cinder-scheduler-0" Feb 17 13:23:30 crc kubenswrapper[4955]: I0217 13:23:30.403955 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6pth\" (UniqueName: \"kubernetes.io/projected/2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7-kube-api-access-m6pth\") pod \"cinder-scheduler-0\" (UID: \"2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7\") " pod="openstack/cinder-scheduler-0" Feb 17 13:23:30 crc kubenswrapper[4955]: I0217 13:23:30.559895 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 17 13:23:31 crc kubenswrapper[4955]: I0217 13:23:31.057290 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 17 13:23:31 crc kubenswrapper[4955]: W0217 13:23:31.057952 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2a5b18d1_9eae_4cb4_a432_9e3c5b68e4e7.slice/crio-5abee1cb8f348b7bcb56dcbd9e1e974b87d526f615295f072c507d9b70f3578c WatchSource:0}: Error finding container 5abee1cb8f348b7bcb56dcbd9e1e974b87d526f615295f072c507d9b70f3578c: Status 404 returned error can't find the container with id 5abee1cb8f348b7bcb56dcbd9e1e974b87d526f615295f072c507d9b70f3578c Feb 17 13:23:31 crc kubenswrapper[4955]: I0217 13:23:31.115720 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7","Type":"ContainerStarted","Data":"5abee1cb8f348b7bcb56dcbd9e1e974b87d526f615295f072c507d9b70f3578c"} Feb 17 13:23:31 crc kubenswrapper[4955]: I0217 13:23:31.244931 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-57f5fb6b66-7vp2j" Feb 17 13:23:31 crc kubenswrapper[4955]: I0217 13:23:31.454826 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-57f5fb6b66-7vp2j" Feb 17 13:23:31 crc kubenswrapper[4955]: I0217 13:23:31.531187 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-65b6bfc874-zgznk"] Feb 17 13:23:31 crc kubenswrapper[4955]: I0217 13:23:31.531481 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-65b6bfc874-zgznk" podUID="353a4cd6-336b-4545-903e-245c9cea97ca" containerName="barbican-api-log" containerID="cri-o://7e9f2e7d38f46f5983cb4dfffe1c3cdfe6f350832d7f2c71a0ef1d278ec760f6" gracePeriod=30 Feb 17 13:23:31 crc kubenswrapper[4955]: I0217 13:23:31.531650 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-65b6bfc874-zgznk" podUID="353a4cd6-336b-4545-903e-245c9cea97ca" containerName="barbican-api" containerID="cri-o://9e1cde53c1148029ca5f31661b223a8898cf091c0dc94ec858633087db1437ec" gracePeriod=30 Feb 17 13:23:32 crc kubenswrapper[4955]: I0217 13:23:32.128276 4955 generic.go:334] "Generic (PLEG): container finished" podID="353a4cd6-336b-4545-903e-245c9cea97ca" containerID="7e9f2e7d38f46f5983cb4dfffe1c3cdfe6f350832d7f2c71a0ef1d278ec760f6" exitCode=143 Feb 17 13:23:32 crc kubenswrapper[4955]: I0217 13:23:32.128375 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65b6bfc874-zgznk" event={"ID":"353a4cd6-336b-4545-903e-245c9cea97ca","Type":"ContainerDied","Data":"7e9f2e7d38f46f5983cb4dfffe1c3cdfe6f350832d7f2c71a0ef1d278ec760f6"} Feb 17 13:23:32 crc kubenswrapper[4955]: I0217 13:23:32.135894 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7","Type":"ContainerStarted","Data":"e67898d8f02497c6fb02301d8ce6b2e47f21342182ce0d7530b00a3d0b1312aa"} Feb 17 13:23:33 crc kubenswrapper[4955]: I0217 13:23:33.145318 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7","Type":"ContainerStarted","Data":"49c1d00f38e40e7efe140486f60a479e74602c80ac171cff49f9fece8a535884"} Feb 17 13:23:33 crc kubenswrapper[4955]: I0217 13:23:33.166214 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.166195508 podStartE2EDuration="3.166195508s" podCreationTimestamp="2026-02-17 13:23:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:23:33.160989819 +0000 UTC m=+1151.683719362" watchObservedRunningTime="2026-02-17 13:23:33.166195508 +0000 UTC m=+1151.688925041" Feb 17 13:23:33 crc kubenswrapper[4955]: I0217 13:23:33.648583 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5788cc4448-245bs" podUID="1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Feb 17 13:23:34 crc kubenswrapper[4955]: I0217 13:23:34.493861 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-c5b8c48f6-zr5fn" Feb 17 13:23:34 crc kubenswrapper[4955]: I0217 13:23:34.574932 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 13:23:34 crc kubenswrapper[4955]: I0217 13:23:34.574990 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 13:23:34 crc kubenswrapper[4955]: I0217 13:23:34.575032 4955 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" Feb 17 13:23:34 crc kubenswrapper[4955]: I0217 13:23:34.575666 4955 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3481e1de1949e001dc1ec5b32fb42f82ef040a09d687f51932cebedf9288edf6"} pod="openshift-machine-config-operator/machine-config-daemon-29qxq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 17 13:23:34 crc kubenswrapper[4955]: I0217 13:23:34.575717 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" containerID="cri-o://3481e1de1949e001dc1ec5b32fb42f82ef040a09d687f51932cebedf9288edf6" gracePeriod=600 Feb 17 13:23:34 crc kubenswrapper[4955]: I0217 13:23:34.780696 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-65b6bfc874-zgznk" podUID="353a4cd6-336b-4545-903e-245c9cea97ca" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.162:9311/healthcheck\": read tcp 10.217.0.2:55468->10.217.0.162:9311: read: connection reset by peer" Feb 17 13:23:34 crc kubenswrapper[4955]: I0217 13:23:34.781992 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-65b6bfc874-zgznk" podUID="353a4cd6-336b-4545-903e-245c9cea97ca" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.162:9311/healthcheck\": read tcp 10.217.0.2:55456->10.217.0.162:9311: read: connection reset by peer" Feb 17 13:23:35 crc kubenswrapper[4955]: I0217 13:23:35.182144 4955 generic.go:334] "Generic (PLEG): container finished" podID="353a4cd6-336b-4545-903e-245c9cea97ca" containerID="9e1cde53c1148029ca5f31661b223a8898cf091c0dc94ec858633087db1437ec" exitCode=0 Feb 17 13:23:35 crc kubenswrapper[4955]: I0217 13:23:35.182220 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65b6bfc874-zgznk" event={"ID":"353a4cd6-336b-4545-903e-245c9cea97ca","Type":"ContainerDied","Data":"9e1cde53c1148029ca5f31661b223a8898cf091c0dc94ec858633087db1437ec"} Feb 17 13:23:35 crc kubenswrapper[4955]: I0217 13:23:35.191492 4955 generic.go:334] "Generic (PLEG): container finished" podID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerID="3481e1de1949e001dc1ec5b32fb42f82ef040a09d687f51932cebedf9288edf6" exitCode=0 Feb 17 13:23:35 crc kubenswrapper[4955]: I0217 13:23:35.191540 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" event={"ID":"5bf36887-2a2b-4193-abdb-a15beaf46300","Type":"ContainerDied","Data":"3481e1de1949e001dc1ec5b32fb42f82ef040a09d687f51932cebedf9288edf6"} Feb 17 13:23:35 crc kubenswrapper[4955]: I0217 13:23:35.191578 4955 scope.go:117] "RemoveContainer" containerID="a72972c52870687e44697a4fc3675d78bf2389f390e43593c993f04221b0f584" Feb 17 13:23:35 crc kubenswrapper[4955]: I0217 13:23:35.547186 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-65b6bfc874-zgznk" Feb 17 13:23:35 crc kubenswrapper[4955]: I0217 13:23:35.560232 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Feb 17 13:23:35 crc kubenswrapper[4955]: I0217 13:23:35.593540 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58mxz\" (UniqueName: \"kubernetes.io/projected/353a4cd6-336b-4545-903e-245c9cea97ca-kube-api-access-58mxz\") pod \"353a4cd6-336b-4545-903e-245c9cea97ca\" (UID: \"353a4cd6-336b-4545-903e-245c9cea97ca\") " Feb 17 13:23:35 crc kubenswrapper[4955]: I0217 13:23:35.593655 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/353a4cd6-336b-4545-903e-245c9cea97ca-combined-ca-bundle\") pod \"353a4cd6-336b-4545-903e-245c9cea97ca\" (UID: \"353a4cd6-336b-4545-903e-245c9cea97ca\") " Feb 17 13:23:35 crc kubenswrapper[4955]: I0217 13:23:35.593715 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/353a4cd6-336b-4545-903e-245c9cea97ca-config-data-custom\") pod \"353a4cd6-336b-4545-903e-245c9cea97ca\" (UID: \"353a4cd6-336b-4545-903e-245c9cea97ca\") " Feb 17 13:23:35 crc kubenswrapper[4955]: I0217 13:23:35.593805 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/353a4cd6-336b-4545-903e-245c9cea97ca-logs\") pod \"353a4cd6-336b-4545-903e-245c9cea97ca\" (UID: \"353a4cd6-336b-4545-903e-245c9cea97ca\") " Feb 17 13:23:35 crc kubenswrapper[4955]: I0217 13:23:35.593862 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/353a4cd6-336b-4545-903e-245c9cea97ca-config-data\") pod \"353a4cd6-336b-4545-903e-245c9cea97ca\" (UID: \"353a4cd6-336b-4545-903e-245c9cea97ca\") " Feb 17 13:23:35 crc kubenswrapper[4955]: I0217 13:23:35.599298 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/353a4cd6-336b-4545-903e-245c9cea97ca-logs" (OuterVolumeSpecName: "logs") pod "353a4cd6-336b-4545-903e-245c9cea97ca" (UID: "353a4cd6-336b-4545-903e-245c9cea97ca"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:23:35 crc kubenswrapper[4955]: I0217 13:23:35.604817 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/353a4cd6-336b-4545-903e-245c9cea97ca-kube-api-access-58mxz" (OuterVolumeSpecName: "kube-api-access-58mxz") pod "353a4cd6-336b-4545-903e-245c9cea97ca" (UID: "353a4cd6-336b-4545-903e-245c9cea97ca"). InnerVolumeSpecName "kube-api-access-58mxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:23:35 crc kubenswrapper[4955]: I0217 13:23:35.606948 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/353a4cd6-336b-4545-903e-245c9cea97ca-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "353a4cd6-336b-4545-903e-245c9cea97ca" (UID: "353a4cd6-336b-4545-903e-245c9cea97ca"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:35 crc kubenswrapper[4955]: I0217 13:23:35.628405 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/353a4cd6-336b-4545-903e-245c9cea97ca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "353a4cd6-336b-4545-903e-245c9cea97ca" (UID: "353a4cd6-336b-4545-903e-245c9cea97ca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:35 crc kubenswrapper[4955]: I0217 13:23:35.644052 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/353a4cd6-336b-4545-903e-245c9cea97ca-config-data" (OuterVolumeSpecName: "config-data") pod "353a4cd6-336b-4545-903e-245c9cea97ca" (UID: "353a4cd6-336b-4545-903e-245c9cea97ca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:35 crc kubenswrapper[4955]: I0217 13:23:35.696072 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58mxz\" (UniqueName: \"kubernetes.io/projected/353a4cd6-336b-4545-903e-245c9cea97ca-kube-api-access-58mxz\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:35 crc kubenswrapper[4955]: I0217 13:23:35.696111 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/353a4cd6-336b-4545-903e-245c9cea97ca-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:35 crc kubenswrapper[4955]: I0217 13:23:35.696123 4955 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/353a4cd6-336b-4545-903e-245c9cea97ca-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:35 crc kubenswrapper[4955]: I0217 13:23:35.696135 4955 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/353a4cd6-336b-4545-903e-245c9cea97ca-logs\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:35 crc kubenswrapper[4955]: I0217 13:23:35.696146 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/353a4cd6-336b-4545-903e-245c9cea97ca-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:36 crc kubenswrapper[4955]: I0217 13:23:36.220101 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65b6bfc874-zgznk" event={"ID":"353a4cd6-336b-4545-903e-245c9cea97ca","Type":"ContainerDied","Data":"2499fcc3732e6ff01b91c42700f4df6b62b3a18843343f3bd5e83c56479e2b5b"} Feb 17 13:23:36 crc kubenswrapper[4955]: I0217 13:23:36.220156 4955 scope.go:117] "RemoveContainer" containerID="9e1cde53c1148029ca5f31661b223a8898cf091c0dc94ec858633087db1437ec" Feb 17 13:23:36 crc kubenswrapper[4955]: I0217 13:23:36.220110 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-65b6bfc874-zgznk" Feb 17 13:23:36 crc kubenswrapper[4955]: I0217 13:23:36.252383 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" event={"ID":"5bf36887-2a2b-4193-abdb-a15beaf46300","Type":"ContainerStarted","Data":"2a2497139f3e8754b75d83788b3f1744445e509419949dab14b01c6f81278a2c"} Feb 17 13:23:36 crc kubenswrapper[4955]: I0217 13:23:36.271039 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-65b6bfc874-zgznk"] Feb 17 13:23:36 crc kubenswrapper[4955]: I0217 13:23:36.275171 4955 scope.go:117] "RemoveContainer" containerID="7e9f2e7d38f46f5983cb4dfffe1c3cdfe6f350832d7f2c71a0ef1d278ec760f6" Feb 17 13:23:36 crc kubenswrapper[4955]: I0217 13:23:36.279671 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-65b6bfc874-zgznk"] Feb 17 13:23:36 crc kubenswrapper[4955]: I0217 13:23:36.868195 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Feb 17 13:23:36 crc kubenswrapper[4955]: E0217 13:23:36.868610 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="353a4cd6-336b-4545-903e-245c9cea97ca" containerName="barbican-api" Feb 17 13:23:36 crc kubenswrapper[4955]: I0217 13:23:36.868633 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="353a4cd6-336b-4545-903e-245c9cea97ca" containerName="barbican-api" Feb 17 13:23:36 crc kubenswrapper[4955]: E0217 13:23:36.868669 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="353a4cd6-336b-4545-903e-245c9cea97ca" containerName="barbican-api-log" Feb 17 13:23:36 crc kubenswrapper[4955]: I0217 13:23:36.868679 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="353a4cd6-336b-4545-903e-245c9cea97ca" containerName="barbican-api-log" Feb 17 13:23:36 crc kubenswrapper[4955]: I0217 13:23:36.868953 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="353a4cd6-336b-4545-903e-245c9cea97ca" containerName="barbican-api-log" Feb 17 13:23:36 crc kubenswrapper[4955]: I0217 13:23:36.868985 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="353a4cd6-336b-4545-903e-245c9cea97ca" containerName="barbican-api" Feb 17 13:23:36 crc kubenswrapper[4955]: I0217 13:23:36.869665 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 17 13:23:36 crc kubenswrapper[4955]: I0217 13:23:36.872281 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Feb 17 13:23:36 crc kubenswrapper[4955]: I0217 13:23:36.872475 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Feb 17 13:23:36 crc kubenswrapper[4955]: I0217 13:23:36.872487 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-vgrgt" Feb 17 13:23:36 crc kubenswrapper[4955]: I0217 13:23:36.887885 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 17 13:23:36 crc kubenswrapper[4955]: I0217 13:23:36.919978 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c2c452a7-9ff5-4f92-a5a9-87925a22302f-openstack-config\") pod \"openstackclient\" (UID: \"c2c452a7-9ff5-4f92-a5a9-87925a22302f\") " pod="openstack/openstackclient" Feb 17 13:23:36 crc kubenswrapper[4955]: I0217 13:23:36.920031 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2c452a7-9ff5-4f92-a5a9-87925a22302f-combined-ca-bundle\") pod \"openstackclient\" (UID: \"c2c452a7-9ff5-4f92-a5a9-87925a22302f\") " pod="openstack/openstackclient" Feb 17 13:23:36 crc kubenswrapper[4955]: I0217 13:23:36.920246 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktbtf\" (UniqueName: \"kubernetes.io/projected/c2c452a7-9ff5-4f92-a5a9-87925a22302f-kube-api-access-ktbtf\") pod \"openstackclient\" (UID: \"c2c452a7-9ff5-4f92-a5a9-87925a22302f\") " pod="openstack/openstackclient" Feb 17 13:23:36 crc kubenswrapper[4955]: I0217 13:23:36.920379 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c2c452a7-9ff5-4f92-a5a9-87925a22302f-openstack-config-secret\") pod \"openstackclient\" (UID: \"c2c452a7-9ff5-4f92-a5a9-87925a22302f\") " pod="openstack/openstackclient" Feb 17 13:23:37 crc kubenswrapper[4955]: I0217 13:23:37.021472 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c2c452a7-9ff5-4f92-a5a9-87925a22302f-openstack-config\") pod \"openstackclient\" (UID: \"c2c452a7-9ff5-4f92-a5a9-87925a22302f\") " pod="openstack/openstackclient" Feb 17 13:23:37 crc kubenswrapper[4955]: I0217 13:23:37.021519 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2c452a7-9ff5-4f92-a5a9-87925a22302f-combined-ca-bundle\") pod \"openstackclient\" (UID: \"c2c452a7-9ff5-4f92-a5a9-87925a22302f\") " pod="openstack/openstackclient" Feb 17 13:23:37 crc kubenswrapper[4955]: I0217 13:23:37.021594 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktbtf\" (UniqueName: \"kubernetes.io/projected/c2c452a7-9ff5-4f92-a5a9-87925a22302f-kube-api-access-ktbtf\") pod \"openstackclient\" (UID: \"c2c452a7-9ff5-4f92-a5a9-87925a22302f\") " pod="openstack/openstackclient" Feb 17 13:23:37 crc kubenswrapper[4955]: I0217 13:23:37.021650 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c2c452a7-9ff5-4f92-a5a9-87925a22302f-openstack-config-secret\") pod \"openstackclient\" (UID: \"c2c452a7-9ff5-4f92-a5a9-87925a22302f\") " pod="openstack/openstackclient" Feb 17 13:23:37 crc kubenswrapper[4955]: I0217 13:23:37.022610 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c2c452a7-9ff5-4f92-a5a9-87925a22302f-openstack-config\") pod \"openstackclient\" (UID: \"c2c452a7-9ff5-4f92-a5a9-87925a22302f\") " pod="openstack/openstackclient" Feb 17 13:23:37 crc kubenswrapper[4955]: I0217 13:23:37.030521 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c2c452a7-9ff5-4f92-a5a9-87925a22302f-openstack-config-secret\") pod \"openstackclient\" (UID: \"c2c452a7-9ff5-4f92-a5a9-87925a22302f\") " pod="openstack/openstackclient" Feb 17 13:23:37 crc kubenswrapper[4955]: I0217 13:23:37.030659 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2c452a7-9ff5-4f92-a5a9-87925a22302f-combined-ca-bundle\") pod \"openstackclient\" (UID: \"c2c452a7-9ff5-4f92-a5a9-87925a22302f\") " pod="openstack/openstackclient" Feb 17 13:23:37 crc kubenswrapper[4955]: I0217 13:23:37.044334 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktbtf\" (UniqueName: \"kubernetes.io/projected/c2c452a7-9ff5-4f92-a5a9-87925a22302f-kube-api-access-ktbtf\") pod \"openstackclient\" (UID: \"c2c452a7-9ff5-4f92-a5a9-87925a22302f\") " pod="openstack/openstackclient" Feb 17 13:23:37 crc kubenswrapper[4955]: I0217 13:23:37.202836 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 17 13:23:37 crc kubenswrapper[4955]: I0217 13:23:37.729641 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 17 13:23:38 crc kubenswrapper[4955]: I0217 13:23:38.234075 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="353a4cd6-336b-4545-903e-245c9cea97ca" path="/var/lib/kubelet/pods/353a4cd6-336b-4545-903e-245c9cea97ca/volumes" Feb 17 13:23:38 crc kubenswrapper[4955]: I0217 13:23:38.257341 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"c2c452a7-9ff5-4f92-a5a9-87925a22302f","Type":"ContainerStarted","Data":"bd898722a05ef7c8fce193d54d1658fd41c8ab3c51f4164e5153c70f365295f7"} Feb 17 13:23:38 crc kubenswrapper[4955]: I0217 13:23:38.812268 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Feb 17 13:23:40 crc kubenswrapper[4955]: I0217 13:23:40.820672 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Feb 17 13:23:41 crc kubenswrapper[4955]: I0217 13:23:41.519400 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-5bcc76b6cf-4r2g9"] Feb 17 13:23:41 crc kubenswrapper[4955]: I0217 13:23:41.523282 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5bcc76b6cf-4r2g9" Feb 17 13:23:41 crc kubenswrapper[4955]: I0217 13:23:41.525372 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Feb 17 13:23:41 crc kubenswrapper[4955]: I0217 13:23:41.525584 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Feb 17 13:23:41 crc kubenswrapper[4955]: I0217 13:23:41.531110 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Feb 17 13:23:41 crc kubenswrapper[4955]: I0217 13:23:41.541869 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5bcc76b6cf-4r2g9"] Feb 17 13:23:41 crc kubenswrapper[4955]: I0217 13:23:41.611371 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0fa0efd9-940a-4580-ab68-c5954c0e732e-run-httpd\") pod \"swift-proxy-5bcc76b6cf-4r2g9\" (UID: \"0fa0efd9-940a-4580-ab68-c5954c0e732e\") " pod="openstack/swift-proxy-5bcc76b6cf-4r2g9" Feb 17 13:23:41 crc kubenswrapper[4955]: I0217 13:23:41.611451 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fa0efd9-940a-4580-ab68-c5954c0e732e-public-tls-certs\") pod \"swift-proxy-5bcc76b6cf-4r2g9\" (UID: \"0fa0efd9-940a-4580-ab68-c5954c0e732e\") " pod="openstack/swift-proxy-5bcc76b6cf-4r2g9" Feb 17 13:23:41 crc kubenswrapper[4955]: I0217 13:23:41.611502 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fa0efd9-940a-4580-ab68-c5954c0e732e-internal-tls-certs\") pod \"swift-proxy-5bcc76b6cf-4r2g9\" (UID: \"0fa0efd9-940a-4580-ab68-c5954c0e732e\") " pod="openstack/swift-proxy-5bcc76b6cf-4r2g9" Feb 17 13:23:41 crc kubenswrapper[4955]: I0217 13:23:41.611545 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0fa0efd9-940a-4580-ab68-c5954c0e732e-etc-swift\") pod \"swift-proxy-5bcc76b6cf-4r2g9\" (UID: \"0fa0efd9-940a-4580-ab68-c5954c0e732e\") " pod="openstack/swift-proxy-5bcc76b6cf-4r2g9" Feb 17 13:23:41 crc kubenswrapper[4955]: I0217 13:23:41.611621 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fa0efd9-940a-4580-ab68-c5954c0e732e-config-data\") pod \"swift-proxy-5bcc76b6cf-4r2g9\" (UID: \"0fa0efd9-940a-4580-ab68-c5954c0e732e\") " pod="openstack/swift-proxy-5bcc76b6cf-4r2g9" Feb 17 13:23:41 crc kubenswrapper[4955]: I0217 13:23:41.611677 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4vbn\" (UniqueName: \"kubernetes.io/projected/0fa0efd9-940a-4580-ab68-c5954c0e732e-kube-api-access-m4vbn\") pod \"swift-proxy-5bcc76b6cf-4r2g9\" (UID: \"0fa0efd9-940a-4580-ab68-c5954c0e732e\") " pod="openstack/swift-proxy-5bcc76b6cf-4r2g9" Feb 17 13:23:41 crc kubenswrapper[4955]: I0217 13:23:41.611703 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fa0efd9-940a-4580-ab68-c5954c0e732e-combined-ca-bundle\") pod \"swift-proxy-5bcc76b6cf-4r2g9\" (UID: \"0fa0efd9-940a-4580-ab68-c5954c0e732e\") " pod="openstack/swift-proxy-5bcc76b6cf-4r2g9" Feb 17 13:23:41 crc kubenswrapper[4955]: I0217 13:23:41.611941 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0fa0efd9-940a-4580-ab68-c5954c0e732e-log-httpd\") pod \"swift-proxy-5bcc76b6cf-4r2g9\" (UID: \"0fa0efd9-940a-4580-ab68-c5954c0e732e\") " pod="openstack/swift-proxy-5bcc76b6cf-4r2g9" Feb 17 13:23:41 crc kubenswrapper[4955]: I0217 13:23:41.713798 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0fa0efd9-940a-4580-ab68-c5954c0e732e-log-httpd\") pod \"swift-proxy-5bcc76b6cf-4r2g9\" (UID: \"0fa0efd9-940a-4580-ab68-c5954c0e732e\") " pod="openstack/swift-proxy-5bcc76b6cf-4r2g9" Feb 17 13:23:41 crc kubenswrapper[4955]: I0217 13:23:41.713872 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0fa0efd9-940a-4580-ab68-c5954c0e732e-run-httpd\") pod \"swift-proxy-5bcc76b6cf-4r2g9\" (UID: \"0fa0efd9-940a-4580-ab68-c5954c0e732e\") " pod="openstack/swift-proxy-5bcc76b6cf-4r2g9" Feb 17 13:23:41 crc kubenswrapper[4955]: I0217 13:23:41.713903 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fa0efd9-940a-4580-ab68-c5954c0e732e-public-tls-certs\") pod \"swift-proxy-5bcc76b6cf-4r2g9\" (UID: \"0fa0efd9-940a-4580-ab68-c5954c0e732e\") " pod="openstack/swift-proxy-5bcc76b6cf-4r2g9" Feb 17 13:23:41 crc kubenswrapper[4955]: I0217 13:23:41.713932 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fa0efd9-940a-4580-ab68-c5954c0e732e-internal-tls-certs\") pod \"swift-proxy-5bcc76b6cf-4r2g9\" (UID: \"0fa0efd9-940a-4580-ab68-c5954c0e732e\") " pod="openstack/swift-proxy-5bcc76b6cf-4r2g9" Feb 17 13:23:41 crc kubenswrapper[4955]: I0217 13:23:41.713960 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0fa0efd9-940a-4580-ab68-c5954c0e732e-etc-swift\") pod \"swift-proxy-5bcc76b6cf-4r2g9\" (UID: \"0fa0efd9-940a-4580-ab68-c5954c0e732e\") " pod="openstack/swift-proxy-5bcc76b6cf-4r2g9" Feb 17 13:23:41 crc kubenswrapper[4955]: I0217 13:23:41.713999 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fa0efd9-940a-4580-ab68-c5954c0e732e-config-data\") pod \"swift-proxy-5bcc76b6cf-4r2g9\" (UID: \"0fa0efd9-940a-4580-ab68-c5954c0e732e\") " pod="openstack/swift-proxy-5bcc76b6cf-4r2g9" Feb 17 13:23:41 crc kubenswrapper[4955]: I0217 13:23:41.714038 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4vbn\" (UniqueName: \"kubernetes.io/projected/0fa0efd9-940a-4580-ab68-c5954c0e732e-kube-api-access-m4vbn\") pod \"swift-proxy-5bcc76b6cf-4r2g9\" (UID: \"0fa0efd9-940a-4580-ab68-c5954c0e732e\") " pod="openstack/swift-proxy-5bcc76b6cf-4r2g9" Feb 17 13:23:41 crc kubenswrapper[4955]: I0217 13:23:41.714059 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fa0efd9-940a-4580-ab68-c5954c0e732e-combined-ca-bundle\") pod \"swift-proxy-5bcc76b6cf-4r2g9\" (UID: \"0fa0efd9-940a-4580-ab68-c5954c0e732e\") " pod="openstack/swift-proxy-5bcc76b6cf-4r2g9" Feb 17 13:23:41 crc kubenswrapper[4955]: I0217 13:23:41.714242 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0fa0efd9-940a-4580-ab68-c5954c0e732e-log-httpd\") pod \"swift-proxy-5bcc76b6cf-4r2g9\" (UID: \"0fa0efd9-940a-4580-ab68-c5954c0e732e\") " pod="openstack/swift-proxy-5bcc76b6cf-4r2g9" Feb 17 13:23:41 crc kubenswrapper[4955]: I0217 13:23:41.715377 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0fa0efd9-940a-4580-ab68-c5954c0e732e-run-httpd\") pod \"swift-proxy-5bcc76b6cf-4r2g9\" (UID: \"0fa0efd9-940a-4580-ab68-c5954c0e732e\") " pod="openstack/swift-proxy-5bcc76b6cf-4r2g9" Feb 17 13:23:41 crc kubenswrapper[4955]: I0217 13:23:41.720963 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fa0efd9-940a-4580-ab68-c5954c0e732e-public-tls-certs\") pod \"swift-proxy-5bcc76b6cf-4r2g9\" (UID: \"0fa0efd9-940a-4580-ab68-c5954c0e732e\") " pod="openstack/swift-proxy-5bcc76b6cf-4r2g9" Feb 17 13:23:41 crc kubenswrapper[4955]: I0217 13:23:41.721113 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fa0efd9-940a-4580-ab68-c5954c0e732e-combined-ca-bundle\") pod \"swift-proxy-5bcc76b6cf-4r2g9\" (UID: \"0fa0efd9-940a-4580-ab68-c5954c0e732e\") " pod="openstack/swift-proxy-5bcc76b6cf-4r2g9" Feb 17 13:23:41 crc kubenswrapper[4955]: I0217 13:23:41.721269 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fa0efd9-940a-4580-ab68-c5954c0e732e-config-data\") pod \"swift-proxy-5bcc76b6cf-4r2g9\" (UID: \"0fa0efd9-940a-4580-ab68-c5954c0e732e\") " pod="openstack/swift-proxy-5bcc76b6cf-4r2g9" Feb 17 13:23:41 crc kubenswrapper[4955]: I0217 13:23:41.723376 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fa0efd9-940a-4580-ab68-c5954c0e732e-internal-tls-certs\") pod \"swift-proxy-5bcc76b6cf-4r2g9\" (UID: \"0fa0efd9-940a-4580-ab68-c5954c0e732e\") " pod="openstack/swift-proxy-5bcc76b6cf-4r2g9" Feb 17 13:23:41 crc kubenswrapper[4955]: I0217 13:23:41.735895 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0fa0efd9-940a-4580-ab68-c5954c0e732e-etc-swift\") pod \"swift-proxy-5bcc76b6cf-4r2g9\" (UID: \"0fa0efd9-940a-4580-ab68-c5954c0e732e\") " pod="openstack/swift-proxy-5bcc76b6cf-4r2g9" Feb 17 13:23:41 crc kubenswrapper[4955]: I0217 13:23:41.745242 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4vbn\" (UniqueName: \"kubernetes.io/projected/0fa0efd9-940a-4580-ab68-c5954c0e732e-kube-api-access-m4vbn\") pod \"swift-proxy-5bcc76b6cf-4r2g9\" (UID: \"0fa0efd9-940a-4580-ab68-c5954c0e732e\") " pod="openstack/swift-proxy-5bcc76b6cf-4r2g9" Feb 17 13:23:41 crc kubenswrapper[4955]: I0217 13:23:41.843263 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5bcc76b6cf-4r2g9" Feb 17 13:23:43 crc kubenswrapper[4955]: I0217 13:23:43.648822 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5788cc4448-245bs" podUID="1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Feb 17 13:23:43 crc kubenswrapper[4955]: I0217 13:23:43.649197 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5788cc4448-245bs" Feb 17 13:23:43 crc kubenswrapper[4955]: I0217 13:23:43.848072 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:23:43 crc kubenswrapper[4955]: I0217 13:23:43.848482 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d20f3280-ab79-434f-a89a-11d6fbfc0edf" containerName="proxy-httpd" containerID="cri-o://4cf2d0a3c593da686ace3c484b767ecbf30fa77a09a675c837b3c46000475432" gracePeriod=30 Feb 17 13:23:43 crc kubenswrapper[4955]: I0217 13:23:43.848529 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d20f3280-ab79-434f-a89a-11d6fbfc0edf" containerName="sg-core" containerID="cri-o://867f11c8b8d3d4d6f0679b281725a532c1d72c6f1cbca9ed91bfbb6f366fd041" gracePeriod=30 Feb 17 13:23:43 crc kubenswrapper[4955]: I0217 13:23:43.849014 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d20f3280-ab79-434f-a89a-11d6fbfc0edf" containerName="ceilometer-notification-agent" containerID="cri-o://812e7e1ffda69c6175b2a09a089bce147530345353d05c3617de2af0ebb6f577" gracePeriod=30 Feb 17 13:23:43 crc kubenswrapper[4955]: I0217 13:23:43.850050 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d20f3280-ab79-434f-a89a-11d6fbfc0edf" containerName="ceilometer-central-agent" containerID="cri-o://e0798da9a93b55d7c2bbe07fbb48b498acc48e833fefe3eff9186170ec9f0299" gracePeriod=30 Feb 17 13:23:43 crc kubenswrapper[4955]: I0217 13:23:43.864754 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="d20f3280-ab79-434f-a89a-11d6fbfc0edf" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Feb 17 13:23:44 crc kubenswrapper[4955]: I0217 13:23:44.326732 4955 generic.go:334] "Generic (PLEG): container finished" podID="d20f3280-ab79-434f-a89a-11d6fbfc0edf" containerID="4cf2d0a3c593da686ace3c484b767ecbf30fa77a09a675c837b3c46000475432" exitCode=0 Feb 17 13:23:44 crc kubenswrapper[4955]: I0217 13:23:44.326761 4955 generic.go:334] "Generic (PLEG): container finished" podID="d20f3280-ab79-434f-a89a-11d6fbfc0edf" containerID="867f11c8b8d3d4d6f0679b281725a532c1d72c6f1cbca9ed91bfbb6f366fd041" exitCode=2 Feb 17 13:23:44 crc kubenswrapper[4955]: I0217 13:23:44.326768 4955 generic.go:334] "Generic (PLEG): container finished" podID="d20f3280-ab79-434f-a89a-11d6fbfc0edf" containerID="e0798da9a93b55d7c2bbe07fbb48b498acc48e833fefe3eff9186170ec9f0299" exitCode=0 Feb 17 13:23:44 crc kubenswrapper[4955]: I0217 13:23:44.326807 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d20f3280-ab79-434f-a89a-11d6fbfc0edf","Type":"ContainerDied","Data":"4cf2d0a3c593da686ace3c484b767ecbf30fa77a09a675c837b3c46000475432"} Feb 17 13:23:44 crc kubenswrapper[4955]: I0217 13:23:44.326886 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d20f3280-ab79-434f-a89a-11d6fbfc0edf","Type":"ContainerDied","Data":"867f11c8b8d3d4d6f0679b281725a532c1d72c6f1cbca9ed91bfbb6f366fd041"} Feb 17 13:23:44 crc kubenswrapper[4955]: I0217 13:23:44.326909 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d20f3280-ab79-434f-a89a-11d6fbfc0edf","Type":"ContainerDied","Data":"e0798da9a93b55d7c2bbe07fbb48b498acc48e833fefe3eff9186170ec9f0299"} Feb 17 13:23:45 crc kubenswrapper[4955]: I0217 13:23:45.338253 4955 generic.go:334] "Generic (PLEG): container finished" podID="d20f3280-ab79-434f-a89a-11d6fbfc0edf" containerID="812e7e1ffda69c6175b2a09a089bce147530345353d05c3617de2af0ebb6f577" exitCode=0 Feb 17 13:23:45 crc kubenswrapper[4955]: I0217 13:23:45.338518 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d20f3280-ab79-434f-a89a-11d6fbfc0edf","Type":"ContainerDied","Data":"812e7e1ffda69c6175b2a09a089bce147530345353d05c3617de2af0ebb6f577"} Feb 17 13:23:46 crc kubenswrapper[4955]: I0217 13:23:46.822309 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 17 13:23:46 crc kubenswrapper[4955]: I0217 13:23:46.924020 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d20f3280-ab79-434f-a89a-11d6fbfc0edf-run-httpd\") pod \"d20f3280-ab79-434f-a89a-11d6fbfc0edf\" (UID: \"d20f3280-ab79-434f-a89a-11d6fbfc0edf\") " Feb 17 13:23:46 crc kubenswrapper[4955]: I0217 13:23:46.924172 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d20f3280-ab79-434f-a89a-11d6fbfc0edf-config-data\") pod \"d20f3280-ab79-434f-a89a-11d6fbfc0edf\" (UID: \"d20f3280-ab79-434f-a89a-11d6fbfc0edf\") " Feb 17 13:23:46 crc kubenswrapper[4955]: I0217 13:23:46.924256 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rslhr\" (UniqueName: \"kubernetes.io/projected/d20f3280-ab79-434f-a89a-11d6fbfc0edf-kube-api-access-rslhr\") pod \"d20f3280-ab79-434f-a89a-11d6fbfc0edf\" (UID: \"d20f3280-ab79-434f-a89a-11d6fbfc0edf\") " Feb 17 13:23:46 crc kubenswrapper[4955]: I0217 13:23:46.924317 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d20f3280-ab79-434f-a89a-11d6fbfc0edf-log-httpd\") pod \"d20f3280-ab79-434f-a89a-11d6fbfc0edf\" (UID: \"d20f3280-ab79-434f-a89a-11d6fbfc0edf\") " Feb 17 13:23:46 crc kubenswrapper[4955]: I0217 13:23:46.924348 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d20f3280-ab79-434f-a89a-11d6fbfc0edf-scripts\") pod \"d20f3280-ab79-434f-a89a-11d6fbfc0edf\" (UID: \"d20f3280-ab79-434f-a89a-11d6fbfc0edf\") " Feb 17 13:23:46 crc kubenswrapper[4955]: I0217 13:23:46.924390 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d20f3280-ab79-434f-a89a-11d6fbfc0edf-combined-ca-bundle\") pod \"d20f3280-ab79-434f-a89a-11d6fbfc0edf\" (UID: \"d20f3280-ab79-434f-a89a-11d6fbfc0edf\") " Feb 17 13:23:46 crc kubenswrapper[4955]: I0217 13:23:46.924510 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d20f3280-ab79-434f-a89a-11d6fbfc0edf-sg-core-conf-yaml\") pod \"d20f3280-ab79-434f-a89a-11d6fbfc0edf\" (UID: \"d20f3280-ab79-434f-a89a-11d6fbfc0edf\") " Feb 17 13:23:46 crc kubenswrapper[4955]: I0217 13:23:46.924732 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d20f3280-ab79-434f-a89a-11d6fbfc0edf-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d20f3280-ab79-434f-a89a-11d6fbfc0edf" (UID: "d20f3280-ab79-434f-a89a-11d6fbfc0edf"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:23:46 crc kubenswrapper[4955]: I0217 13:23:46.925003 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d20f3280-ab79-434f-a89a-11d6fbfc0edf-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d20f3280-ab79-434f-a89a-11d6fbfc0edf" (UID: "d20f3280-ab79-434f-a89a-11d6fbfc0edf"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:23:46 crc kubenswrapper[4955]: I0217 13:23:46.925033 4955 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d20f3280-ab79-434f-a89a-11d6fbfc0edf-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:46 crc kubenswrapper[4955]: I0217 13:23:46.928571 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d20f3280-ab79-434f-a89a-11d6fbfc0edf-kube-api-access-rslhr" (OuterVolumeSpecName: "kube-api-access-rslhr") pod "d20f3280-ab79-434f-a89a-11d6fbfc0edf" (UID: "d20f3280-ab79-434f-a89a-11d6fbfc0edf"). InnerVolumeSpecName "kube-api-access-rslhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:23:46 crc kubenswrapper[4955]: I0217 13:23:46.928631 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d20f3280-ab79-434f-a89a-11d6fbfc0edf-scripts" (OuterVolumeSpecName: "scripts") pod "d20f3280-ab79-434f-a89a-11d6fbfc0edf" (UID: "d20f3280-ab79-434f-a89a-11d6fbfc0edf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:46 crc kubenswrapper[4955]: I0217 13:23:46.954430 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d20f3280-ab79-434f-a89a-11d6fbfc0edf-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d20f3280-ab79-434f-a89a-11d6fbfc0edf" (UID: "d20f3280-ab79-434f-a89a-11d6fbfc0edf"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.026850 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d20f3280-ab79-434f-a89a-11d6fbfc0edf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d20f3280-ab79-434f-a89a-11d6fbfc0edf" (UID: "d20f3280-ab79-434f-a89a-11d6fbfc0edf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.028037 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rslhr\" (UniqueName: \"kubernetes.io/projected/d20f3280-ab79-434f-a89a-11d6fbfc0edf-kube-api-access-rslhr\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.028074 4955 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d20f3280-ab79-434f-a89a-11d6fbfc0edf-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.028092 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d20f3280-ab79-434f-a89a-11d6fbfc0edf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.028108 4955 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d20f3280-ab79-434f-a89a-11d6fbfc0edf-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.028122 4955 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d20f3280-ab79-434f-a89a-11d6fbfc0edf-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.047313 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d20f3280-ab79-434f-a89a-11d6fbfc0edf-config-data" (OuterVolumeSpecName: "config-data") pod "d20f3280-ab79-434f-a89a-11d6fbfc0edf" (UID: "d20f3280-ab79-434f-a89a-11d6fbfc0edf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.086408 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-9ms22"] Feb 17 13:23:47 crc kubenswrapper[4955]: E0217 13:23:47.087179 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d20f3280-ab79-434f-a89a-11d6fbfc0edf" containerName="sg-core" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.087205 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="d20f3280-ab79-434f-a89a-11d6fbfc0edf" containerName="sg-core" Feb 17 13:23:47 crc kubenswrapper[4955]: E0217 13:23:47.087217 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d20f3280-ab79-434f-a89a-11d6fbfc0edf" containerName="ceilometer-notification-agent" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.087225 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="d20f3280-ab79-434f-a89a-11d6fbfc0edf" containerName="ceilometer-notification-agent" Feb 17 13:23:47 crc kubenswrapper[4955]: E0217 13:23:47.087246 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d20f3280-ab79-434f-a89a-11d6fbfc0edf" containerName="ceilometer-central-agent" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.087254 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="d20f3280-ab79-434f-a89a-11d6fbfc0edf" containerName="ceilometer-central-agent" Feb 17 13:23:47 crc kubenswrapper[4955]: E0217 13:23:47.087277 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d20f3280-ab79-434f-a89a-11d6fbfc0edf" containerName="proxy-httpd" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.087284 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="d20f3280-ab79-434f-a89a-11d6fbfc0edf" containerName="proxy-httpd" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.087506 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="d20f3280-ab79-434f-a89a-11d6fbfc0edf" containerName="ceilometer-notification-agent" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.087524 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="d20f3280-ab79-434f-a89a-11d6fbfc0edf" containerName="proxy-httpd" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.087542 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="d20f3280-ab79-434f-a89a-11d6fbfc0edf" containerName="ceilometer-central-agent" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.087554 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="d20f3280-ab79-434f-a89a-11d6fbfc0edf" containerName="sg-core" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.088315 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-9ms22" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.098629 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-9ms22"] Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.129789 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fkd2\" (UniqueName: \"kubernetes.io/projected/ff0b8526-a0fe-425e-92e8-a4933f3faf61-kube-api-access-4fkd2\") pod \"nova-api-db-create-9ms22\" (UID: \"ff0b8526-a0fe-425e-92e8-a4933f3faf61\") " pod="openstack/nova-api-db-create-9ms22" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.129905 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff0b8526-a0fe-425e-92e8-a4933f3faf61-operator-scripts\") pod \"nova-api-db-create-9ms22\" (UID: \"ff0b8526-a0fe-425e-92e8-a4933f3faf61\") " pod="openstack/nova-api-db-create-9ms22" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.129952 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d20f3280-ab79-434f-a89a-11d6fbfc0edf-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.192248 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5bcc76b6cf-4r2g9"] Feb 17 13:23:47 crc kubenswrapper[4955]: W0217 13:23:47.192559 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0fa0efd9_940a_4580_ab68_c5954c0e732e.slice/crio-60c01737518118422362bf3a61ecb04c287cfe843ccf4b274cd6bcec80d61593 WatchSource:0}: Error finding container 60c01737518118422362bf3a61ecb04c287cfe843ccf4b274cd6bcec80d61593: Status 404 returned error can't find the container with id 60c01737518118422362bf3a61ecb04c287cfe843ccf4b274cd6bcec80d61593 Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.231322 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fkd2\" (UniqueName: \"kubernetes.io/projected/ff0b8526-a0fe-425e-92e8-a4933f3faf61-kube-api-access-4fkd2\") pod \"nova-api-db-create-9ms22\" (UID: \"ff0b8526-a0fe-425e-92e8-a4933f3faf61\") " pod="openstack/nova-api-db-create-9ms22" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.231459 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff0b8526-a0fe-425e-92e8-a4933f3faf61-operator-scripts\") pod \"nova-api-db-create-9ms22\" (UID: \"ff0b8526-a0fe-425e-92e8-a4933f3faf61\") " pod="openstack/nova-api-db-create-9ms22" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.232404 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff0b8526-a0fe-425e-92e8-a4933f3faf61-operator-scripts\") pod \"nova-api-db-create-9ms22\" (UID: \"ff0b8526-a0fe-425e-92e8-a4933f3faf61\") " pod="openstack/nova-api-db-create-9ms22" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.263434 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fkd2\" (UniqueName: \"kubernetes.io/projected/ff0b8526-a0fe-425e-92e8-a4933f3faf61-kube-api-access-4fkd2\") pod \"nova-api-db-create-9ms22\" (UID: \"ff0b8526-a0fe-425e-92e8-a4933f3faf61\") " pod="openstack/nova-api-db-create-9ms22" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.283521 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-6qshk"] Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.285883 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-6qshk" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.297842 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-301f-account-create-update-wc7xh"] Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.298983 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-301f-account-create-update-wc7xh" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.301347 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.328845 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-6qshk"] Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.334177 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c4bbd808-631c-4a9b-81e6-dee199484728-operator-scripts\") pod \"nova-cell0-db-create-6qshk\" (UID: \"c4bbd808-631c-4a9b-81e6-dee199484728\") " pod="openstack/nova-cell0-db-create-6qshk" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.334211 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcnr8\" (UniqueName: \"kubernetes.io/projected/c4bbd808-631c-4a9b-81e6-dee199484728-kube-api-access-mcnr8\") pod \"nova-cell0-db-create-6qshk\" (UID: \"c4bbd808-631c-4a9b-81e6-dee199484728\") " pod="openstack/nova-cell0-db-create-6qshk" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.334247 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84b8be7f-2891-44a0-beb3-850911ac8b2a-operator-scripts\") pod \"nova-api-301f-account-create-update-wc7xh\" (UID: \"84b8be7f-2891-44a0-beb3-850911ac8b2a\") " pod="openstack/nova-api-301f-account-create-update-wc7xh" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.334302 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bns5\" (UniqueName: \"kubernetes.io/projected/84b8be7f-2891-44a0-beb3-850911ac8b2a-kube-api-access-8bns5\") pod \"nova-api-301f-account-create-update-wc7xh\" (UID: \"84b8be7f-2891-44a0-beb3-850911ac8b2a\") " pod="openstack/nova-api-301f-account-create-update-wc7xh" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.350083 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-301f-account-create-update-wc7xh"] Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.385842 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d20f3280-ab79-434f-a89a-11d6fbfc0edf","Type":"ContainerDied","Data":"74359edeb30c4e7e7464c3a8d030d1e01cfdbd7f32b7d2c458c9d18b55ebb128"} Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.386059 4955 scope.go:117] "RemoveContainer" containerID="4cf2d0a3c593da686ace3c484b767ecbf30fa77a09a675c837b3c46000475432" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.386066 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.388269 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5bcc76b6cf-4r2g9" event={"ID":"0fa0efd9-940a-4580-ab68-c5954c0e732e","Type":"ContainerStarted","Data":"60c01737518118422362bf3a61ecb04c287cfe843ccf4b274cd6bcec80d61593"} Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.394484 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"c2c452a7-9ff5-4f92-a5a9-87925a22302f","Type":"ContainerStarted","Data":"fb3b194acdcbce4310f8b22b4f8d9d5ee50fb114073be7f03ef8d729fd4a1f47"} Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.414898 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-9ms22" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.419143 4955 scope.go:117] "RemoveContainer" containerID="867f11c8b8d3d4d6f0679b281725a532c1d72c6f1cbca9ed91bfbb6f366fd041" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.432932 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-rc66h"] Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.434226 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-rc66h" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.437375 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c4bbd808-631c-4a9b-81e6-dee199484728-operator-scripts\") pod \"nova-cell0-db-create-6qshk\" (UID: \"c4bbd808-631c-4a9b-81e6-dee199484728\") " pod="openstack/nova-cell0-db-create-6qshk" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.437452 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcnr8\" (UniqueName: \"kubernetes.io/projected/c4bbd808-631c-4a9b-81e6-dee199484728-kube-api-access-mcnr8\") pod \"nova-cell0-db-create-6qshk\" (UID: \"c4bbd808-631c-4a9b-81e6-dee199484728\") " pod="openstack/nova-cell0-db-create-6qshk" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.438100 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84b8be7f-2891-44a0-beb3-850911ac8b2a-operator-scripts\") pod \"nova-api-301f-account-create-update-wc7xh\" (UID: \"84b8be7f-2891-44a0-beb3-850911ac8b2a\") " pod="openstack/nova-api-301f-account-create-update-wc7xh" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.438358 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bns5\" (UniqueName: \"kubernetes.io/projected/84b8be7f-2891-44a0-beb3-850911ac8b2a-kube-api-access-8bns5\") pod \"nova-api-301f-account-create-update-wc7xh\" (UID: \"84b8be7f-2891-44a0-beb3-850911ac8b2a\") " pod="openstack/nova-api-301f-account-create-update-wc7xh" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.439633 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84b8be7f-2891-44a0-beb3-850911ac8b2a-operator-scripts\") pod \"nova-api-301f-account-create-update-wc7xh\" (UID: \"84b8be7f-2891-44a0-beb3-850911ac8b2a\") " pod="openstack/nova-api-301f-account-create-update-wc7xh" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.440474 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c4bbd808-631c-4a9b-81e6-dee199484728-operator-scripts\") pod \"nova-cell0-db-create-6qshk\" (UID: \"c4bbd808-631c-4a9b-81e6-dee199484728\") " pod="openstack/nova-cell0-db-create-6qshk" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.455165 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-rc66h"] Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.458513 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bns5\" (UniqueName: \"kubernetes.io/projected/84b8be7f-2891-44a0-beb3-850911ac8b2a-kube-api-access-8bns5\") pod \"nova-api-301f-account-create-update-wc7xh\" (UID: \"84b8be7f-2891-44a0-beb3-850911ac8b2a\") " pod="openstack/nova-api-301f-account-create-update-wc7xh" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.466212 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcnr8\" (UniqueName: \"kubernetes.io/projected/c4bbd808-631c-4a9b-81e6-dee199484728-kube-api-access-mcnr8\") pod \"nova-cell0-db-create-6qshk\" (UID: \"c4bbd808-631c-4a9b-81e6-dee199484728\") " pod="openstack/nova-cell0-db-create-6qshk" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.471093 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.57569313 podStartE2EDuration="11.471066915s" podCreationTimestamp="2026-02-17 13:23:36 +0000 UTC" firstStartedPulling="2026-02-17 13:23:37.71289685 +0000 UTC m=+1156.235626393" lastFinishedPulling="2026-02-17 13:23:46.608270635 +0000 UTC m=+1165.131000178" observedRunningTime="2026-02-17 13:23:47.430421926 +0000 UTC m=+1165.953151469" watchObservedRunningTime="2026-02-17 13:23:47.471066915 +0000 UTC m=+1165.993796458" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.498219 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.508129 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.511921 4955 scope.go:117] "RemoveContainer" containerID="812e7e1ffda69c6175b2a09a089bce147530345353d05c3617de2af0ebb6f577" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.517318 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.520214 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.526492 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.526502 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.545483 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4031cf36-67c1-425d-98f9-0c4f3a4a6065-operator-scripts\") pod \"nova-cell1-db-create-rc66h\" (UID: \"4031cf36-67c1-425d-98f9-0c4f3a4a6065\") " pod="openstack/nova-cell1-db-create-rc66h" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.549061 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zsp5\" (UniqueName: \"kubernetes.io/projected/4031cf36-67c1-425d-98f9-0c4f3a4a6065-kube-api-access-6zsp5\") pod \"nova-cell1-db-create-rc66h\" (UID: \"4031cf36-67c1-425d-98f9-0c4f3a4a6065\") " pod="openstack/nova-cell1-db-create-rc66h" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.585910 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-a9af-account-create-update-l42hs"] Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.587187 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-a9af-account-create-update-l42hs" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.589515 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.613389 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.619176 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-6qshk" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.626140 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-301f-account-create-update-wc7xh" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.632323 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-a9af-account-create-update-l42hs"] Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.651000 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e56111b4-4b8a-4e07-bba9-68d34fd179fb-log-httpd\") pod \"ceilometer-0\" (UID: \"e56111b4-4b8a-4e07-bba9-68d34fd179fb\") " pod="openstack/ceilometer-0" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.651057 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e56111b4-4b8a-4e07-bba9-68d34fd179fb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e56111b4-4b8a-4e07-bba9-68d34fd179fb\") " pod="openstack/ceilometer-0" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.651086 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zsp5\" (UniqueName: \"kubernetes.io/projected/4031cf36-67c1-425d-98f9-0c4f3a4a6065-kube-api-access-6zsp5\") pod \"nova-cell1-db-create-rc66h\" (UID: \"4031cf36-67c1-425d-98f9-0c4f3a4a6065\") " pod="openstack/nova-cell1-db-create-rc66h" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.651115 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf4h6\" (UniqueName: \"kubernetes.io/projected/e56111b4-4b8a-4e07-bba9-68d34fd179fb-kube-api-access-wf4h6\") pod \"ceilometer-0\" (UID: \"e56111b4-4b8a-4e07-bba9-68d34fd179fb\") " pod="openstack/ceilometer-0" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.651309 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4031cf36-67c1-425d-98f9-0c4f3a4a6065-operator-scripts\") pod \"nova-cell1-db-create-rc66h\" (UID: \"4031cf36-67c1-425d-98f9-0c4f3a4a6065\") " pod="openstack/nova-cell1-db-create-rc66h" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.651338 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e56111b4-4b8a-4e07-bba9-68d34fd179fb-scripts\") pod \"ceilometer-0\" (UID: \"e56111b4-4b8a-4e07-bba9-68d34fd179fb\") " pod="openstack/ceilometer-0" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.651362 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e56111b4-4b8a-4e07-bba9-68d34fd179fb-run-httpd\") pod \"ceilometer-0\" (UID: \"e56111b4-4b8a-4e07-bba9-68d34fd179fb\") " pod="openstack/ceilometer-0" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.651413 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e56111b4-4b8a-4e07-bba9-68d34fd179fb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e56111b4-4b8a-4e07-bba9-68d34fd179fb\") " pod="openstack/ceilometer-0" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.651436 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3d747bbd-45e4-4c6c-a6cb-64a259440231-operator-scripts\") pod \"nova-cell0-a9af-account-create-update-l42hs\" (UID: \"3d747bbd-45e4-4c6c-a6cb-64a259440231\") " pod="openstack/nova-cell0-a9af-account-create-update-l42hs" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.651453 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rnf9\" (UniqueName: \"kubernetes.io/projected/3d747bbd-45e4-4c6c-a6cb-64a259440231-kube-api-access-5rnf9\") pod \"nova-cell0-a9af-account-create-update-l42hs\" (UID: \"3d747bbd-45e4-4c6c-a6cb-64a259440231\") " pod="openstack/nova-cell0-a9af-account-create-update-l42hs" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.651480 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e56111b4-4b8a-4e07-bba9-68d34fd179fb-config-data\") pod \"ceilometer-0\" (UID: \"e56111b4-4b8a-4e07-bba9-68d34fd179fb\") " pod="openstack/ceilometer-0" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.653502 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4031cf36-67c1-425d-98f9-0c4f3a4a6065-operator-scripts\") pod \"nova-cell1-db-create-rc66h\" (UID: \"4031cf36-67c1-425d-98f9-0c4f3a4a6065\") " pod="openstack/nova-cell1-db-create-rc66h" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.679899 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-aa00-account-create-update-r9kcs"] Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.680628 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zsp5\" (UniqueName: \"kubernetes.io/projected/4031cf36-67c1-425d-98f9-0c4f3a4a6065-kube-api-access-6zsp5\") pod \"nova-cell1-db-create-rc66h\" (UID: \"4031cf36-67c1-425d-98f9-0c4f3a4a6065\") " pod="openstack/nova-cell1-db-create-rc66h" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.681029 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-aa00-account-create-update-r9kcs" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.682580 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.688125 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-aa00-account-create-update-r9kcs"] Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.714977 4955 scope.go:117] "RemoveContainer" containerID="e0798da9a93b55d7c2bbe07fbb48b498acc48e833fefe3eff9186170ec9f0299" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.753940 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e56111b4-4b8a-4e07-bba9-68d34fd179fb-scripts\") pod \"ceilometer-0\" (UID: \"e56111b4-4b8a-4e07-bba9-68d34fd179fb\") " pod="openstack/ceilometer-0" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.753989 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e56111b4-4b8a-4e07-bba9-68d34fd179fb-run-httpd\") pod \"ceilometer-0\" (UID: \"e56111b4-4b8a-4e07-bba9-68d34fd179fb\") " pod="openstack/ceilometer-0" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.754008 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e56111b4-4b8a-4e07-bba9-68d34fd179fb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e56111b4-4b8a-4e07-bba9-68d34fd179fb\") " pod="openstack/ceilometer-0" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.754031 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rnf9\" (UniqueName: \"kubernetes.io/projected/3d747bbd-45e4-4c6c-a6cb-64a259440231-kube-api-access-5rnf9\") pod \"nova-cell0-a9af-account-create-update-l42hs\" (UID: \"3d747bbd-45e4-4c6c-a6cb-64a259440231\") " pod="openstack/nova-cell0-a9af-account-create-update-l42hs" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.754072 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3d747bbd-45e4-4c6c-a6cb-64a259440231-operator-scripts\") pod \"nova-cell0-a9af-account-create-update-l42hs\" (UID: \"3d747bbd-45e4-4c6c-a6cb-64a259440231\") " pod="openstack/nova-cell0-a9af-account-create-update-l42hs" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.754100 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e56111b4-4b8a-4e07-bba9-68d34fd179fb-config-data\") pod \"ceilometer-0\" (UID: \"e56111b4-4b8a-4e07-bba9-68d34fd179fb\") " pod="openstack/ceilometer-0" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.754132 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d0c9416-2de8-4d45-8818-87ae59a2cd6f-operator-scripts\") pod \"nova-cell1-aa00-account-create-update-r9kcs\" (UID: \"0d0c9416-2de8-4d45-8818-87ae59a2cd6f\") " pod="openstack/nova-cell1-aa00-account-create-update-r9kcs" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.754156 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcpqr\" (UniqueName: \"kubernetes.io/projected/0d0c9416-2de8-4d45-8818-87ae59a2cd6f-kube-api-access-bcpqr\") pod \"nova-cell1-aa00-account-create-update-r9kcs\" (UID: \"0d0c9416-2de8-4d45-8818-87ae59a2cd6f\") " pod="openstack/nova-cell1-aa00-account-create-update-r9kcs" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.754204 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e56111b4-4b8a-4e07-bba9-68d34fd179fb-log-httpd\") pod \"ceilometer-0\" (UID: \"e56111b4-4b8a-4e07-bba9-68d34fd179fb\") " pod="openstack/ceilometer-0" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.754231 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e56111b4-4b8a-4e07-bba9-68d34fd179fb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e56111b4-4b8a-4e07-bba9-68d34fd179fb\") " pod="openstack/ceilometer-0" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.754263 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wf4h6\" (UniqueName: \"kubernetes.io/projected/e56111b4-4b8a-4e07-bba9-68d34fd179fb-kube-api-access-wf4h6\") pod \"ceilometer-0\" (UID: \"e56111b4-4b8a-4e07-bba9-68d34fd179fb\") " pod="openstack/ceilometer-0" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.755265 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e56111b4-4b8a-4e07-bba9-68d34fd179fb-log-httpd\") pod \"ceilometer-0\" (UID: \"e56111b4-4b8a-4e07-bba9-68d34fd179fb\") " pod="openstack/ceilometer-0" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.759329 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3d747bbd-45e4-4c6c-a6cb-64a259440231-operator-scripts\") pod \"nova-cell0-a9af-account-create-update-l42hs\" (UID: \"3d747bbd-45e4-4c6c-a6cb-64a259440231\") " pod="openstack/nova-cell0-a9af-account-create-update-l42hs" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.761633 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e56111b4-4b8a-4e07-bba9-68d34fd179fb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e56111b4-4b8a-4e07-bba9-68d34fd179fb\") " pod="openstack/ceilometer-0" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.761639 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e56111b4-4b8a-4e07-bba9-68d34fd179fb-config-data\") pod \"ceilometer-0\" (UID: \"e56111b4-4b8a-4e07-bba9-68d34fd179fb\") " pod="openstack/ceilometer-0" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.764005 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e56111b4-4b8a-4e07-bba9-68d34fd179fb-scripts\") pod \"ceilometer-0\" (UID: \"e56111b4-4b8a-4e07-bba9-68d34fd179fb\") " pod="openstack/ceilometer-0" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.764185 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e56111b4-4b8a-4e07-bba9-68d34fd179fb-run-httpd\") pod \"ceilometer-0\" (UID: \"e56111b4-4b8a-4e07-bba9-68d34fd179fb\") " pod="openstack/ceilometer-0" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.767844 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-rc66h" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.768628 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e56111b4-4b8a-4e07-bba9-68d34fd179fb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e56111b4-4b8a-4e07-bba9-68d34fd179fb\") " pod="openstack/ceilometer-0" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.778297 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rnf9\" (UniqueName: \"kubernetes.io/projected/3d747bbd-45e4-4c6c-a6cb-64a259440231-kube-api-access-5rnf9\") pod \"nova-cell0-a9af-account-create-update-l42hs\" (UID: \"3d747bbd-45e4-4c6c-a6cb-64a259440231\") " pod="openstack/nova-cell0-a9af-account-create-update-l42hs" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.781766 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf4h6\" (UniqueName: \"kubernetes.io/projected/e56111b4-4b8a-4e07-bba9-68d34fd179fb-kube-api-access-wf4h6\") pod \"ceilometer-0\" (UID: \"e56111b4-4b8a-4e07-bba9-68d34fd179fb\") " pod="openstack/ceilometer-0" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.855925 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d0c9416-2de8-4d45-8818-87ae59a2cd6f-operator-scripts\") pod \"nova-cell1-aa00-account-create-update-r9kcs\" (UID: \"0d0c9416-2de8-4d45-8818-87ae59a2cd6f\") " pod="openstack/nova-cell1-aa00-account-create-update-r9kcs" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.855966 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcpqr\" (UniqueName: \"kubernetes.io/projected/0d0c9416-2de8-4d45-8818-87ae59a2cd6f-kube-api-access-bcpqr\") pod \"nova-cell1-aa00-account-create-update-r9kcs\" (UID: \"0d0c9416-2de8-4d45-8818-87ae59a2cd6f\") " pod="openstack/nova-cell1-aa00-account-create-update-r9kcs" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.862520 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.863457 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d0c9416-2de8-4d45-8818-87ae59a2cd6f-operator-scripts\") pod \"nova-cell1-aa00-account-create-update-r9kcs\" (UID: \"0d0c9416-2de8-4d45-8818-87ae59a2cd6f\") " pod="openstack/nova-cell1-aa00-account-create-update-r9kcs" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.878080 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-a9af-account-create-update-l42hs" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.886629 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcpqr\" (UniqueName: \"kubernetes.io/projected/0d0c9416-2de8-4d45-8818-87ae59a2cd6f-kube-api-access-bcpqr\") pod \"nova-cell1-aa00-account-create-update-r9kcs\" (UID: \"0d0c9416-2de8-4d45-8818-87ae59a2cd6f\") " pod="openstack/nova-cell1-aa00-account-create-update-r9kcs" Feb 17 13:23:47 crc kubenswrapper[4955]: I0217 13:23:47.898118 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-aa00-account-create-update-r9kcs" Feb 17 13:23:48 crc kubenswrapper[4955]: W0217 13:23:48.042516 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff0b8526_a0fe_425e_92e8_a4933f3faf61.slice/crio-a01e54d519f4f1cfde55a3c130ff7e3d25f332c27f75bb1c307ce797129d3802 WatchSource:0}: Error finding container a01e54d519f4f1cfde55a3c130ff7e3d25f332c27f75bb1c307ce797129d3802: Status 404 returned error can't find the container with id a01e54d519f4f1cfde55a3c130ff7e3d25f332c27f75bb1c307ce797129d3802 Feb 17 13:23:48 crc kubenswrapper[4955]: I0217 13:23:48.043548 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-9ms22"] Feb 17 13:23:48 crc kubenswrapper[4955]: I0217 13:23:48.236437 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d20f3280-ab79-434f-a89a-11d6fbfc0edf" path="/var/lib/kubelet/pods/d20f3280-ab79-434f-a89a-11d6fbfc0edf/volumes" Feb 17 13:23:48 crc kubenswrapper[4955]: I0217 13:23:48.284293 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-301f-account-create-update-wc7xh"] Feb 17 13:23:48 crc kubenswrapper[4955]: I0217 13:23:48.402206 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-6qshk"] Feb 17 13:23:48 crc kubenswrapper[4955]: I0217 13:23:48.417056 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-9ms22" event={"ID":"ff0b8526-a0fe-425e-92e8-a4933f3faf61","Type":"ContainerStarted","Data":"6fe3e468aec84cc15d844b1d42ba7b177b31ca0841242ceea64587d233ef0056"} Feb 17 13:23:48 crc kubenswrapper[4955]: I0217 13:23:48.417099 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-9ms22" event={"ID":"ff0b8526-a0fe-425e-92e8-a4933f3faf61","Type":"ContainerStarted","Data":"a01e54d519f4f1cfde55a3c130ff7e3d25f332c27f75bb1c307ce797129d3802"} Feb 17 13:23:48 crc kubenswrapper[4955]: I0217 13:23:48.422302 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-301f-account-create-update-wc7xh" event={"ID":"84b8be7f-2891-44a0-beb3-850911ac8b2a","Type":"ContainerStarted","Data":"0f8a34b069dc094f91778ff6fe018180755afc9e25b32d460b9683220dc90124"} Feb 17 13:23:48 crc kubenswrapper[4955]: I0217 13:23:48.424247 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5bcc76b6cf-4r2g9" event={"ID":"0fa0efd9-940a-4580-ab68-c5954c0e732e","Type":"ContainerStarted","Data":"c615fef6c596974f6dc66af43790c25f22dccf6180e9f1f5dd3c129c2231e0c6"} Feb 17 13:23:48 crc kubenswrapper[4955]: I0217 13:23:48.424282 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5bcc76b6cf-4r2g9" event={"ID":"0fa0efd9-940a-4580-ab68-c5954c0e732e","Type":"ContainerStarted","Data":"d8f4c0a1d04f9cb0777829fe5b4d4718d4c860b67571ee54c5807caf44fe1fdd"} Feb 17 13:23:48 crc kubenswrapper[4955]: I0217 13:23:48.424395 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5bcc76b6cf-4r2g9" Feb 17 13:23:48 crc kubenswrapper[4955]: I0217 13:23:48.424489 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5bcc76b6cf-4r2g9" Feb 17 13:23:48 crc kubenswrapper[4955]: I0217 13:23:48.464974 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-9ms22" podStartSLOduration=1.464949486 podStartE2EDuration="1.464949486s" podCreationTimestamp="2026-02-17 13:23:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:23:48.433092057 +0000 UTC m=+1166.955821600" watchObservedRunningTime="2026-02-17 13:23:48.464949486 +0000 UTC m=+1166.987679029" Feb 17 13:23:48 crc kubenswrapper[4955]: I0217 13:23:48.465401 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-5bcc76b6cf-4r2g9" podStartSLOduration=7.465395618 podStartE2EDuration="7.465395618s" podCreationTimestamp="2026-02-17 13:23:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:23:48.458397799 +0000 UTC m=+1166.981127342" watchObservedRunningTime="2026-02-17 13:23:48.465395618 +0000 UTC m=+1166.988125171" Feb 17 13:23:48 crc kubenswrapper[4955]: I0217 13:23:48.510872 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:23:48 crc kubenswrapper[4955]: W0217 13:23:48.513318 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode56111b4_4b8a_4e07_bba9_68d34fd179fb.slice/crio-6b71c055acde1fa41a4d33dd3bfa35d899ed637a32058d82efe419386c8dd1e2 WatchSource:0}: Error finding container 6b71c055acde1fa41a4d33dd3bfa35d899ed637a32058d82efe419386c8dd1e2: Status 404 returned error can't find the container with id 6b71c055acde1fa41a4d33dd3bfa35d899ed637a32058d82efe419386c8dd1e2 Feb 17 13:23:48 crc kubenswrapper[4955]: I0217 13:23:48.523500 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-rc66h"] Feb 17 13:23:48 crc kubenswrapper[4955]: I0217 13:23:48.616094 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-aa00-account-create-update-r9kcs"] Feb 17 13:23:48 crc kubenswrapper[4955]: I0217 13:23:48.643370 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-a9af-account-create-update-l42hs"] Feb 17 13:23:48 crc kubenswrapper[4955]: I0217 13:23:48.957894 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 17 13:23:48 crc kubenswrapper[4955]: I0217 13:23:48.958424 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="03690a1e-7d74-482e-be94-78c5074595b6" containerName="glance-log" containerID="cri-o://76e742fa32650a687849fb64674d1d2f47974c7b99026b751786a56009f0ba57" gracePeriod=30 Feb 17 13:23:48 crc kubenswrapper[4955]: I0217 13:23:48.958926 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="03690a1e-7d74-482e-be94-78c5074595b6" containerName="glance-httpd" containerID="cri-o://ddb90ae2171d4c56825cc492fcade33f9f4837d699bc2112954ddeddd9d99958" gracePeriod=30 Feb 17 13:23:49 crc kubenswrapper[4955]: I0217 13:23:49.440196 4955 generic.go:334] "Generic (PLEG): container finished" podID="1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf" containerID="32e18fff4a97bffae1a71c95b67ec8d234531f3a2c24d149ec79a131c6c4a75a" exitCode=137 Feb 17 13:23:49 crc kubenswrapper[4955]: I0217 13:23:49.440532 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5788cc4448-245bs" event={"ID":"1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf","Type":"ContainerDied","Data":"32e18fff4a97bffae1a71c95b67ec8d234531f3a2c24d149ec79a131c6c4a75a"} Feb 17 13:23:49 crc kubenswrapper[4955]: I0217 13:23:49.448401 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-a9af-account-create-update-l42hs" event={"ID":"3d747bbd-45e4-4c6c-a6cb-64a259440231","Type":"ContainerStarted","Data":"ab7acc4aa73173917c4542bceac2bd95c370beb085d9ca33215c40e9a5c13f31"} Feb 17 13:23:49 crc kubenswrapper[4955]: I0217 13:23:49.448444 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-a9af-account-create-update-l42hs" event={"ID":"3d747bbd-45e4-4c6c-a6cb-64a259440231","Type":"ContainerStarted","Data":"4ebd828e2628219a5c55fddc42bc8e9907d016aa9018b84c40413616930fcef7"} Feb 17 13:23:49 crc kubenswrapper[4955]: I0217 13:23:49.461537 4955 generic.go:334] "Generic (PLEG): container finished" podID="03690a1e-7d74-482e-be94-78c5074595b6" containerID="76e742fa32650a687849fb64674d1d2f47974c7b99026b751786a56009f0ba57" exitCode=143 Feb 17 13:23:49 crc kubenswrapper[4955]: I0217 13:23:49.461617 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"03690a1e-7d74-482e-be94-78c5074595b6","Type":"ContainerDied","Data":"76e742fa32650a687849fb64674d1d2f47974c7b99026b751786a56009f0ba57"} Feb 17 13:23:49 crc kubenswrapper[4955]: I0217 13:23:49.472238 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-a9af-account-create-update-l42hs" podStartSLOduration=2.472218527 podStartE2EDuration="2.472218527s" podCreationTimestamp="2026-02-17 13:23:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:23:49.465745702 +0000 UTC m=+1167.988475245" watchObservedRunningTime="2026-02-17 13:23:49.472218527 +0000 UTC m=+1167.994948070" Feb 17 13:23:49 crc kubenswrapper[4955]: I0217 13:23:49.482066 4955 generic.go:334] "Generic (PLEG): container finished" podID="4031cf36-67c1-425d-98f9-0c4f3a4a6065" containerID="1d4a8311652addf9faf8c3a9147990dbfb37dd838fdc0d55c76b3ff9aa5ebaef" exitCode=0 Feb 17 13:23:49 crc kubenswrapper[4955]: I0217 13:23:49.482145 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-rc66h" event={"ID":"4031cf36-67c1-425d-98f9-0c4f3a4a6065","Type":"ContainerDied","Data":"1d4a8311652addf9faf8c3a9147990dbfb37dd838fdc0d55c76b3ff9aa5ebaef"} Feb 17 13:23:49 crc kubenswrapper[4955]: I0217 13:23:49.482171 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-rc66h" event={"ID":"4031cf36-67c1-425d-98f9-0c4f3a4a6065","Type":"ContainerStarted","Data":"4fa55ef9674899c026a319d48e19e7ee41d1d349483cc73edc4a4995fbf5b096"} Feb 17 13:23:49 crc kubenswrapper[4955]: I0217 13:23:49.501924 4955 generic.go:334] "Generic (PLEG): container finished" podID="ff0b8526-a0fe-425e-92e8-a4933f3faf61" containerID="6fe3e468aec84cc15d844b1d42ba7b177b31ca0841242ceea64587d233ef0056" exitCode=0 Feb 17 13:23:49 crc kubenswrapper[4955]: I0217 13:23:49.502026 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-9ms22" event={"ID":"ff0b8526-a0fe-425e-92e8-a4933f3faf61","Type":"ContainerDied","Data":"6fe3e468aec84cc15d844b1d42ba7b177b31ca0841242ceea64587d233ef0056"} Feb 17 13:23:49 crc kubenswrapper[4955]: I0217 13:23:49.518990 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-aa00-account-create-update-r9kcs" event={"ID":"0d0c9416-2de8-4d45-8818-87ae59a2cd6f","Type":"ContainerStarted","Data":"bdd7c1b366095bdefcf87a6ccecb349fc25188a4c12fc60e1cbb38b9b6833d28"} Feb 17 13:23:49 crc kubenswrapper[4955]: I0217 13:23:49.519034 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-aa00-account-create-update-r9kcs" event={"ID":"0d0c9416-2de8-4d45-8818-87ae59a2cd6f","Type":"ContainerStarted","Data":"c683ac3fda8effab3aebcd4f470d4d721975cf8d55fd210079223f0a15b8d00e"} Feb 17 13:23:49 crc kubenswrapper[4955]: I0217 13:23:49.522814 4955 generic.go:334] "Generic (PLEG): container finished" podID="c4bbd808-631c-4a9b-81e6-dee199484728" containerID="6722a7996f5d1b7e07e4582a2b2872753d0693e485f080ab97224b9c557c3cc6" exitCode=0 Feb 17 13:23:49 crc kubenswrapper[4955]: I0217 13:23:49.522902 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-6qshk" event={"ID":"c4bbd808-631c-4a9b-81e6-dee199484728","Type":"ContainerDied","Data":"6722a7996f5d1b7e07e4582a2b2872753d0693e485f080ab97224b9c557c3cc6"} Feb 17 13:23:49 crc kubenswrapper[4955]: I0217 13:23:49.522930 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-6qshk" event={"ID":"c4bbd808-631c-4a9b-81e6-dee199484728","Type":"ContainerStarted","Data":"d86d58af659f47013372ddf40eedd0721e15397b196767b70315096c9dc327fb"} Feb 17 13:23:49 crc kubenswrapper[4955]: I0217 13:23:49.524744 4955 generic.go:334] "Generic (PLEG): container finished" podID="84b8be7f-2891-44a0-beb3-850911ac8b2a" containerID="c527171461761a90d74648e6aa7ee6ea4bb27ea42c3a36ca0863784547769d13" exitCode=0 Feb 17 13:23:49 crc kubenswrapper[4955]: I0217 13:23:49.524803 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-301f-account-create-update-wc7xh" event={"ID":"84b8be7f-2891-44a0-beb3-850911ac8b2a","Type":"ContainerDied","Data":"c527171461761a90d74648e6aa7ee6ea4bb27ea42c3a36ca0863784547769d13"} Feb 17 13:23:49 crc kubenswrapper[4955]: I0217 13:23:49.538613 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e56111b4-4b8a-4e07-bba9-68d34fd179fb","Type":"ContainerStarted","Data":"f520260e9c3bbb1b63ec8f9522bd9fcc6cd69d7e6ab2a4af3b330077c6fc553f"} Feb 17 13:23:49 crc kubenswrapper[4955]: I0217 13:23:49.538655 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e56111b4-4b8a-4e07-bba9-68d34fd179fb","Type":"ContainerStarted","Data":"6b71c055acde1fa41a4d33dd3bfa35d899ed637a32058d82efe419386c8dd1e2"} Feb 17 13:23:49 crc kubenswrapper[4955]: I0217 13:23:49.558693 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-aa00-account-create-update-r9kcs" podStartSLOduration=2.558673723 podStartE2EDuration="2.558673723s" podCreationTimestamp="2026-02-17 13:23:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:23:49.552747944 +0000 UTC m=+1168.075477507" watchObservedRunningTime="2026-02-17 13:23:49.558673723 +0000 UTC m=+1168.081403256" Feb 17 13:23:50 crc kubenswrapper[4955]: I0217 13:23:50.414368 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7c5ddc6895-552ts" Feb 17 13:23:50 crc kubenswrapper[4955]: I0217 13:23:50.481914 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-746fcf4988-4q7tw"] Feb 17 13:23:50 crc kubenswrapper[4955]: I0217 13:23:50.482401 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-746fcf4988-4q7tw" podUID="47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e" containerName="neutron-api" containerID="cri-o://6d565c62d191904d43404bc26ded506c5b6c75a7f0d89b1ca2702cf83d5ad913" gracePeriod=30 Feb 17 13:23:50 crc kubenswrapper[4955]: I0217 13:23:50.482694 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-746fcf4988-4q7tw" podUID="47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e" containerName="neutron-httpd" containerID="cri-o://6df53ab637202b08040479088ed897dde6492ba6c8f79fb9bae89a3e98c0cd26" gracePeriod=30 Feb 17 13:23:50 crc kubenswrapper[4955]: I0217 13:23:50.621702 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 17 13:23:50 crc kubenswrapper[4955]: I0217 13:23:50.622493 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5da72340-6ea2-42dd-855b-dec4f29f061c" containerName="glance-log" containerID="cri-o://29b22587a44a1d3f014b0d32e0afc120c7ed49e76d1b13702dd03d797cd374cc" gracePeriod=30 Feb 17 13:23:50 crc kubenswrapper[4955]: I0217 13:23:50.623147 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5da72340-6ea2-42dd-855b-dec4f29f061c" containerName="glance-httpd" containerID="cri-o://172f55b5031dad85034b5db9a36dfbaf6236f4dcc6a4d44cdd9d2c42d86253b2" gracePeriod=30 Feb 17 13:23:50 crc kubenswrapper[4955]: I0217 13:23:50.864712 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5788cc4448-245bs" Feb 17 13:23:50 crc kubenswrapper[4955]: I0217 13:23:50.941906 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf-horizon-secret-key\") pod \"1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf\" (UID: \"1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf\") " Feb 17 13:23:50 crc kubenswrapper[4955]: I0217 13:23:50.942246 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf-scripts\") pod \"1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf\" (UID: \"1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf\") " Feb 17 13:23:50 crc kubenswrapper[4955]: I0217 13:23:50.942409 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf-logs\") pod \"1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf\" (UID: \"1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf\") " Feb 17 13:23:50 crc kubenswrapper[4955]: I0217 13:23:50.942531 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbs6\" (UniqueName: \"kubernetes.io/projected/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf-kube-api-access-jhbs6\") pod \"1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf\" (UID: \"1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf\") " Feb 17 13:23:50 crc kubenswrapper[4955]: I0217 13:23:50.942645 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf-combined-ca-bundle\") pod \"1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf\" (UID: \"1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf\") " Feb 17 13:23:50 crc kubenswrapper[4955]: I0217 13:23:50.942886 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf-horizon-tls-certs\") pod \"1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf\" (UID: \"1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf\") " Feb 17 13:23:50 crc kubenswrapper[4955]: I0217 13:23:50.943050 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf-config-data\") pod \"1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf\" (UID: \"1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf\") " Feb 17 13:23:50 crc kubenswrapper[4955]: I0217 13:23:50.943049 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf-logs" (OuterVolumeSpecName: "logs") pod "1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf" (UID: "1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:23:50 crc kubenswrapper[4955]: I0217 13:23:50.952188 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf" (UID: "1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:50 crc kubenswrapper[4955]: I0217 13:23:50.952673 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf-kube-api-access-jhbs6" (OuterVolumeSpecName: "kube-api-access-jhbs6") pod "1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf" (UID: "1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf"). InnerVolumeSpecName "kube-api-access-jhbs6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:23:50 crc kubenswrapper[4955]: I0217 13:23:50.987063 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf-scripts" (OuterVolumeSpecName: "scripts") pod "1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf" (UID: "1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:23:50 crc kubenswrapper[4955]: I0217 13:23:50.990348 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf" (UID: "1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:50 crc kubenswrapper[4955]: I0217 13:23:50.994675 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf-config-data" (OuterVolumeSpecName: "config-data") pod "1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf" (UID: "1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.023771 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf" (UID: "1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.045158 4955 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.045189 4955 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf-logs\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.045211 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbs6\" (UniqueName: \"kubernetes.io/projected/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf-kube-api-access-jhbs6\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.045224 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.045233 4955 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.045241 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.045249 4955 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.160331 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-9ms22" Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.248290 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fkd2\" (UniqueName: \"kubernetes.io/projected/ff0b8526-a0fe-425e-92e8-a4933f3faf61-kube-api-access-4fkd2\") pod \"ff0b8526-a0fe-425e-92e8-a4933f3faf61\" (UID: \"ff0b8526-a0fe-425e-92e8-a4933f3faf61\") " Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.248573 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff0b8526-a0fe-425e-92e8-a4933f3faf61-operator-scripts\") pod \"ff0b8526-a0fe-425e-92e8-a4933f3faf61\" (UID: \"ff0b8526-a0fe-425e-92e8-a4933f3faf61\") " Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.256357 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff0b8526-a0fe-425e-92e8-a4933f3faf61-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ff0b8526-a0fe-425e-92e8-a4933f3faf61" (UID: "ff0b8526-a0fe-425e-92e8-a4933f3faf61"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.256373 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff0b8526-a0fe-425e-92e8-a4933f3faf61-kube-api-access-4fkd2" (OuterVolumeSpecName: "kube-api-access-4fkd2") pod "ff0b8526-a0fe-425e-92e8-a4933f3faf61" (UID: "ff0b8526-a0fe-425e-92e8-a4933f3faf61"). InnerVolumeSpecName "kube-api-access-4fkd2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.328302 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-6qshk" Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.351408 4955 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff0b8526-a0fe-425e-92e8-a4933f3faf61-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.351441 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fkd2\" (UniqueName: \"kubernetes.io/projected/ff0b8526-a0fe-425e-92e8-a4933f3faf61-kube-api-access-4fkd2\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.357090 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-301f-account-create-update-wc7xh" Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.405393 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-rc66h" Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.452741 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6zsp5\" (UniqueName: \"kubernetes.io/projected/4031cf36-67c1-425d-98f9-0c4f3a4a6065-kube-api-access-6zsp5\") pod \"4031cf36-67c1-425d-98f9-0c4f3a4a6065\" (UID: \"4031cf36-67c1-425d-98f9-0c4f3a4a6065\") " Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.452826 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4031cf36-67c1-425d-98f9-0c4f3a4a6065-operator-scripts\") pod \"4031cf36-67c1-425d-98f9-0c4f3a4a6065\" (UID: \"4031cf36-67c1-425d-98f9-0c4f3a4a6065\") " Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.453533 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8bns5\" (UniqueName: \"kubernetes.io/projected/84b8be7f-2891-44a0-beb3-850911ac8b2a-kube-api-access-8bns5\") pod \"84b8be7f-2891-44a0-beb3-850911ac8b2a\" (UID: \"84b8be7f-2891-44a0-beb3-850911ac8b2a\") " Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.453963 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84b8be7f-2891-44a0-beb3-850911ac8b2a-operator-scripts\") pod \"84b8be7f-2891-44a0-beb3-850911ac8b2a\" (UID: \"84b8be7f-2891-44a0-beb3-850911ac8b2a\") " Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.454015 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mcnr8\" (UniqueName: \"kubernetes.io/projected/c4bbd808-631c-4a9b-81e6-dee199484728-kube-api-access-mcnr8\") pod \"c4bbd808-631c-4a9b-81e6-dee199484728\" (UID: \"c4bbd808-631c-4a9b-81e6-dee199484728\") " Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.454230 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c4bbd808-631c-4a9b-81e6-dee199484728-operator-scripts\") pod \"c4bbd808-631c-4a9b-81e6-dee199484728\" (UID: \"c4bbd808-631c-4a9b-81e6-dee199484728\") " Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.454454 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4031cf36-67c1-425d-98f9-0c4f3a4a6065-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4031cf36-67c1-425d-98f9-0c4f3a4a6065" (UID: "4031cf36-67c1-425d-98f9-0c4f3a4a6065"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.454879 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4bbd808-631c-4a9b-81e6-dee199484728-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c4bbd808-631c-4a9b-81e6-dee199484728" (UID: "c4bbd808-631c-4a9b-81e6-dee199484728"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.454927 4955 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4031cf36-67c1-425d-98f9-0c4f3a4a6065-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.455228 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84b8be7f-2891-44a0-beb3-850911ac8b2a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "84b8be7f-2891-44a0-beb3-850911ac8b2a" (UID: "84b8be7f-2891-44a0-beb3-850911ac8b2a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.460707 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84b8be7f-2891-44a0-beb3-850911ac8b2a-kube-api-access-8bns5" (OuterVolumeSpecName: "kube-api-access-8bns5") pod "84b8be7f-2891-44a0-beb3-850911ac8b2a" (UID: "84b8be7f-2891-44a0-beb3-850911ac8b2a"). InnerVolumeSpecName "kube-api-access-8bns5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.465906 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4031cf36-67c1-425d-98f9-0c4f3a4a6065-kube-api-access-6zsp5" (OuterVolumeSpecName: "kube-api-access-6zsp5") pod "4031cf36-67c1-425d-98f9-0c4f3a4a6065" (UID: "4031cf36-67c1-425d-98f9-0c4f3a4a6065"). InnerVolumeSpecName "kube-api-access-6zsp5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.472767 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4bbd808-631c-4a9b-81e6-dee199484728-kube-api-access-mcnr8" (OuterVolumeSpecName: "kube-api-access-mcnr8") pod "c4bbd808-631c-4a9b-81e6-dee199484728" (UID: "c4bbd808-631c-4a9b-81e6-dee199484728"). InnerVolumeSpecName "kube-api-access-mcnr8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.556460 4955 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c4bbd808-631c-4a9b-81e6-dee199484728-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.556491 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6zsp5\" (UniqueName: \"kubernetes.io/projected/4031cf36-67c1-425d-98f9-0c4f3a4a6065-kube-api-access-6zsp5\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.556506 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8bns5\" (UniqueName: \"kubernetes.io/projected/84b8be7f-2891-44a0-beb3-850911ac8b2a-kube-api-access-8bns5\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.556514 4955 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84b8be7f-2891-44a0-beb3-850911ac8b2a-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.556523 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mcnr8\" (UniqueName: \"kubernetes.io/projected/c4bbd808-631c-4a9b-81e6-dee199484728-kube-api-access-mcnr8\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.572718 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-301f-account-create-update-wc7xh" event={"ID":"84b8be7f-2891-44a0-beb3-850911ac8b2a","Type":"ContainerDied","Data":"0f8a34b069dc094f91778ff6fe018180755afc9e25b32d460b9683220dc90124"} Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.572757 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f8a34b069dc094f91778ff6fe018180755afc9e25b32d460b9683220dc90124" Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.572833 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-301f-account-create-update-wc7xh" Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.579973 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-rc66h" event={"ID":"4031cf36-67c1-425d-98f9-0c4f3a4a6065","Type":"ContainerDied","Data":"4fa55ef9674899c026a319d48e19e7ee41d1d349483cc73edc4a4995fbf5b096"} Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.580015 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4fa55ef9674899c026a319d48e19e7ee41d1d349483cc73edc4a4995fbf5b096" Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.580096 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-rc66h" Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.595167 4955 generic.go:334] "Generic (PLEG): container finished" podID="5da72340-6ea2-42dd-855b-dec4f29f061c" containerID="29b22587a44a1d3f014b0d32e0afc120c7ed49e76d1b13702dd03d797cd374cc" exitCode=143 Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.595301 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5da72340-6ea2-42dd-855b-dec4f29f061c","Type":"ContainerDied","Data":"29b22587a44a1d3f014b0d32e0afc120c7ed49e76d1b13702dd03d797cd374cc"} Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.609597 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5788cc4448-245bs" event={"ID":"1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf","Type":"ContainerDied","Data":"937a259e06c16c941790cb4506e60ff90561c28dea54c31c6838afbcee91cd93"} Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.609700 4955 scope.go:117] "RemoveContainer" containerID="52cdaba09384ae938ed1a6b17e232b9095c1468ccdd7386221ad6a01d6be8ff4" Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.610005 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5788cc4448-245bs" Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.627957 4955 generic.go:334] "Generic (PLEG): container finished" podID="47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e" containerID="6df53ab637202b08040479088ed897dde6492ba6c8f79fb9bae89a3e98c0cd26" exitCode=0 Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.628051 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-746fcf4988-4q7tw" event={"ID":"47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e","Type":"ContainerDied","Data":"6df53ab637202b08040479088ed897dde6492ba6c8f79fb9bae89a3e98c0cd26"} Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.629709 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-9ms22" event={"ID":"ff0b8526-a0fe-425e-92e8-a4933f3faf61","Type":"ContainerDied","Data":"a01e54d519f4f1cfde55a3c130ff7e3d25f332c27f75bb1c307ce797129d3802"} Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.629750 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a01e54d519f4f1cfde55a3c130ff7e3d25f332c27f75bb1c307ce797129d3802" Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.629907 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-9ms22" Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.636846 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e56111b4-4b8a-4e07-bba9-68d34fd179fb","Type":"ContainerStarted","Data":"89e46bb8b72cfb0f7210ef620927b81062fa54f4cd2c210711b1668eae6ffcac"} Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.638397 4955 generic.go:334] "Generic (PLEG): container finished" podID="0d0c9416-2de8-4d45-8818-87ae59a2cd6f" containerID="bdd7c1b366095bdefcf87a6ccecb349fc25188a4c12fc60e1cbb38b9b6833d28" exitCode=0 Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.638488 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-aa00-account-create-update-r9kcs" event={"ID":"0d0c9416-2de8-4d45-8818-87ae59a2cd6f","Type":"ContainerDied","Data":"bdd7c1b366095bdefcf87a6ccecb349fc25188a4c12fc60e1cbb38b9b6833d28"} Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.650456 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-6qshk" event={"ID":"c4bbd808-631c-4a9b-81e6-dee199484728","Type":"ContainerDied","Data":"d86d58af659f47013372ddf40eedd0721e15397b196767b70315096c9dc327fb"} Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.650514 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d86d58af659f47013372ddf40eedd0721e15397b196767b70315096c9dc327fb" Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.650603 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-6qshk" Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.675208 4955 generic.go:334] "Generic (PLEG): container finished" podID="3d747bbd-45e4-4c6c-a6cb-64a259440231" containerID="ab7acc4aa73173917c4542bceac2bd95c370beb085d9ca33215c40e9a5c13f31" exitCode=0 Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.675254 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-a9af-account-create-update-l42hs" event={"ID":"3d747bbd-45e4-4c6c-a6cb-64a259440231","Type":"ContainerDied","Data":"ab7acc4aa73173917c4542bceac2bd95c370beb085d9ca33215c40e9a5c13f31"} Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.813460 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5788cc4448-245bs"] Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.822974 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5788cc4448-245bs"] Feb 17 13:23:51 crc kubenswrapper[4955]: I0217 13:23:51.913962 4955 scope.go:117] "RemoveContainer" containerID="32e18fff4a97bffae1a71c95b67ec8d234531f3a2c24d149ec79a131c6c4a75a" Feb 17 13:23:52 crc kubenswrapper[4955]: I0217 13:23:52.234341 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf" path="/var/lib/kubelet/pods/1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf/volumes" Feb 17 13:23:52 crc kubenswrapper[4955]: I0217 13:23:52.686437 4955 generic.go:334] "Generic (PLEG): container finished" podID="03690a1e-7d74-482e-be94-78c5074595b6" containerID="ddb90ae2171d4c56825cc492fcade33f9f4837d699bc2112954ddeddd9d99958" exitCode=0 Feb 17 13:23:52 crc kubenswrapper[4955]: I0217 13:23:52.686624 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"03690a1e-7d74-482e-be94-78c5074595b6","Type":"ContainerDied","Data":"ddb90ae2171d4c56825cc492fcade33f9f4837d699bc2112954ddeddd9d99958"} Feb 17 13:23:52 crc kubenswrapper[4955]: I0217 13:23:52.686707 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"03690a1e-7d74-482e-be94-78c5074595b6","Type":"ContainerDied","Data":"0eb8f2a6d19208279ef41122b8a09aee30ec45e12c795c772236ca4839d4e9cf"} Feb 17 13:23:52 crc kubenswrapper[4955]: I0217 13:23:52.686719 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0eb8f2a6d19208279ef41122b8a09aee30ec45e12c795c772236ca4839d4e9cf" Feb 17 13:23:52 crc kubenswrapper[4955]: I0217 13:23:52.688708 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e56111b4-4b8a-4e07-bba9-68d34fd179fb","Type":"ContainerStarted","Data":"4683a112963c8aed791f163ff609e4a0b3f319e5243c44eac71422cec74ea161"} Feb 17 13:23:52 crc kubenswrapper[4955]: I0217 13:23:52.742955 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 17 13:23:52 crc kubenswrapper[4955]: I0217 13:23:52.796066 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03690a1e-7d74-482e-be94-78c5074595b6-scripts\") pod \"03690a1e-7d74-482e-be94-78c5074595b6\" (UID: \"03690a1e-7d74-482e-be94-78c5074595b6\") " Feb 17 13:23:52 crc kubenswrapper[4955]: I0217 13:23:52.796166 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/03690a1e-7d74-482e-be94-78c5074595b6-httpd-run\") pod \"03690a1e-7d74-482e-be94-78c5074595b6\" (UID: \"03690a1e-7d74-482e-be94-78c5074595b6\") " Feb 17 13:23:52 crc kubenswrapper[4955]: I0217 13:23:52.796191 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"03690a1e-7d74-482e-be94-78c5074595b6\" (UID: \"03690a1e-7d74-482e-be94-78c5074595b6\") " Feb 17 13:23:52 crc kubenswrapper[4955]: I0217 13:23:52.796275 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/03690a1e-7d74-482e-be94-78c5074595b6-logs\") pod \"03690a1e-7d74-482e-be94-78c5074595b6\" (UID: \"03690a1e-7d74-482e-be94-78c5074595b6\") " Feb 17 13:23:52 crc kubenswrapper[4955]: I0217 13:23:52.796314 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/03690a1e-7d74-482e-be94-78c5074595b6-public-tls-certs\") pod \"03690a1e-7d74-482e-be94-78c5074595b6\" (UID: \"03690a1e-7d74-482e-be94-78c5074595b6\") " Feb 17 13:23:52 crc kubenswrapper[4955]: I0217 13:23:52.796401 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ssjqh\" (UniqueName: \"kubernetes.io/projected/03690a1e-7d74-482e-be94-78c5074595b6-kube-api-access-ssjqh\") pod \"03690a1e-7d74-482e-be94-78c5074595b6\" (UID: \"03690a1e-7d74-482e-be94-78c5074595b6\") " Feb 17 13:23:52 crc kubenswrapper[4955]: I0217 13:23:52.796453 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03690a1e-7d74-482e-be94-78c5074595b6-combined-ca-bundle\") pod \"03690a1e-7d74-482e-be94-78c5074595b6\" (UID: \"03690a1e-7d74-482e-be94-78c5074595b6\") " Feb 17 13:23:52 crc kubenswrapper[4955]: I0217 13:23:52.796545 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03690a1e-7d74-482e-be94-78c5074595b6-config-data\") pod \"03690a1e-7d74-482e-be94-78c5074595b6\" (UID: \"03690a1e-7d74-482e-be94-78c5074595b6\") " Feb 17 13:23:52 crc kubenswrapper[4955]: I0217 13:23:52.798107 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03690a1e-7d74-482e-be94-78c5074595b6-logs" (OuterVolumeSpecName: "logs") pod "03690a1e-7d74-482e-be94-78c5074595b6" (UID: "03690a1e-7d74-482e-be94-78c5074595b6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:23:52 crc kubenswrapper[4955]: I0217 13:23:52.800209 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03690a1e-7d74-482e-be94-78c5074595b6-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "03690a1e-7d74-482e-be94-78c5074595b6" (UID: "03690a1e-7d74-482e-be94-78c5074595b6"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:23:52 crc kubenswrapper[4955]: I0217 13:23:52.831415 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03690a1e-7d74-482e-be94-78c5074595b6-scripts" (OuterVolumeSpecName: "scripts") pod "03690a1e-7d74-482e-be94-78c5074595b6" (UID: "03690a1e-7d74-482e-be94-78c5074595b6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:52 crc kubenswrapper[4955]: I0217 13:23:52.831689 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03690a1e-7d74-482e-be94-78c5074595b6-kube-api-access-ssjqh" (OuterVolumeSpecName: "kube-api-access-ssjqh") pod "03690a1e-7d74-482e-be94-78c5074595b6" (UID: "03690a1e-7d74-482e-be94-78c5074595b6"). InnerVolumeSpecName "kube-api-access-ssjqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:23:52 crc kubenswrapper[4955]: I0217 13:23:52.832173 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "03690a1e-7d74-482e-be94-78c5074595b6" (UID: "03690a1e-7d74-482e-be94-78c5074595b6"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 17 13:23:52 crc kubenswrapper[4955]: I0217 13:23:52.878944 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03690a1e-7d74-482e-be94-78c5074595b6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "03690a1e-7d74-482e-be94-78c5074595b6" (UID: "03690a1e-7d74-482e-be94-78c5074595b6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:52 crc kubenswrapper[4955]: I0217 13:23:52.899333 4955 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03690a1e-7d74-482e-be94-78c5074595b6-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:52 crc kubenswrapper[4955]: I0217 13:23:52.899357 4955 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/03690a1e-7d74-482e-be94-78c5074595b6-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:52 crc kubenswrapper[4955]: I0217 13:23:52.899377 4955 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Feb 17 13:23:52 crc kubenswrapper[4955]: I0217 13:23:52.899388 4955 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/03690a1e-7d74-482e-be94-78c5074595b6-logs\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:52 crc kubenswrapper[4955]: I0217 13:23:52.899397 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ssjqh\" (UniqueName: \"kubernetes.io/projected/03690a1e-7d74-482e-be94-78c5074595b6-kube-api-access-ssjqh\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:52 crc kubenswrapper[4955]: I0217 13:23:52.899593 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03690a1e-7d74-482e-be94-78c5074595b6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:52 crc kubenswrapper[4955]: I0217 13:23:52.929875 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03690a1e-7d74-482e-be94-78c5074595b6-config-data" (OuterVolumeSpecName: "config-data") pod "03690a1e-7d74-482e-be94-78c5074595b6" (UID: "03690a1e-7d74-482e-be94-78c5074595b6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:52 crc kubenswrapper[4955]: I0217 13:23:52.959486 4955 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Feb 17 13:23:52 crc kubenswrapper[4955]: I0217 13:23:52.965430 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03690a1e-7d74-482e-be94-78c5074595b6-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "03690a1e-7d74-482e-be94-78c5074595b6" (UID: "03690a1e-7d74-482e-be94-78c5074595b6"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.002040 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03690a1e-7d74-482e-be94-78c5074595b6-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.002086 4955 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.002099 4955 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/03690a1e-7d74-482e-be94-78c5074595b6-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.136996 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-aa00-account-create-update-r9kcs" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.205373 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bcpqr\" (UniqueName: \"kubernetes.io/projected/0d0c9416-2de8-4d45-8818-87ae59a2cd6f-kube-api-access-bcpqr\") pod \"0d0c9416-2de8-4d45-8818-87ae59a2cd6f\" (UID: \"0d0c9416-2de8-4d45-8818-87ae59a2cd6f\") " Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.205452 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d0c9416-2de8-4d45-8818-87ae59a2cd6f-operator-scripts\") pod \"0d0c9416-2de8-4d45-8818-87ae59a2cd6f\" (UID: \"0d0c9416-2de8-4d45-8818-87ae59a2cd6f\") " Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.206996 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d0c9416-2de8-4d45-8818-87ae59a2cd6f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0d0c9416-2de8-4d45-8818-87ae59a2cd6f" (UID: "0d0c9416-2de8-4d45-8818-87ae59a2cd6f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.209621 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d0c9416-2de8-4d45-8818-87ae59a2cd6f-kube-api-access-bcpqr" (OuterVolumeSpecName: "kube-api-access-bcpqr") pod "0d0c9416-2de8-4d45-8818-87ae59a2cd6f" (UID: "0d0c9416-2de8-4d45-8818-87ae59a2cd6f"). InnerVolumeSpecName "kube-api-access-bcpqr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.215734 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-a9af-account-create-update-l42hs" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.306794 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rnf9\" (UniqueName: \"kubernetes.io/projected/3d747bbd-45e4-4c6c-a6cb-64a259440231-kube-api-access-5rnf9\") pod \"3d747bbd-45e4-4c6c-a6cb-64a259440231\" (UID: \"3d747bbd-45e4-4c6c-a6cb-64a259440231\") " Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.306962 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3d747bbd-45e4-4c6c-a6cb-64a259440231-operator-scripts\") pod \"3d747bbd-45e4-4c6c-a6cb-64a259440231\" (UID: \"3d747bbd-45e4-4c6c-a6cb-64a259440231\") " Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.307399 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bcpqr\" (UniqueName: \"kubernetes.io/projected/0d0c9416-2de8-4d45-8818-87ae59a2cd6f-kube-api-access-bcpqr\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.307416 4955 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d0c9416-2de8-4d45-8818-87ae59a2cd6f-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.307427 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d747bbd-45e4-4c6c-a6cb-64a259440231-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3d747bbd-45e4-4c6c-a6cb-64a259440231" (UID: "3d747bbd-45e4-4c6c-a6cb-64a259440231"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.310887 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d747bbd-45e4-4c6c-a6cb-64a259440231-kube-api-access-5rnf9" (OuterVolumeSpecName: "kube-api-access-5rnf9") pod "3d747bbd-45e4-4c6c-a6cb-64a259440231" (UID: "3d747bbd-45e4-4c6c-a6cb-64a259440231"). InnerVolumeSpecName "kube-api-access-5rnf9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.408717 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rnf9\" (UniqueName: \"kubernetes.io/projected/3d747bbd-45e4-4c6c-a6cb-64a259440231-kube-api-access-5rnf9\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.408760 4955 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3d747bbd-45e4-4c6c-a6cb-64a259440231-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.424431 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.698766 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-a9af-account-create-update-l42hs" event={"ID":"3d747bbd-45e4-4c6c-a6cb-64a259440231","Type":"ContainerDied","Data":"4ebd828e2628219a5c55fddc42bc8e9907d016aa9018b84c40413616930fcef7"} Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.699119 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ebd828e2628219a5c55fddc42bc8e9907d016aa9018b84c40413616930fcef7" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.698883 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-a9af-account-create-update-l42hs" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.700486 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.710165 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-aa00-account-create-update-r9kcs" event={"ID":"0d0c9416-2de8-4d45-8818-87ae59a2cd6f","Type":"ContainerDied","Data":"c683ac3fda8effab3aebcd4f470d4d721975cf8d55fd210079223f0a15b8d00e"} Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.710224 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c683ac3fda8effab3aebcd4f470d4d721975cf8d55fd210079223f0a15b8d00e" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.710300 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-aa00-account-create-update-r9kcs" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.762833 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.771188 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.793678 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 17 13:23:53 crc kubenswrapper[4955]: E0217 13:23:53.794290 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4bbd808-631c-4a9b-81e6-dee199484728" containerName="mariadb-database-create" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.794314 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4bbd808-631c-4a9b-81e6-dee199484728" containerName="mariadb-database-create" Feb 17 13:23:53 crc kubenswrapper[4955]: E0217 13:23:53.794334 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d747bbd-45e4-4c6c-a6cb-64a259440231" containerName="mariadb-account-create-update" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.794364 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d747bbd-45e4-4c6c-a6cb-64a259440231" containerName="mariadb-account-create-update" Feb 17 13:23:53 crc kubenswrapper[4955]: E0217 13:23:53.794380 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03690a1e-7d74-482e-be94-78c5074595b6" containerName="glance-httpd" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.794388 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="03690a1e-7d74-482e-be94-78c5074595b6" containerName="glance-httpd" Feb 17 13:23:53 crc kubenswrapper[4955]: E0217 13:23:53.794400 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d0c9416-2de8-4d45-8818-87ae59a2cd6f" containerName="mariadb-account-create-update" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.794407 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d0c9416-2de8-4d45-8818-87ae59a2cd6f" containerName="mariadb-account-create-update" Feb 17 13:23:53 crc kubenswrapper[4955]: E0217 13:23:53.794447 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf" containerName="horizon" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.794456 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf" containerName="horizon" Feb 17 13:23:53 crc kubenswrapper[4955]: E0217 13:23:53.794477 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4031cf36-67c1-425d-98f9-0c4f3a4a6065" containerName="mariadb-database-create" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.794485 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="4031cf36-67c1-425d-98f9-0c4f3a4a6065" containerName="mariadb-database-create" Feb 17 13:23:53 crc kubenswrapper[4955]: E0217 13:23:53.794526 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf" containerName="horizon-log" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.794536 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf" containerName="horizon-log" Feb 17 13:23:53 crc kubenswrapper[4955]: E0217 13:23:53.794551 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff0b8526-a0fe-425e-92e8-a4933f3faf61" containerName="mariadb-database-create" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.794601 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff0b8526-a0fe-425e-92e8-a4933f3faf61" containerName="mariadb-database-create" Feb 17 13:23:53 crc kubenswrapper[4955]: E0217 13:23:53.794618 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84b8be7f-2891-44a0-beb3-850911ac8b2a" containerName="mariadb-account-create-update" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.794626 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="84b8be7f-2891-44a0-beb3-850911ac8b2a" containerName="mariadb-account-create-update" Feb 17 13:23:53 crc kubenswrapper[4955]: E0217 13:23:53.794643 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03690a1e-7d74-482e-be94-78c5074595b6" containerName="glance-log" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.794651 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="03690a1e-7d74-482e-be94-78c5074595b6" containerName="glance-log" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.794984 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf" containerName="horizon" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.795003 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="84b8be7f-2891-44a0-beb3-850911ac8b2a" containerName="mariadb-account-create-update" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.795014 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="4031cf36-67c1-425d-98f9-0c4f3a4a6065" containerName="mariadb-database-create" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.795052 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d0c9416-2de8-4d45-8818-87ae59a2cd6f" containerName="mariadb-account-create-update" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.795067 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d747bbd-45e4-4c6c-a6cb-64a259440231" containerName="mariadb-account-create-update" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.795082 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="1aa41a8e-bb33-44bd-abe0-27bbe4ae16cf" containerName="horizon-log" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.795123 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="03690a1e-7d74-482e-be94-78c5074595b6" containerName="glance-httpd" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.795142 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="03690a1e-7d74-482e-be94-78c5074595b6" containerName="glance-log" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.795157 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4bbd808-631c-4a9b-81e6-dee199484728" containerName="mariadb-database-create" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.795169 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff0b8526-a0fe-425e-92e8-a4933f3faf61" containerName="mariadb-database-create" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.796377 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.799461 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.801190 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.823578 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.918225 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72354e86-fe70-4224-b6ea-c7fed8a90f5d-config-data\") pod \"glance-default-external-api-0\" (UID: \"72354e86-fe70-4224-b6ea-c7fed8a90f5d\") " pod="openstack/glance-default-external-api-0" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.918364 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72354e86-fe70-4224-b6ea-c7fed8a90f5d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"72354e86-fe70-4224-b6ea-c7fed8a90f5d\") " pod="openstack/glance-default-external-api-0" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.918407 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/72354e86-fe70-4224-b6ea-c7fed8a90f5d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"72354e86-fe70-4224-b6ea-c7fed8a90f5d\") " pod="openstack/glance-default-external-api-0" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.918447 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/72354e86-fe70-4224-b6ea-c7fed8a90f5d-logs\") pod \"glance-default-external-api-0\" (UID: \"72354e86-fe70-4224-b6ea-c7fed8a90f5d\") " pod="openstack/glance-default-external-api-0" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.918613 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smhmr\" (UniqueName: \"kubernetes.io/projected/72354e86-fe70-4224-b6ea-c7fed8a90f5d-kube-api-access-smhmr\") pod \"glance-default-external-api-0\" (UID: \"72354e86-fe70-4224-b6ea-c7fed8a90f5d\") " pod="openstack/glance-default-external-api-0" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.918704 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72354e86-fe70-4224-b6ea-c7fed8a90f5d-scripts\") pod \"glance-default-external-api-0\" (UID: \"72354e86-fe70-4224-b6ea-c7fed8a90f5d\") " pod="openstack/glance-default-external-api-0" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.918756 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/72354e86-fe70-4224-b6ea-c7fed8a90f5d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"72354e86-fe70-4224-b6ea-c7fed8a90f5d\") " pod="openstack/glance-default-external-api-0" Feb 17 13:23:53 crc kubenswrapper[4955]: I0217 13:23:53.918872 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"72354e86-fe70-4224-b6ea-c7fed8a90f5d\") " pod="openstack/glance-default-external-api-0" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.020386 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72354e86-fe70-4224-b6ea-c7fed8a90f5d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"72354e86-fe70-4224-b6ea-c7fed8a90f5d\") " pod="openstack/glance-default-external-api-0" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.020454 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/72354e86-fe70-4224-b6ea-c7fed8a90f5d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"72354e86-fe70-4224-b6ea-c7fed8a90f5d\") " pod="openstack/glance-default-external-api-0" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.020509 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/72354e86-fe70-4224-b6ea-c7fed8a90f5d-logs\") pod \"glance-default-external-api-0\" (UID: \"72354e86-fe70-4224-b6ea-c7fed8a90f5d\") " pod="openstack/glance-default-external-api-0" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.020538 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smhmr\" (UniqueName: \"kubernetes.io/projected/72354e86-fe70-4224-b6ea-c7fed8a90f5d-kube-api-access-smhmr\") pod \"glance-default-external-api-0\" (UID: \"72354e86-fe70-4224-b6ea-c7fed8a90f5d\") " pod="openstack/glance-default-external-api-0" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.020949 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72354e86-fe70-4224-b6ea-c7fed8a90f5d-scripts\") pod \"glance-default-external-api-0\" (UID: \"72354e86-fe70-4224-b6ea-c7fed8a90f5d\") " pod="openstack/glance-default-external-api-0" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.020976 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/72354e86-fe70-4224-b6ea-c7fed8a90f5d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"72354e86-fe70-4224-b6ea-c7fed8a90f5d\") " pod="openstack/glance-default-external-api-0" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.020897 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/72354e86-fe70-4224-b6ea-c7fed8a90f5d-logs\") pod \"glance-default-external-api-0\" (UID: \"72354e86-fe70-4224-b6ea-c7fed8a90f5d\") " pod="openstack/glance-default-external-api-0" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.021016 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"72354e86-fe70-4224-b6ea-c7fed8a90f5d\") " pod="openstack/glance-default-external-api-0" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.021122 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72354e86-fe70-4224-b6ea-c7fed8a90f5d-config-data\") pod \"glance-default-external-api-0\" (UID: \"72354e86-fe70-4224-b6ea-c7fed8a90f5d\") " pod="openstack/glance-default-external-api-0" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.021664 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/72354e86-fe70-4224-b6ea-c7fed8a90f5d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"72354e86-fe70-4224-b6ea-c7fed8a90f5d\") " pod="openstack/glance-default-external-api-0" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.022013 4955 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"72354e86-fe70-4224-b6ea-c7fed8a90f5d\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-external-api-0" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.024767 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72354e86-fe70-4224-b6ea-c7fed8a90f5d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"72354e86-fe70-4224-b6ea-c7fed8a90f5d\") " pod="openstack/glance-default-external-api-0" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.025071 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/72354e86-fe70-4224-b6ea-c7fed8a90f5d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"72354e86-fe70-4224-b6ea-c7fed8a90f5d\") " pod="openstack/glance-default-external-api-0" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.027525 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72354e86-fe70-4224-b6ea-c7fed8a90f5d-scripts\") pod \"glance-default-external-api-0\" (UID: \"72354e86-fe70-4224-b6ea-c7fed8a90f5d\") " pod="openstack/glance-default-external-api-0" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.041491 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smhmr\" (UniqueName: \"kubernetes.io/projected/72354e86-fe70-4224-b6ea-c7fed8a90f5d-kube-api-access-smhmr\") pod \"glance-default-external-api-0\" (UID: \"72354e86-fe70-4224-b6ea-c7fed8a90f5d\") " pod="openstack/glance-default-external-api-0" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.042527 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72354e86-fe70-4224-b6ea-c7fed8a90f5d-config-data\") pod \"glance-default-external-api-0\" (UID: \"72354e86-fe70-4224-b6ea-c7fed8a90f5d\") " pod="openstack/glance-default-external-api-0" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.051402 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"72354e86-fe70-4224-b6ea-c7fed8a90f5d\") " pod="openstack/glance-default-external-api-0" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.126671 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.165942 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="5da72340-6ea2-42dd-855b-dec4f29f061c" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.156:9292/healthcheck\": read tcp 10.217.0.2:54730->10.217.0.156:9292: read: connection reset by peer" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.166053 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="5da72340-6ea2-42dd-855b-dec4f29f061c" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.156:9292/healthcheck\": read tcp 10.217.0.2:54738->10.217.0.156:9292: read: connection reset by peer" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.253754 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03690a1e-7d74-482e-be94-78c5074595b6" path="/var/lib/kubelet/pods/03690a1e-7d74-482e-be94-78c5074595b6/volumes" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.363697 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-746fcf4988-4q7tw" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.429423 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e-ovndb-tls-certs\") pod \"47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e\" (UID: \"47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e\") " Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.430279 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e-config\") pod \"47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e\" (UID: \"47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e\") " Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.430429 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e-combined-ca-bundle\") pod \"47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e\" (UID: \"47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e\") " Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.430622 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-954vg\" (UniqueName: \"kubernetes.io/projected/47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e-kube-api-access-954vg\") pod \"47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e\" (UID: \"47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e\") " Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.430708 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e-httpd-config\") pod \"47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e\" (UID: \"47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e\") " Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.440911 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e-kube-api-access-954vg" (OuterVolumeSpecName: "kube-api-access-954vg") pod "47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e" (UID: "47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e"). InnerVolumeSpecName "kube-api-access-954vg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.444040 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e" (UID: "47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.502948 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e-config" (OuterVolumeSpecName: "config") pod "47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e" (UID: "47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.509271 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e" (UID: "47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.530965 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e" (UID: "47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.533196 4955 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.533239 4955 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.533254 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.533268 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-954vg\" (UniqueName: \"kubernetes.io/projected/47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e-kube-api-access-954vg\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.533285 4955 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e-httpd-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.752819 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e56111b4-4b8a-4e07-bba9-68d34fd179fb","Type":"ContainerStarted","Data":"d11dcdf9098672bbd8b3304411081345a023f689c8d7cd17d0406b3a267eb291"} Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.753116 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e56111b4-4b8a-4e07-bba9-68d34fd179fb" containerName="ceilometer-central-agent" containerID="cri-o://f520260e9c3bbb1b63ec8f9522bd9fcc6cd69d7e6ab2a4af3b330077c6fc553f" gracePeriod=30 Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.753237 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.753271 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e56111b4-4b8a-4e07-bba9-68d34fd179fb" containerName="proxy-httpd" containerID="cri-o://d11dcdf9098672bbd8b3304411081345a023f689c8d7cd17d0406b3a267eb291" gracePeriod=30 Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.753331 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e56111b4-4b8a-4e07-bba9-68d34fd179fb" containerName="sg-core" containerID="cri-o://4683a112963c8aed791f163ff609e4a0b3f319e5243c44eac71422cec74ea161" gracePeriod=30 Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.753395 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e56111b4-4b8a-4e07-bba9-68d34fd179fb" containerName="ceilometer-notification-agent" containerID="cri-o://89e46bb8b72cfb0f7210ef620927b81062fa54f4cd2c210711b1668eae6ffcac" gracePeriod=30 Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.777530 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.466426082 podStartE2EDuration="7.777511357s" podCreationTimestamp="2026-02-17 13:23:47 +0000 UTC" firstStartedPulling="2026-02-17 13:23:48.516786064 +0000 UTC m=+1167.039515607" lastFinishedPulling="2026-02-17 13:23:53.827871339 +0000 UTC m=+1172.350600882" observedRunningTime="2026-02-17 13:23:54.775707636 +0000 UTC m=+1173.298437189" watchObservedRunningTime="2026-02-17 13:23:54.777511357 +0000 UTC m=+1173.300240900" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.786279 4955 generic.go:334] "Generic (PLEG): container finished" podID="47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e" containerID="6d565c62d191904d43404bc26ded506c5b6c75a7f0d89b1ca2702cf83d5ad913" exitCode=0 Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.786400 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-746fcf4988-4q7tw" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.786756 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-746fcf4988-4q7tw" event={"ID":"47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e","Type":"ContainerDied","Data":"6d565c62d191904d43404bc26ded506c5b6c75a7f0d89b1ca2702cf83d5ad913"} Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.786856 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-746fcf4988-4q7tw" event={"ID":"47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e","Type":"ContainerDied","Data":"d94ba57dd3ef3b201c37616bab2b39c14a4bdb073ea40f077bf62a736e4f5253"} Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.786881 4955 scope.go:117] "RemoveContainer" containerID="6df53ab637202b08040479088ed897dde6492ba6c8f79fb9bae89a3e98c0cd26" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.787671 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.795635 4955 generic.go:334] "Generic (PLEG): container finished" podID="5da72340-6ea2-42dd-855b-dec4f29f061c" containerID="172f55b5031dad85034b5db9a36dfbaf6236f4dcc6a4d44cdd9d2c42d86253b2" exitCode=0 Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.795709 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5da72340-6ea2-42dd-855b-dec4f29f061c","Type":"ContainerDied","Data":"172f55b5031dad85034b5db9a36dfbaf6236f4dcc6a4d44cdd9d2c42d86253b2"} Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.796059 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5da72340-6ea2-42dd-855b-dec4f29f061c","Type":"ContainerDied","Data":"e1132736f08468db8c3016fd720a91e1991cb8ec1e936471c62231597135f2ee"} Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.842931 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.850762 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5da72340-6ea2-42dd-855b-dec4f29f061c-httpd-run\") pod \"5da72340-6ea2-42dd-855b-dec4f29f061c\" (UID: \"5da72340-6ea2-42dd-855b-dec4f29f061c\") " Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.850863 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5da72340-6ea2-42dd-855b-dec4f29f061c-internal-tls-certs\") pod \"5da72340-6ea2-42dd-855b-dec4f29f061c\" (UID: \"5da72340-6ea2-42dd-855b-dec4f29f061c\") " Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.850927 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5da72340-6ea2-42dd-855b-dec4f29f061c-combined-ca-bundle\") pod \"5da72340-6ea2-42dd-855b-dec4f29f061c\" (UID: \"5da72340-6ea2-42dd-855b-dec4f29f061c\") " Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.850958 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5da72340-6ea2-42dd-855b-dec4f29f061c-config-data\") pod \"5da72340-6ea2-42dd-855b-dec4f29f061c\" (UID: \"5da72340-6ea2-42dd-855b-dec4f29f061c\") " Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.851054 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5da72340-6ea2-42dd-855b-dec4f29f061c-logs\") pod \"5da72340-6ea2-42dd-855b-dec4f29f061c\" (UID: \"5da72340-6ea2-42dd-855b-dec4f29f061c\") " Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.851084 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"5da72340-6ea2-42dd-855b-dec4f29f061c\" (UID: \"5da72340-6ea2-42dd-855b-dec4f29f061c\") " Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.851107 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-48w4x\" (UniqueName: \"kubernetes.io/projected/5da72340-6ea2-42dd-855b-dec4f29f061c-kube-api-access-48w4x\") pod \"5da72340-6ea2-42dd-855b-dec4f29f061c\" (UID: \"5da72340-6ea2-42dd-855b-dec4f29f061c\") " Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.851162 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5da72340-6ea2-42dd-855b-dec4f29f061c-scripts\") pod \"5da72340-6ea2-42dd-855b-dec4f29f061c\" (UID: \"5da72340-6ea2-42dd-855b-dec4f29f061c\") " Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.856418 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5da72340-6ea2-42dd-855b-dec4f29f061c-scripts" (OuterVolumeSpecName: "scripts") pod "5da72340-6ea2-42dd-855b-dec4f29f061c" (UID: "5da72340-6ea2-42dd-855b-dec4f29f061c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.858638 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5da72340-6ea2-42dd-855b-dec4f29f061c-logs" (OuterVolumeSpecName: "logs") pod "5da72340-6ea2-42dd-855b-dec4f29f061c" (UID: "5da72340-6ea2-42dd-855b-dec4f29f061c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.860327 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5da72340-6ea2-42dd-855b-dec4f29f061c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5da72340-6ea2-42dd-855b-dec4f29f061c" (UID: "5da72340-6ea2-42dd-855b-dec4f29f061c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.863971 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "5da72340-6ea2-42dd-855b-dec4f29f061c" (UID: "5da72340-6ea2-42dd-855b-dec4f29f061c"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.880199 4955 scope.go:117] "RemoveContainer" containerID="6d565c62d191904d43404bc26ded506c5b6c75a7f0d89b1ca2702cf83d5ad913" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.907989 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5da72340-6ea2-42dd-855b-dec4f29f061c-kube-api-access-48w4x" (OuterVolumeSpecName: "kube-api-access-48w4x") pod "5da72340-6ea2-42dd-855b-dec4f29f061c" (UID: "5da72340-6ea2-42dd-855b-dec4f29f061c"). InnerVolumeSpecName "kube-api-access-48w4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.927990 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-746fcf4988-4q7tw"] Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.929042 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5da72340-6ea2-42dd-855b-dec4f29f061c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5da72340-6ea2-42dd-855b-dec4f29f061c" (UID: "5da72340-6ea2-42dd-855b-dec4f29f061c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.953242 4955 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5da72340-6ea2-42dd-855b-dec4f29f061c-logs\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.953295 4955 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.953311 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-48w4x\" (UniqueName: \"kubernetes.io/projected/5da72340-6ea2-42dd-855b-dec4f29f061c-kube-api-access-48w4x\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.953323 4955 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5da72340-6ea2-42dd-855b-dec4f29f061c-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.953333 4955 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5da72340-6ea2-42dd-855b-dec4f29f061c-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.953344 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5da72340-6ea2-42dd-855b-dec4f29f061c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.954683 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-746fcf4988-4q7tw"] Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.964947 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5da72340-6ea2-42dd-855b-dec4f29f061c-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "5da72340-6ea2-42dd-855b-dec4f29f061c" (UID: "5da72340-6ea2-42dd-855b-dec4f29f061c"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:54 crc kubenswrapper[4955]: I0217 13:23:54.974308 4955 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.015074 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5da72340-6ea2-42dd-855b-dec4f29f061c-config-data" (OuterVolumeSpecName: "config-data") pod "5da72340-6ea2-42dd-855b-dec4f29f061c" (UID: "5da72340-6ea2-42dd-855b-dec4f29f061c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.040913 4955 scope.go:117] "RemoveContainer" containerID="6df53ab637202b08040479088ed897dde6492ba6c8f79fb9bae89a3e98c0cd26" Feb 17 13:23:55 crc kubenswrapper[4955]: E0217 13:23:55.047693 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6df53ab637202b08040479088ed897dde6492ba6c8f79fb9bae89a3e98c0cd26\": container with ID starting with 6df53ab637202b08040479088ed897dde6492ba6c8f79fb9bae89a3e98c0cd26 not found: ID does not exist" containerID="6df53ab637202b08040479088ed897dde6492ba6c8f79fb9bae89a3e98c0cd26" Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.047748 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6df53ab637202b08040479088ed897dde6492ba6c8f79fb9bae89a3e98c0cd26"} err="failed to get container status \"6df53ab637202b08040479088ed897dde6492ba6c8f79fb9bae89a3e98c0cd26\": rpc error: code = NotFound desc = could not find container \"6df53ab637202b08040479088ed897dde6492ba6c8f79fb9bae89a3e98c0cd26\": container with ID starting with 6df53ab637202b08040479088ed897dde6492ba6c8f79fb9bae89a3e98c0cd26 not found: ID does not exist" Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.047784 4955 scope.go:117] "RemoveContainer" containerID="6d565c62d191904d43404bc26ded506c5b6c75a7f0d89b1ca2702cf83d5ad913" Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.055242 4955 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5da72340-6ea2-42dd-855b-dec4f29f061c-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.055285 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5da72340-6ea2-42dd-855b-dec4f29f061c-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.055299 4955 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:55 crc kubenswrapper[4955]: E0217 13:23:55.056908 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d565c62d191904d43404bc26ded506c5b6c75a7f0d89b1ca2702cf83d5ad913\": container with ID starting with 6d565c62d191904d43404bc26ded506c5b6c75a7f0d89b1ca2702cf83d5ad913 not found: ID does not exist" containerID="6d565c62d191904d43404bc26ded506c5b6c75a7f0d89b1ca2702cf83d5ad913" Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.056955 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d565c62d191904d43404bc26ded506c5b6c75a7f0d89b1ca2702cf83d5ad913"} err="failed to get container status \"6d565c62d191904d43404bc26ded506c5b6c75a7f0d89b1ca2702cf83d5ad913\": rpc error: code = NotFound desc = could not find container \"6d565c62d191904d43404bc26ded506c5b6c75a7f0d89b1ca2702cf83d5ad913\": container with ID starting with 6d565c62d191904d43404bc26ded506c5b6c75a7f0d89b1ca2702cf83d5ad913 not found: ID does not exist" Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.056989 4955 scope.go:117] "RemoveContainer" containerID="172f55b5031dad85034b5db9a36dfbaf6236f4dcc6a4d44cdd9d2c42d86253b2" Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.088268 4955 scope.go:117] "RemoveContainer" containerID="29b22587a44a1d3f014b0d32e0afc120c7ed49e76d1b13702dd03d797cd374cc" Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.120227 4955 scope.go:117] "RemoveContainer" containerID="172f55b5031dad85034b5db9a36dfbaf6236f4dcc6a4d44cdd9d2c42d86253b2" Feb 17 13:23:55 crc kubenswrapper[4955]: E0217 13:23:55.121437 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"172f55b5031dad85034b5db9a36dfbaf6236f4dcc6a4d44cdd9d2c42d86253b2\": container with ID starting with 172f55b5031dad85034b5db9a36dfbaf6236f4dcc6a4d44cdd9d2c42d86253b2 not found: ID does not exist" containerID="172f55b5031dad85034b5db9a36dfbaf6236f4dcc6a4d44cdd9d2c42d86253b2" Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.121463 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"172f55b5031dad85034b5db9a36dfbaf6236f4dcc6a4d44cdd9d2c42d86253b2"} err="failed to get container status \"172f55b5031dad85034b5db9a36dfbaf6236f4dcc6a4d44cdd9d2c42d86253b2\": rpc error: code = NotFound desc = could not find container \"172f55b5031dad85034b5db9a36dfbaf6236f4dcc6a4d44cdd9d2c42d86253b2\": container with ID starting with 172f55b5031dad85034b5db9a36dfbaf6236f4dcc6a4d44cdd9d2c42d86253b2 not found: ID does not exist" Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.121484 4955 scope.go:117] "RemoveContainer" containerID="29b22587a44a1d3f014b0d32e0afc120c7ed49e76d1b13702dd03d797cd374cc" Feb 17 13:23:55 crc kubenswrapper[4955]: E0217 13:23:55.121758 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29b22587a44a1d3f014b0d32e0afc120c7ed49e76d1b13702dd03d797cd374cc\": container with ID starting with 29b22587a44a1d3f014b0d32e0afc120c7ed49e76d1b13702dd03d797cd374cc not found: ID does not exist" containerID="29b22587a44a1d3f014b0d32e0afc120c7ed49e76d1b13702dd03d797cd374cc" Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.121783 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29b22587a44a1d3f014b0d32e0afc120c7ed49e76d1b13702dd03d797cd374cc"} err="failed to get container status \"29b22587a44a1d3f014b0d32e0afc120c7ed49e76d1b13702dd03d797cd374cc\": rpc error: code = NotFound desc = could not find container \"29b22587a44a1d3f014b0d32e0afc120c7ed49e76d1b13702dd03d797cd374cc\": container with ID starting with 29b22587a44a1d3f014b0d32e0afc120c7ed49e76d1b13702dd03d797cd374cc not found: ID does not exist" Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.812004 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.822310 4955 generic.go:334] "Generic (PLEG): container finished" podID="e56111b4-4b8a-4e07-bba9-68d34fd179fb" containerID="d11dcdf9098672bbd8b3304411081345a023f689c8d7cd17d0406b3a267eb291" exitCode=0 Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.822339 4955 generic.go:334] "Generic (PLEG): container finished" podID="e56111b4-4b8a-4e07-bba9-68d34fd179fb" containerID="4683a112963c8aed791f163ff609e4a0b3f319e5243c44eac71422cec74ea161" exitCode=2 Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.822349 4955 generic.go:334] "Generic (PLEG): container finished" podID="e56111b4-4b8a-4e07-bba9-68d34fd179fb" containerID="89e46bb8b72cfb0f7210ef620927b81062fa54f4cd2c210711b1668eae6ffcac" exitCode=0 Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.822406 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e56111b4-4b8a-4e07-bba9-68d34fd179fb","Type":"ContainerDied","Data":"d11dcdf9098672bbd8b3304411081345a023f689c8d7cd17d0406b3a267eb291"} Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.822429 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e56111b4-4b8a-4e07-bba9-68d34fd179fb","Type":"ContainerDied","Data":"4683a112963c8aed791f163ff609e4a0b3f319e5243c44eac71422cec74ea161"} Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.822459 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e56111b4-4b8a-4e07-bba9-68d34fd179fb","Type":"ContainerDied","Data":"89e46bb8b72cfb0f7210ef620927b81062fa54f4cd2c210711b1668eae6ffcac"} Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.824405 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"72354e86-fe70-4224-b6ea-c7fed8a90f5d","Type":"ContainerStarted","Data":"fc1756a32643e117a438c5ec0c9f03822194f0ff7d44378d5c0e3bb8003595a5"} Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.824458 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"72354e86-fe70-4224-b6ea-c7fed8a90f5d","Type":"ContainerStarted","Data":"e721f3875dfc742b31de8af0eafa5a9f0103a02d4cde55fc8506ee9f198b3a34"} Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.863884 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.877851 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.901916 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 17 13:23:55 crc kubenswrapper[4955]: E0217 13:23:55.902354 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5da72340-6ea2-42dd-855b-dec4f29f061c" containerName="glance-httpd" Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.902373 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="5da72340-6ea2-42dd-855b-dec4f29f061c" containerName="glance-httpd" Feb 17 13:23:55 crc kubenswrapper[4955]: E0217 13:23:55.902399 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e" containerName="neutron-api" Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.902407 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e" containerName="neutron-api" Feb 17 13:23:55 crc kubenswrapper[4955]: E0217 13:23:55.902424 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e" containerName="neutron-httpd" Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.902432 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e" containerName="neutron-httpd" Feb 17 13:23:55 crc kubenswrapper[4955]: E0217 13:23:55.902453 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5da72340-6ea2-42dd-855b-dec4f29f061c" containerName="glance-log" Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.902461 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="5da72340-6ea2-42dd-855b-dec4f29f061c" containerName="glance-log" Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.902675 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="5da72340-6ea2-42dd-855b-dec4f29f061c" containerName="glance-httpd" Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.902694 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e" containerName="neutron-api" Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.902718 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="5da72340-6ea2-42dd-855b-dec4f29f061c" containerName="glance-log" Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.902729 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e" containerName="neutron-httpd" Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.903886 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.906427 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.906452 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.915767 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.971207 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggqkc\" (UniqueName: \"kubernetes.io/projected/539c0e46-aba7-4b5b-a176-3e9ac9089560-kube-api-access-ggqkc\") pod \"glance-default-internal-api-0\" (UID: \"539c0e46-aba7-4b5b-a176-3e9ac9089560\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.971266 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/539c0e46-aba7-4b5b-a176-3e9ac9089560-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"539c0e46-aba7-4b5b-a176-3e9ac9089560\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.971314 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/539c0e46-aba7-4b5b-a176-3e9ac9089560-scripts\") pod \"glance-default-internal-api-0\" (UID: \"539c0e46-aba7-4b5b-a176-3e9ac9089560\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.971436 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/539c0e46-aba7-4b5b-a176-3e9ac9089560-config-data\") pod \"glance-default-internal-api-0\" (UID: \"539c0e46-aba7-4b5b-a176-3e9ac9089560\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.971493 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/539c0e46-aba7-4b5b-a176-3e9ac9089560-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"539c0e46-aba7-4b5b-a176-3e9ac9089560\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.971532 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/539c0e46-aba7-4b5b-a176-3e9ac9089560-logs\") pod \"glance-default-internal-api-0\" (UID: \"539c0e46-aba7-4b5b-a176-3e9ac9089560\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.971563 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/539c0e46-aba7-4b5b-a176-3e9ac9089560-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"539c0e46-aba7-4b5b-a176-3e9ac9089560\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:23:55 crc kubenswrapper[4955]: I0217 13:23:55.971588 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"539c0e46-aba7-4b5b-a176-3e9ac9089560\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:23:56 crc kubenswrapper[4955]: I0217 13:23:56.073138 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/539c0e46-aba7-4b5b-a176-3e9ac9089560-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"539c0e46-aba7-4b5b-a176-3e9ac9089560\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:23:56 crc kubenswrapper[4955]: I0217 13:23:56.073199 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/539c0e46-aba7-4b5b-a176-3e9ac9089560-logs\") pod \"glance-default-internal-api-0\" (UID: \"539c0e46-aba7-4b5b-a176-3e9ac9089560\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:23:56 crc kubenswrapper[4955]: I0217 13:23:56.073242 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/539c0e46-aba7-4b5b-a176-3e9ac9089560-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"539c0e46-aba7-4b5b-a176-3e9ac9089560\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:23:56 crc kubenswrapper[4955]: I0217 13:23:56.073272 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"539c0e46-aba7-4b5b-a176-3e9ac9089560\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:23:56 crc kubenswrapper[4955]: I0217 13:23:56.073381 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggqkc\" (UniqueName: \"kubernetes.io/projected/539c0e46-aba7-4b5b-a176-3e9ac9089560-kube-api-access-ggqkc\") pod \"glance-default-internal-api-0\" (UID: \"539c0e46-aba7-4b5b-a176-3e9ac9089560\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:23:56 crc kubenswrapper[4955]: I0217 13:23:56.073405 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/539c0e46-aba7-4b5b-a176-3e9ac9089560-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"539c0e46-aba7-4b5b-a176-3e9ac9089560\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:23:56 crc kubenswrapper[4955]: I0217 13:23:56.073435 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/539c0e46-aba7-4b5b-a176-3e9ac9089560-scripts\") pod \"glance-default-internal-api-0\" (UID: \"539c0e46-aba7-4b5b-a176-3e9ac9089560\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:23:56 crc kubenswrapper[4955]: I0217 13:23:56.073491 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/539c0e46-aba7-4b5b-a176-3e9ac9089560-config-data\") pod \"glance-default-internal-api-0\" (UID: \"539c0e46-aba7-4b5b-a176-3e9ac9089560\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:23:56 crc kubenswrapper[4955]: I0217 13:23:56.074341 4955 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"539c0e46-aba7-4b5b-a176-3e9ac9089560\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Feb 17 13:23:56 crc kubenswrapper[4955]: I0217 13:23:56.076123 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/539c0e46-aba7-4b5b-a176-3e9ac9089560-logs\") pod \"glance-default-internal-api-0\" (UID: \"539c0e46-aba7-4b5b-a176-3e9ac9089560\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:23:56 crc kubenswrapper[4955]: I0217 13:23:56.076635 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/539c0e46-aba7-4b5b-a176-3e9ac9089560-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"539c0e46-aba7-4b5b-a176-3e9ac9089560\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:23:56 crc kubenswrapper[4955]: I0217 13:23:56.079882 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/539c0e46-aba7-4b5b-a176-3e9ac9089560-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"539c0e46-aba7-4b5b-a176-3e9ac9089560\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:23:56 crc kubenswrapper[4955]: I0217 13:23:56.081758 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/539c0e46-aba7-4b5b-a176-3e9ac9089560-config-data\") pod \"glance-default-internal-api-0\" (UID: \"539c0e46-aba7-4b5b-a176-3e9ac9089560\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:23:56 crc kubenswrapper[4955]: I0217 13:23:56.087404 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/539c0e46-aba7-4b5b-a176-3e9ac9089560-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"539c0e46-aba7-4b5b-a176-3e9ac9089560\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:23:56 crc kubenswrapper[4955]: I0217 13:23:56.102079 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/539c0e46-aba7-4b5b-a176-3e9ac9089560-scripts\") pod \"glance-default-internal-api-0\" (UID: \"539c0e46-aba7-4b5b-a176-3e9ac9089560\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:23:56 crc kubenswrapper[4955]: I0217 13:23:56.139143 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggqkc\" (UniqueName: \"kubernetes.io/projected/539c0e46-aba7-4b5b-a176-3e9ac9089560-kube-api-access-ggqkc\") pod \"glance-default-internal-api-0\" (UID: \"539c0e46-aba7-4b5b-a176-3e9ac9089560\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:23:56 crc kubenswrapper[4955]: I0217 13:23:56.153210 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"539c0e46-aba7-4b5b-a176-3e9ac9089560\") " pod="openstack/glance-default-internal-api-0" Feb 17 13:23:56 crc kubenswrapper[4955]: I0217 13:23:56.229244 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 17 13:23:56 crc kubenswrapper[4955]: I0217 13:23:56.246668 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e" path="/var/lib/kubelet/pods/47ec3f8a-3b61-4d9c-8f9a-b1b2b581093e/volumes" Feb 17 13:23:56 crc kubenswrapper[4955]: I0217 13:23:56.247555 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5da72340-6ea2-42dd-855b-dec4f29f061c" path="/var/lib/kubelet/pods/5da72340-6ea2-42dd-855b-dec4f29f061c/volumes" Feb 17 13:23:56 crc kubenswrapper[4955]: I0217 13:23:56.779147 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 17 13:23:56 crc kubenswrapper[4955]: I0217 13:23:56.839835 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"72354e86-fe70-4224-b6ea-c7fed8a90f5d","Type":"ContainerStarted","Data":"dca623aa1c39bc203d7507942bc6d95d33109ff1b4f0a9a1bf2c6ecdbfe1dfb4"} Feb 17 13:23:56 crc kubenswrapper[4955]: I0217 13:23:56.842606 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"539c0e46-aba7-4b5b-a176-3e9ac9089560","Type":"ContainerStarted","Data":"11787a9e3e989382fb40dbcade6da2974cc9450227833513b2b0008da6e6f834"} Feb 17 13:23:56 crc kubenswrapper[4955]: I0217 13:23:56.854248 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5bcc76b6cf-4r2g9" Feb 17 13:23:56 crc kubenswrapper[4955]: I0217 13:23:56.857723 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5bcc76b6cf-4r2g9" Feb 17 13:23:56 crc kubenswrapper[4955]: I0217 13:23:56.868886 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.8688671230000002 podStartE2EDuration="3.868867123s" podCreationTimestamp="2026-02-17 13:23:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:23:56.867577576 +0000 UTC m=+1175.390307119" watchObservedRunningTime="2026-02-17 13:23:56.868867123 +0000 UTC m=+1175.391596666" Feb 17 13:23:57 crc kubenswrapper[4955]: I0217 13:23:57.868040 4955 generic.go:334] "Generic (PLEG): container finished" podID="e56111b4-4b8a-4e07-bba9-68d34fd179fb" containerID="f520260e9c3bbb1b63ec8f9522bd9fcc6cd69d7e6ab2a4af3b330077c6fc553f" exitCode=0 Feb 17 13:23:57 crc kubenswrapper[4955]: I0217 13:23:57.868182 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e56111b4-4b8a-4e07-bba9-68d34fd179fb","Type":"ContainerDied","Data":"f520260e9c3bbb1b63ec8f9522bd9fcc6cd69d7e6ab2a4af3b330077c6fc553f"} Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.278184 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ffnlw"] Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.279591 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-ffnlw" Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.283570 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-rz62x" Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.283765 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.283799 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.288330 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ffnlw"] Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.318817 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6zcm\" (UniqueName: \"kubernetes.io/projected/fbdb6999-85bf-4bd2-b7f4-7b7c8566941f-kube-api-access-x6zcm\") pod \"nova-cell0-conductor-db-sync-ffnlw\" (UID: \"fbdb6999-85bf-4bd2-b7f4-7b7c8566941f\") " pod="openstack/nova-cell0-conductor-db-sync-ffnlw" Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.318890 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbdb6999-85bf-4bd2-b7f4-7b7c8566941f-scripts\") pod \"nova-cell0-conductor-db-sync-ffnlw\" (UID: \"fbdb6999-85bf-4bd2-b7f4-7b7c8566941f\") " pod="openstack/nova-cell0-conductor-db-sync-ffnlw" Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.319049 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbdb6999-85bf-4bd2-b7f4-7b7c8566941f-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-ffnlw\" (UID: \"fbdb6999-85bf-4bd2-b7f4-7b7c8566941f\") " pod="openstack/nova-cell0-conductor-db-sync-ffnlw" Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.319138 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbdb6999-85bf-4bd2-b7f4-7b7c8566941f-config-data\") pod \"nova-cell0-conductor-db-sync-ffnlw\" (UID: \"fbdb6999-85bf-4bd2-b7f4-7b7c8566941f\") " pod="openstack/nova-cell0-conductor-db-sync-ffnlw" Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.421327 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbdb6999-85bf-4bd2-b7f4-7b7c8566941f-scripts\") pod \"nova-cell0-conductor-db-sync-ffnlw\" (UID: \"fbdb6999-85bf-4bd2-b7f4-7b7c8566941f\") " pod="openstack/nova-cell0-conductor-db-sync-ffnlw" Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.421402 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbdb6999-85bf-4bd2-b7f4-7b7c8566941f-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-ffnlw\" (UID: \"fbdb6999-85bf-4bd2-b7f4-7b7c8566941f\") " pod="openstack/nova-cell0-conductor-db-sync-ffnlw" Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.421449 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbdb6999-85bf-4bd2-b7f4-7b7c8566941f-config-data\") pod \"nova-cell0-conductor-db-sync-ffnlw\" (UID: \"fbdb6999-85bf-4bd2-b7f4-7b7c8566941f\") " pod="openstack/nova-cell0-conductor-db-sync-ffnlw" Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.421568 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6zcm\" (UniqueName: \"kubernetes.io/projected/fbdb6999-85bf-4bd2-b7f4-7b7c8566941f-kube-api-access-x6zcm\") pod \"nova-cell0-conductor-db-sync-ffnlw\" (UID: \"fbdb6999-85bf-4bd2-b7f4-7b7c8566941f\") " pod="openstack/nova-cell0-conductor-db-sync-ffnlw" Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.430490 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbdb6999-85bf-4bd2-b7f4-7b7c8566941f-scripts\") pod \"nova-cell0-conductor-db-sync-ffnlw\" (UID: \"fbdb6999-85bf-4bd2-b7f4-7b7c8566941f\") " pod="openstack/nova-cell0-conductor-db-sync-ffnlw" Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.433345 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbdb6999-85bf-4bd2-b7f4-7b7c8566941f-config-data\") pod \"nova-cell0-conductor-db-sync-ffnlw\" (UID: \"fbdb6999-85bf-4bd2-b7f4-7b7c8566941f\") " pod="openstack/nova-cell0-conductor-db-sync-ffnlw" Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.444163 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbdb6999-85bf-4bd2-b7f4-7b7c8566941f-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-ffnlw\" (UID: \"fbdb6999-85bf-4bd2-b7f4-7b7c8566941f\") " pod="openstack/nova-cell0-conductor-db-sync-ffnlw" Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.455623 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6zcm\" (UniqueName: \"kubernetes.io/projected/fbdb6999-85bf-4bd2-b7f4-7b7c8566941f-kube-api-access-x6zcm\") pod \"nova-cell0-conductor-db-sync-ffnlw\" (UID: \"fbdb6999-85bf-4bd2-b7f4-7b7c8566941f\") " pod="openstack/nova-cell0-conductor-db-sync-ffnlw" Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.613572 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5d56c994-tx7w7" Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.626469 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-ffnlw" Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.712013 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.725840 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e56111b4-4b8a-4e07-bba9-68d34fd179fb-run-httpd\") pod \"e56111b4-4b8a-4e07-bba9-68d34fd179fb\" (UID: \"e56111b4-4b8a-4e07-bba9-68d34fd179fb\") " Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.725883 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e56111b4-4b8a-4e07-bba9-68d34fd179fb-log-httpd\") pod \"e56111b4-4b8a-4e07-bba9-68d34fd179fb\" (UID: \"e56111b4-4b8a-4e07-bba9-68d34fd179fb\") " Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.725902 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wf4h6\" (UniqueName: \"kubernetes.io/projected/e56111b4-4b8a-4e07-bba9-68d34fd179fb-kube-api-access-wf4h6\") pod \"e56111b4-4b8a-4e07-bba9-68d34fd179fb\" (UID: \"e56111b4-4b8a-4e07-bba9-68d34fd179fb\") " Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.725933 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e56111b4-4b8a-4e07-bba9-68d34fd179fb-scripts\") pod \"e56111b4-4b8a-4e07-bba9-68d34fd179fb\" (UID: \"e56111b4-4b8a-4e07-bba9-68d34fd179fb\") " Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.726050 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e56111b4-4b8a-4e07-bba9-68d34fd179fb-sg-core-conf-yaml\") pod \"e56111b4-4b8a-4e07-bba9-68d34fd179fb\" (UID: \"e56111b4-4b8a-4e07-bba9-68d34fd179fb\") " Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.726179 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e56111b4-4b8a-4e07-bba9-68d34fd179fb-combined-ca-bundle\") pod \"e56111b4-4b8a-4e07-bba9-68d34fd179fb\" (UID: \"e56111b4-4b8a-4e07-bba9-68d34fd179fb\") " Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.726227 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e56111b4-4b8a-4e07-bba9-68d34fd179fb-config-data\") pod \"e56111b4-4b8a-4e07-bba9-68d34fd179fb\" (UID: \"e56111b4-4b8a-4e07-bba9-68d34fd179fb\") " Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.726278 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e56111b4-4b8a-4e07-bba9-68d34fd179fb-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e56111b4-4b8a-4e07-bba9-68d34fd179fb" (UID: "e56111b4-4b8a-4e07-bba9-68d34fd179fb"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.726537 4955 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e56111b4-4b8a-4e07-bba9-68d34fd179fb-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.727269 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e56111b4-4b8a-4e07-bba9-68d34fd179fb-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e56111b4-4b8a-4e07-bba9-68d34fd179fb" (UID: "e56111b4-4b8a-4e07-bba9-68d34fd179fb"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.731806 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5d56c994-tx7w7" Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.740205 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e56111b4-4b8a-4e07-bba9-68d34fd179fb-scripts" (OuterVolumeSpecName: "scripts") pod "e56111b4-4b8a-4e07-bba9-68d34fd179fb" (UID: "e56111b4-4b8a-4e07-bba9-68d34fd179fb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.740670 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e56111b4-4b8a-4e07-bba9-68d34fd179fb-kube-api-access-wf4h6" (OuterVolumeSpecName: "kube-api-access-wf4h6") pod "e56111b4-4b8a-4e07-bba9-68d34fd179fb" (UID: "e56111b4-4b8a-4e07-bba9-68d34fd179fb"). InnerVolumeSpecName "kube-api-access-wf4h6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.806967 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e56111b4-4b8a-4e07-bba9-68d34fd179fb-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e56111b4-4b8a-4e07-bba9-68d34fd179fb" (UID: "e56111b4-4b8a-4e07-bba9-68d34fd179fb"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.831398 4955 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e56111b4-4b8a-4e07-bba9-68d34fd179fb-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.831432 4955 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e56111b4-4b8a-4e07-bba9-68d34fd179fb-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.831445 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wf4h6\" (UniqueName: \"kubernetes.io/projected/e56111b4-4b8a-4e07-bba9-68d34fd179fb-kube-api-access-wf4h6\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.831456 4955 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e56111b4-4b8a-4e07-bba9-68d34fd179fb-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.901264 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-845ffbbf5d-ntk67"] Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.901497 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-845ffbbf5d-ntk67" podUID="20d6bdbe-1dd5-4790-add1-2ea3f3496315" containerName="placement-log" containerID="cri-o://a20011351ac1920632e1b4557ffb5d35fab17f1d712096d2aa96c1115c17a75a" gracePeriod=30 Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.901908 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-845ffbbf5d-ntk67" podUID="20d6bdbe-1dd5-4790-add1-2ea3f3496315" containerName="placement-api" containerID="cri-o://f550116db1414f335a606a7e1ad151d2ee5761dfeb702452fed5a5ea730d9b78" gracePeriod=30 Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.935972 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.936356 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e56111b4-4b8a-4e07-bba9-68d34fd179fb","Type":"ContainerDied","Data":"6b71c055acde1fa41a4d33dd3bfa35d899ed637a32058d82efe419386c8dd1e2"} Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.936423 4955 scope.go:117] "RemoveContainer" containerID="d11dcdf9098672bbd8b3304411081345a023f689c8d7cd17d0406b3a267eb291" Feb 17 13:23:58 crc kubenswrapper[4955]: E0217 13:23:58.947695 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e56111b4-4b8a-4e07-bba9-68d34fd179fb-config-data podName:e56111b4-4b8a-4e07-bba9-68d34fd179fb nodeName:}" failed. No retries permitted until 2026-02-17 13:23:59.4476617 +0000 UTC m=+1177.970391243 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config-data" (UniqueName: "kubernetes.io/secret/e56111b4-4b8a-4e07-bba9-68d34fd179fb-config-data") pod "e56111b4-4b8a-4e07-bba9-68d34fd179fb" (UID: "e56111b4-4b8a-4e07-bba9-68d34fd179fb") : error deleting /var/lib/kubelet/pods/e56111b4-4b8a-4e07-bba9-68d34fd179fb/volume-subpaths: remove /var/lib/kubelet/pods/e56111b4-4b8a-4e07-bba9-68d34fd179fb/volume-subpaths: no such file or directory Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.954948 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e56111b4-4b8a-4e07-bba9-68d34fd179fb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e56111b4-4b8a-4e07-bba9-68d34fd179fb" (UID: "e56111b4-4b8a-4e07-bba9-68d34fd179fb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:58 crc kubenswrapper[4955]: I0217 13:23:58.973737 4955 scope.go:117] "RemoveContainer" containerID="4683a112963c8aed791f163ff609e4a0b3f319e5243c44eac71422cec74ea161" Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.000226 4955 scope.go:117] "RemoveContainer" containerID="89e46bb8b72cfb0f7210ef620927b81062fa54f4cd2c210711b1668eae6ffcac" Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.034662 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e56111b4-4b8a-4e07-bba9-68d34fd179fb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.060940 4955 scope.go:117] "RemoveContainer" containerID="f520260e9c3bbb1b63ec8f9522bd9fcc6cd69d7e6ab2a4af3b330077c6fc553f" Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.312921 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ffnlw"] Feb 17 13:23:59 crc kubenswrapper[4955]: W0217 13:23:59.321939 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfbdb6999_85bf_4bd2_b7f4_7b7c8566941f.slice/crio-593388dc8214c4cc4b502c47af3a68e45c32c7f61a07184ffd6a6daa832a6168 WatchSource:0}: Error finding container 593388dc8214c4cc4b502c47af3a68e45c32c7f61a07184ffd6a6daa832a6168: Status 404 returned error can't find the container with id 593388dc8214c4cc4b502c47af3a68e45c32c7f61a07184ffd6a6daa832a6168 Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.452597 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e56111b4-4b8a-4e07-bba9-68d34fd179fb-config-data\") pod \"e56111b4-4b8a-4e07-bba9-68d34fd179fb\" (UID: \"e56111b4-4b8a-4e07-bba9-68d34fd179fb\") " Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.457866 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e56111b4-4b8a-4e07-bba9-68d34fd179fb-config-data" (OuterVolumeSpecName: "config-data") pod "e56111b4-4b8a-4e07-bba9-68d34fd179fb" (UID: "e56111b4-4b8a-4e07-bba9-68d34fd179fb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.554759 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e56111b4-4b8a-4e07-bba9-68d34fd179fb-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.590697 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.602858 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.619320 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:23:59 crc kubenswrapper[4955]: E0217 13:23:59.620257 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e56111b4-4b8a-4e07-bba9-68d34fd179fb" containerName="ceilometer-central-agent" Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.620283 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="e56111b4-4b8a-4e07-bba9-68d34fd179fb" containerName="ceilometer-central-agent" Feb 17 13:23:59 crc kubenswrapper[4955]: E0217 13:23:59.620309 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e56111b4-4b8a-4e07-bba9-68d34fd179fb" containerName="sg-core" Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.620318 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="e56111b4-4b8a-4e07-bba9-68d34fd179fb" containerName="sg-core" Feb 17 13:23:59 crc kubenswrapper[4955]: E0217 13:23:59.620355 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e56111b4-4b8a-4e07-bba9-68d34fd179fb" containerName="proxy-httpd" Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.620363 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="e56111b4-4b8a-4e07-bba9-68d34fd179fb" containerName="proxy-httpd" Feb 17 13:23:59 crc kubenswrapper[4955]: E0217 13:23:59.620374 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e56111b4-4b8a-4e07-bba9-68d34fd179fb" containerName="ceilometer-notification-agent" Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.620381 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="e56111b4-4b8a-4e07-bba9-68d34fd179fb" containerName="ceilometer-notification-agent" Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.620768 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="e56111b4-4b8a-4e07-bba9-68d34fd179fb" containerName="ceilometer-central-agent" Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.620909 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="e56111b4-4b8a-4e07-bba9-68d34fd179fb" containerName="ceilometer-notification-agent" Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.620949 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="e56111b4-4b8a-4e07-bba9-68d34fd179fb" containerName="proxy-httpd" Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.620966 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="e56111b4-4b8a-4e07-bba9-68d34fd179fb" containerName="sg-core" Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.633522 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.637663 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.637880 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.657552 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f17c56b0-743a-4e97-9455-746034ea1df2-run-httpd\") pod \"ceilometer-0\" (UID: \"f17c56b0-743a-4e97-9455-746034ea1df2\") " pod="openstack/ceilometer-0" Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.657597 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f17c56b0-743a-4e97-9455-746034ea1df2-log-httpd\") pod \"ceilometer-0\" (UID: \"f17c56b0-743a-4e97-9455-746034ea1df2\") " pod="openstack/ceilometer-0" Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.657874 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f17c56b0-743a-4e97-9455-746034ea1df2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f17c56b0-743a-4e97-9455-746034ea1df2\") " pod="openstack/ceilometer-0" Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.657945 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f17c56b0-743a-4e97-9455-746034ea1df2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f17c56b0-743a-4e97-9455-746034ea1df2\") " pod="openstack/ceilometer-0" Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.657970 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f17c56b0-743a-4e97-9455-746034ea1df2-config-data\") pod \"ceilometer-0\" (UID: \"f17c56b0-743a-4e97-9455-746034ea1df2\") " pod="openstack/ceilometer-0" Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.659268 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f17c56b0-743a-4e97-9455-746034ea1df2-scripts\") pod \"ceilometer-0\" (UID: \"f17c56b0-743a-4e97-9455-746034ea1df2\") " pod="openstack/ceilometer-0" Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.659361 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pk9kj\" (UniqueName: \"kubernetes.io/projected/f17c56b0-743a-4e97-9455-746034ea1df2-kube-api-access-pk9kj\") pod \"ceilometer-0\" (UID: \"f17c56b0-743a-4e97-9455-746034ea1df2\") " pod="openstack/ceilometer-0" Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.671990 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.761530 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f17c56b0-743a-4e97-9455-746034ea1df2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f17c56b0-743a-4e97-9455-746034ea1df2\") " pod="openstack/ceilometer-0" Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.761585 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f17c56b0-743a-4e97-9455-746034ea1df2-config-data\") pod \"ceilometer-0\" (UID: \"f17c56b0-743a-4e97-9455-746034ea1df2\") " pod="openstack/ceilometer-0" Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.761626 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f17c56b0-743a-4e97-9455-746034ea1df2-scripts\") pod \"ceilometer-0\" (UID: \"f17c56b0-743a-4e97-9455-746034ea1df2\") " pod="openstack/ceilometer-0" Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.761644 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pk9kj\" (UniqueName: \"kubernetes.io/projected/f17c56b0-743a-4e97-9455-746034ea1df2-kube-api-access-pk9kj\") pod \"ceilometer-0\" (UID: \"f17c56b0-743a-4e97-9455-746034ea1df2\") " pod="openstack/ceilometer-0" Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.761683 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f17c56b0-743a-4e97-9455-746034ea1df2-run-httpd\") pod \"ceilometer-0\" (UID: \"f17c56b0-743a-4e97-9455-746034ea1df2\") " pod="openstack/ceilometer-0" Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.761700 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f17c56b0-743a-4e97-9455-746034ea1df2-log-httpd\") pod \"ceilometer-0\" (UID: \"f17c56b0-743a-4e97-9455-746034ea1df2\") " pod="openstack/ceilometer-0" Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.761764 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f17c56b0-743a-4e97-9455-746034ea1df2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f17c56b0-743a-4e97-9455-746034ea1df2\") " pod="openstack/ceilometer-0" Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.762372 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f17c56b0-743a-4e97-9455-746034ea1df2-run-httpd\") pod \"ceilometer-0\" (UID: \"f17c56b0-743a-4e97-9455-746034ea1df2\") " pod="openstack/ceilometer-0" Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.762514 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f17c56b0-743a-4e97-9455-746034ea1df2-log-httpd\") pod \"ceilometer-0\" (UID: \"f17c56b0-743a-4e97-9455-746034ea1df2\") " pod="openstack/ceilometer-0" Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.771007 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f17c56b0-743a-4e97-9455-746034ea1df2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f17c56b0-743a-4e97-9455-746034ea1df2\") " pod="openstack/ceilometer-0" Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.771155 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f17c56b0-743a-4e97-9455-746034ea1df2-scripts\") pod \"ceilometer-0\" (UID: \"f17c56b0-743a-4e97-9455-746034ea1df2\") " pod="openstack/ceilometer-0" Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.771173 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f17c56b0-743a-4e97-9455-746034ea1df2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f17c56b0-743a-4e97-9455-746034ea1df2\") " pod="openstack/ceilometer-0" Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.771400 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f17c56b0-743a-4e97-9455-746034ea1df2-config-data\") pod \"ceilometer-0\" (UID: \"f17c56b0-743a-4e97-9455-746034ea1df2\") " pod="openstack/ceilometer-0" Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.780031 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pk9kj\" (UniqueName: \"kubernetes.io/projected/f17c56b0-743a-4e97-9455-746034ea1df2-kube-api-access-pk9kj\") pod \"ceilometer-0\" (UID: \"f17c56b0-743a-4e97-9455-746034ea1df2\") " pod="openstack/ceilometer-0" Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.947220 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"539c0e46-aba7-4b5b-a176-3e9ac9089560","Type":"ContainerStarted","Data":"dd45e7727a3fe833ddc3300f027fba80dafd26ad8f6f525fb79c50d49039a77c"} Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.947269 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"539c0e46-aba7-4b5b-a176-3e9ac9089560","Type":"ContainerStarted","Data":"77825be6734fc7876a593a7f56c21828def3bee5b618c6ac0587e50f164f373d"} Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.949329 4955 generic.go:334] "Generic (PLEG): container finished" podID="20d6bdbe-1dd5-4790-add1-2ea3f3496315" containerID="a20011351ac1920632e1b4557ffb5d35fab17f1d712096d2aa96c1115c17a75a" exitCode=143 Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.949427 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-845ffbbf5d-ntk67" event={"ID":"20d6bdbe-1dd5-4790-add1-2ea3f3496315","Type":"ContainerDied","Data":"a20011351ac1920632e1b4557ffb5d35fab17f1d712096d2aa96c1115c17a75a"} Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.950599 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-ffnlw" event={"ID":"fbdb6999-85bf-4bd2-b7f4-7b7c8566941f","Type":"ContainerStarted","Data":"593388dc8214c4cc4b502c47af3a68e45c32c7f61a07184ffd6a6daa832a6168"} Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.956978 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 17 13:23:59 crc kubenswrapper[4955]: I0217 13:23:59.977345 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.97732441 podStartE2EDuration="4.97732441s" podCreationTimestamp="2026-02-17 13:23:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:23:59.966913213 +0000 UTC m=+1178.489642756" watchObservedRunningTime="2026-02-17 13:23:59.97732441 +0000 UTC m=+1178.500053973" Feb 17 13:24:00 crc kubenswrapper[4955]: I0217 13:24:00.240177 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e56111b4-4b8a-4e07-bba9-68d34fd179fb" path="/var/lib/kubelet/pods/e56111b4-4b8a-4e07-bba9-68d34fd179fb/volumes" Feb 17 13:24:00 crc kubenswrapper[4955]: I0217 13:24:00.405370 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:24:00 crc kubenswrapper[4955]: W0217 13:24:00.406285 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf17c56b0_743a_4e97_9455_746034ea1df2.slice/crio-7261edc80223418f10c9ee2a524b48e6101e2b77162c561674613b75a4cb43a8 WatchSource:0}: Error finding container 7261edc80223418f10c9ee2a524b48e6101e2b77162c561674613b75a4cb43a8: Status 404 returned error can't find the container with id 7261edc80223418f10c9ee2a524b48e6101e2b77162c561674613b75a4cb43a8 Feb 17 13:24:00 crc kubenswrapper[4955]: I0217 13:24:00.964245 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f17c56b0-743a-4e97-9455-746034ea1df2","Type":"ContainerStarted","Data":"7261edc80223418f10c9ee2a524b48e6101e2b77162c561674613b75a4cb43a8"} Feb 17 13:24:01 crc kubenswrapper[4955]: I0217 13:24:01.729525 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:24:01 crc kubenswrapper[4955]: I0217 13:24:01.979315 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f17c56b0-743a-4e97-9455-746034ea1df2","Type":"ContainerStarted","Data":"8be55cc8213bb18664b09a11fd1fe3bf8e06cdf2293d5aeb137df67e4a9a6b57"} Feb 17 13:24:02 crc kubenswrapper[4955]: I0217 13:24:02.503520 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-845ffbbf5d-ntk67" Feb 17 13:24:02 crc kubenswrapper[4955]: I0217 13:24:02.620613 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20d6bdbe-1dd5-4790-add1-2ea3f3496315-config-data\") pod \"20d6bdbe-1dd5-4790-add1-2ea3f3496315\" (UID: \"20d6bdbe-1dd5-4790-add1-2ea3f3496315\") " Feb 17 13:24:02 crc kubenswrapper[4955]: I0217 13:24:02.620702 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmctr\" (UniqueName: \"kubernetes.io/projected/20d6bdbe-1dd5-4790-add1-2ea3f3496315-kube-api-access-zmctr\") pod \"20d6bdbe-1dd5-4790-add1-2ea3f3496315\" (UID: \"20d6bdbe-1dd5-4790-add1-2ea3f3496315\") " Feb 17 13:24:02 crc kubenswrapper[4955]: I0217 13:24:02.620854 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20d6bdbe-1dd5-4790-add1-2ea3f3496315-logs\") pod \"20d6bdbe-1dd5-4790-add1-2ea3f3496315\" (UID: \"20d6bdbe-1dd5-4790-add1-2ea3f3496315\") " Feb 17 13:24:02 crc kubenswrapper[4955]: I0217 13:24:02.620890 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/20d6bdbe-1dd5-4790-add1-2ea3f3496315-public-tls-certs\") pod \"20d6bdbe-1dd5-4790-add1-2ea3f3496315\" (UID: \"20d6bdbe-1dd5-4790-add1-2ea3f3496315\") " Feb 17 13:24:02 crc kubenswrapper[4955]: I0217 13:24:02.620955 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20d6bdbe-1dd5-4790-add1-2ea3f3496315-scripts\") pod \"20d6bdbe-1dd5-4790-add1-2ea3f3496315\" (UID: \"20d6bdbe-1dd5-4790-add1-2ea3f3496315\") " Feb 17 13:24:02 crc kubenswrapper[4955]: I0217 13:24:02.621010 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20d6bdbe-1dd5-4790-add1-2ea3f3496315-combined-ca-bundle\") pod \"20d6bdbe-1dd5-4790-add1-2ea3f3496315\" (UID: \"20d6bdbe-1dd5-4790-add1-2ea3f3496315\") " Feb 17 13:24:02 crc kubenswrapper[4955]: I0217 13:24:02.621097 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/20d6bdbe-1dd5-4790-add1-2ea3f3496315-internal-tls-certs\") pod \"20d6bdbe-1dd5-4790-add1-2ea3f3496315\" (UID: \"20d6bdbe-1dd5-4790-add1-2ea3f3496315\") " Feb 17 13:24:02 crc kubenswrapper[4955]: I0217 13:24:02.621368 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20d6bdbe-1dd5-4790-add1-2ea3f3496315-logs" (OuterVolumeSpecName: "logs") pod "20d6bdbe-1dd5-4790-add1-2ea3f3496315" (UID: "20d6bdbe-1dd5-4790-add1-2ea3f3496315"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:24:02 crc kubenswrapper[4955]: I0217 13:24:02.621796 4955 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20d6bdbe-1dd5-4790-add1-2ea3f3496315-logs\") on node \"crc\" DevicePath \"\"" Feb 17 13:24:02 crc kubenswrapper[4955]: I0217 13:24:02.628332 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20d6bdbe-1dd5-4790-add1-2ea3f3496315-kube-api-access-zmctr" (OuterVolumeSpecName: "kube-api-access-zmctr") pod "20d6bdbe-1dd5-4790-add1-2ea3f3496315" (UID: "20d6bdbe-1dd5-4790-add1-2ea3f3496315"). InnerVolumeSpecName "kube-api-access-zmctr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:24:02 crc kubenswrapper[4955]: I0217 13:24:02.630966 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20d6bdbe-1dd5-4790-add1-2ea3f3496315-scripts" (OuterVolumeSpecName: "scripts") pod "20d6bdbe-1dd5-4790-add1-2ea3f3496315" (UID: "20d6bdbe-1dd5-4790-add1-2ea3f3496315"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:24:02 crc kubenswrapper[4955]: I0217 13:24:02.711373 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20d6bdbe-1dd5-4790-add1-2ea3f3496315-config-data" (OuterVolumeSpecName: "config-data") pod "20d6bdbe-1dd5-4790-add1-2ea3f3496315" (UID: "20d6bdbe-1dd5-4790-add1-2ea3f3496315"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:24:02 crc kubenswrapper[4955]: I0217 13:24:02.712075 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20d6bdbe-1dd5-4790-add1-2ea3f3496315-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "20d6bdbe-1dd5-4790-add1-2ea3f3496315" (UID: "20d6bdbe-1dd5-4790-add1-2ea3f3496315"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:24:02 crc kubenswrapper[4955]: I0217 13:24:02.723877 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20d6bdbe-1dd5-4790-add1-2ea3f3496315-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:24:02 crc kubenswrapper[4955]: I0217 13:24:02.723918 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmctr\" (UniqueName: \"kubernetes.io/projected/20d6bdbe-1dd5-4790-add1-2ea3f3496315-kube-api-access-zmctr\") on node \"crc\" DevicePath \"\"" Feb 17 13:24:02 crc kubenswrapper[4955]: I0217 13:24:02.723933 4955 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20d6bdbe-1dd5-4790-add1-2ea3f3496315-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:24:02 crc kubenswrapper[4955]: I0217 13:24:02.723946 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20d6bdbe-1dd5-4790-add1-2ea3f3496315-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:24:02 crc kubenswrapper[4955]: I0217 13:24:02.793143 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20d6bdbe-1dd5-4790-add1-2ea3f3496315-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "20d6bdbe-1dd5-4790-add1-2ea3f3496315" (UID: "20d6bdbe-1dd5-4790-add1-2ea3f3496315"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:24:02 crc kubenswrapper[4955]: I0217 13:24:02.826864 4955 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/20d6bdbe-1dd5-4790-add1-2ea3f3496315-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 17 13:24:02 crc kubenswrapper[4955]: I0217 13:24:02.828710 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20d6bdbe-1dd5-4790-add1-2ea3f3496315-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "20d6bdbe-1dd5-4790-add1-2ea3f3496315" (UID: "20d6bdbe-1dd5-4790-add1-2ea3f3496315"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:24:02 crc kubenswrapper[4955]: I0217 13:24:02.928503 4955 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/20d6bdbe-1dd5-4790-add1-2ea3f3496315-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 17 13:24:02 crc kubenswrapper[4955]: I0217 13:24:02.990817 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f17c56b0-743a-4e97-9455-746034ea1df2","Type":"ContainerStarted","Data":"7d11ab0a96376b29e0ca8fc1fd4bdd2b742428e7bc790d9317d46ffe72865c03"} Feb 17 13:24:02 crc kubenswrapper[4955]: I0217 13:24:02.990871 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f17c56b0-743a-4e97-9455-746034ea1df2","Type":"ContainerStarted","Data":"377f4f6678ae5109e754ce319972e7be2aff5f0bfa80dee682db08a0bc10a9f9"} Feb 17 13:24:02 crc kubenswrapper[4955]: I0217 13:24:02.992758 4955 generic.go:334] "Generic (PLEG): container finished" podID="20d6bdbe-1dd5-4790-add1-2ea3f3496315" containerID="f550116db1414f335a606a7e1ad151d2ee5761dfeb702452fed5a5ea730d9b78" exitCode=0 Feb 17 13:24:02 crc kubenswrapper[4955]: I0217 13:24:02.992801 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-845ffbbf5d-ntk67" event={"ID":"20d6bdbe-1dd5-4790-add1-2ea3f3496315","Type":"ContainerDied","Data":"f550116db1414f335a606a7e1ad151d2ee5761dfeb702452fed5a5ea730d9b78"} Feb 17 13:24:02 crc kubenswrapper[4955]: I0217 13:24:02.992817 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-845ffbbf5d-ntk67" event={"ID":"20d6bdbe-1dd5-4790-add1-2ea3f3496315","Type":"ContainerDied","Data":"fdf006e63d380ac6f9c5df66f616039d3417717224751ebcdf4e4435267586bd"} Feb 17 13:24:02 crc kubenswrapper[4955]: I0217 13:24:02.992833 4955 scope.go:117] "RemoveContainer" containerID="f550116db1414f335a606a7e1ad151d2ee5761dfeb702452fed5a5ea730d9b78" Feb 17 13:24:02 crc kubenswrapper[4955]: I0217 13:24:02.992962 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-845ffbbf5d-ntk67" Feb 17 13:24:03 crc kubenswrapper[4955]: I0217 13:24:03.029471 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-845ffbbf5d-ntk67"] Feb 17 13:24:03 crc kubenswrapper[4955]: I0217 13:24:03.038385 4955 scope.go:117] "RemoveContainer" containerID="a20011351ac1920632e1b4557ffb5d35fab17f1d712096d2aa96c1115c17a75a" Feb 17 13:24:03 crc kubenswrapper[4955]: I0217 13:24:03.043063 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-845ffbbf5d-ntk67"] Feb 17 13:24:03 crc kubenswrapper[4955]: I0217 13:24:03.089914 4955 scope.go:117] "RemoveContainer" containerID="f550116db1414f335a606a7e1ad151d2ee5761dfeb702452fed5a5ea730d9b78" Feb 17 13:24:03 crc kubenswrapper[4955]: E0217 13:24:03.090409 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f550116db1414f335a606a7e1ad151d2ee5761dfeb702452fed5a5ea730d9b78\": container with ID starting with f550116db1414f335a606a7e1ad151d2ee5761dfeb702452fed5a5ea730d9b78 not found: ID does not exist" containerID="f550116db1414f335a606a7e1ad151d2ee5761dfeb702452fed5a5ea730d9b78" Feb 17 13:24:03 crc kubenswrapper[4955]: I0217 13:24:03.090446 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f550116db1414f335a606a7e1ad151d2ee5761dfeb702452fed5a5ea730d9b78"} err="failed to get container status \"f550116db1414f335a606a7e1ad151d2ee5761dfeb702452fed5a5ea730d9b78\": rpc error: code = NotFound desc = could not find container \"f550116db1414f335a606a7e1ad151d2ee5761dfeb702452fed5a5ea730d9b78\": container with ID starting with f550116db1414f335a606a7e1ad151d2ee5761dfeb702452fed5a5ea730d9b78 not found: ID does not exist" Feb 17 13:24:03 crc kubenswrapper[4955]: I0217 13:24:03.090473 4955 scope.go:117] "RemoveContainer" containerID="a20011351ac1920632e1b4557ffb5d35fab17f1d712096d2aa96c1115c17a75a" Feb 17 13:24:03 crc kubenswrapper[4955]: E0217 13:24:03.090819 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a20011351ac1920632e1b4557ffb5d35fab17f1d712096d2aa96c1115c17a75a\": container with ID starting with a20011351ac1920632e1b4557ffb5d35fab17f1d712096d2aa96c1115c17a75a not found: ID does not exist" containerID="a20011351ac1920632e1b4557ffb5d35fab17f1d712096d2aa96c1115c17a75a" Feb 17 13:24:03 crc kubenswrapper[4955]: I0217 13:24:03.090873 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a20011351ac1920632e1b4557ffb5d35fab17f1d712096d2aa96c1115c17a75a"} err="failed to get container status \"a20011351ac1920632e1b4557ffb5d35fab17f1d712096d2aa96c1115c17a75a\": rpc error: code = NotFound desc = could not find container \"a20011351ac1920632e1b4557ffb5d35fab17f1d712096d2aa96c1115c17a75a\": container with ID starting with a20011351ac1920632e1b4557ffb5d35fab17f1d712096d2aa96c1115c17a75a not found: ID does not exist" Feb 17 13:24:04 crc kubenswrapper[4955]: I0217 13:24:04.127610 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 17 13:24:04 crc kubenswrapper[4955]: I0217 13:24:04.127976 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 17 13:24:04 crc kubenswrapper[4955]: I0217 13:24:04.160193 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 17 13:24:04 crc kubenswrapper[4955]: I0217 13:24:04.179617 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 17 13:24:04 crc kubenswrapper[4955]: I0217 13:24:04.238378 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20d6bdbe-1dd5-4790-add1-2ea3f3496315" path="/var/lib/kubelet/pods/20d6bdbe-1dd5-4790-add1-2ea3f3496315/volumes" Feb 17 13:24:05 crc kubenswrapper[4955]: I0217 13:24:05.013184 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 17 13:24:05 crc kubenswrapper[4955]: I0217 13:24:05.013216 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 17 13:24:06 crc kubenswrapper[4955]: I0217 13:24:06.252860 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 17 13:24:06 crc kubenswrapper[4955]: I0217 13:24:06.252973 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 17 13:24:06 crc kubenswrapper[4955]: I0217 13:24:06.280868 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 17 13:24:06 crc kubenswrapper[4955]: I0217 13:24:06.289051 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 17 13:24:07 crc kubenswrapper[4955]: I0217 13:24:07.030432 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 17 13:24:07 crc kubenswrapper[4955]: I0217 13:24:07.030487 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 17 13:24:07 crc kubenswrapper[4955]: I0217 13:24:07.120368 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 17 13:24:07 crc kubenswrapper[4955]: I0217 13:24:07.120804 4955 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 17 13:24:07 crc kubenswrapper[4955]: I0217 13:24:07.129318 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 17 13:24:09 crc kubenswrapper[4955]: I0217 13:24:09.046064 4955 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 17 13:24:09 crc kubenswrapper[4955]: I0217 13:24:09.046599 4955 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 17 13:24:09 crc kubenswrapper[4955]: I0217 13:24:09.086908 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 17 13:24:09 crc kubenswrapper[4955]: I0217 13:24:09.159484 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 17 13:24:10 crc kubenswrapper[4955]: I0217 13:24:10.131957 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f17c56b0-743a-4e97-9455-746034ea1df2","Type":"ContainerStarted","Data":"6d6774c0f57b0cbcdb3df2c3e8efd46b66acd753069d6d28503baa6ad9ab112e"} Feb 17 13:24:10 crc kubenswrapper[4955]: I0217 13:24:10.132118 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f17c56b0-743a-4e97-9455-746034ea1df2" containerName="ceilometer-central-agent" containerID="cri-o://8be55cc8213bb18664b09a11fd1fe3bf8e06cdf2293d5aeb137df67e4a9a6b57" gracePeriod=30 Feb 17 13:24:10 crc kubenswrapper[4955]: I0217 13:24:10.132205 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f17c56b0-743a-4e97-9455-746034ea1df2" containerName="proxy-httpd" containerID="cri-o://6d6774c0f57b0cbcdb3df2c3e8efd46b66acd753069d6d28503baa6ad9ab112e" gracePeriod=30 Feb 17 13:24:10 crc kubenswrapper[4955]: I0217 13:24:10.132238 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f17c56b0-743a-4e97-9455-746034ea1df2" containerName="sg-core" containerID="cri-o://7d11ab0a96376b29e0ca8fc1fd4bdd2b742428e7bc790d9317d46ffe72865c03" gracePeriod=30 Feb 17 13:24:10 crc kubenswrapper[4955]: I0217 13:24:10.132268 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f17c56b0-743a-4e97-9455-746034ea1df2" containerName="ceilometer-notification-agent" containerID="cri-o://377f4f6678ae5109e754ce319972e7be2aff5f0bfa80dee682db08a0bc10a9f9" gracePeriod=30 Feb 17 13:24:10 crc kubenswrapper[4955]: I0217 13:24:10.132445 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 17 13:24:10 crc kubenswrapper[4955]: I0217 13:24:10.134635 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-ffnlw" event={"ID":"fbdb6999-85bf-4bd2-b7f4-7b7c8566941f","Type":"ContainerStarted","Data":"73d6d40c5cbbe29be944855ed25368d36864681d3acca1c2904cc68409ee2622"} Feb 17 13:24:10 crc kubenswrapper[4955]: I0217 13:24:10.177219 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.69588274 podStartE2EDuration="11.177204656s" podCreationTimestamp="2026-02-17 13:23:59 +0000 UTC" firstStartedPulling="2026-02-17 13:24:00.408681104 +0000 UTC m=+1178.931410637" lastFinishedPulling="2026-02-17 13:24:08.89000301 +0000 UTC m=+1187.412732553" observedRunningTime="2026-02-17 13:24:10.164484163 +0000 UTC m=+1188.687213706" watchObservedRunningTime="2026-02-17 13:24:10.177204656 +0000 UTC m=+1188.699934199" Feb 17 13:24:10 crc kubenswrapper[4955]: I0217 13:24:10.187027 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-ffnlw" podStartSLOduration=2.621421502 podStartE2EDuration="12.187010015s" podCreationTimestamp="2026-02-17 13:23:58 +0000 UTC" firstStartedPulling="2026-02-17 13:23:59.324911801 +0000 UTC m=+1177.847641344" lastFinishedPulling="2026-02-17 13:24:08.890500314 +0000 UTC m=+1187.413229857" observedRunningTime="2026-02-17 13:24:10.182153247 +0000 UTC m=+1188.704882810" watchObservedRunningTime="2026-02-17 13:24:10.187010015 +0000 UTC m=+1188.709739558" Feb 17 13:24:10 crc kubenswrapper[4955]: I0217 13:24:10.920207 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.084722 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f17c56b0-743a-4e97-9455-746034ea1df2-combined-ca-bundle\") pod \"f17c56b0-743a-4e97-9455-746034ea1df2\" (UID: \"f17c56b0-743a-4e97-9455-746034ea1df2\") " Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.084769 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f17c56b0-743a-4e97-9455-746034ea1df2-config-data\") pod \"f17c56b0-743a-4e97-9455-746034ea1df2\" (UID: \"f17c56b0-743a-4e97-9455-746034ea1df2\") " Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.084843 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f17c56b0-743a-4e97-9455-746034ea1df2-sg-core-conf-yaml\") pod \"f17c56b0-743a-4e97-9455-746034ea1df2\" (UID: \"f17c56b0-743a-4e97-9455-746034ea1df2\") " Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.084882 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pk9kj\" (UniqueName: \"kubernetes.io/projected/f17c56b0-743a-4e97-9455-746034ea1df2-kube-api-access-pk9kj\") pod \"f17c56b0-743a-4e97-9455-746034ea1df2\" (UID: \"f17c56b0-743a-4e97-9455-746034ea1df2\") " Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.084929 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f17c56b0-743a-4e97-9455-746034ea1df2-run-httpd\") pod \"f17c56b0-743a-4e97-9455-746034ea1df2\" (UID: \"f17c56b0-743a-4e97-9455-746034ea1df2\") " Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.085008 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f17c56b0-743a-4e97-9455-746034ea1df2-log-httpd\") pod \"f17c56b0-743a-4e97-9455-746034ea1df2\" (UID: \"f17c56b0-743a-4e97-9455-746034ea1df2\") " Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.085032 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f17c56b0-743a-4e97-9455-746034ea1df2-scripts\") pod \"f17c56b0-743a-4e97-9455-746034ea1df2\" (UID: \"f17c56b0-743a-4e97-9455-746034ea1df2\") " Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.086976 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f17c56b0-743a-4e97-9455-746034ea1df2-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f17c56b0-743a-4e97-9455-746034ea1df2" (UID: "f17c56b0-743a-4e97-9455-746034ea1df2"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.087032 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f17c56b0-743a-4e97-9455-746034ea1df2-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f17c56b0-743a-4e97-9455-746034ea1df2" (UID: "f17c56b0-743a-4e97-9455-746034ea1df2"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.093899 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f17c56b0-743a-4e97-9455-746034ea1df2-scripts" (OuterVolumeSpecName: "scripts") pod "f17c56b0-743a-4e97-9455-746034ea1df2" (UID: "f17c56b0-743a-4e97-9455-746034ea1df2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.094305 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f17c56b0-743a-4e97-9455-746034ea1df2-kube-api-access-pk9kj" (OuterVolumeSpecName: "kube-api-access-pk9kj") pod "f17c56b0-743a-4e97-9455-746034ea1df2" (UID: "f17c56b0-743a-4e97-9455-746034ea1df2"). InnerVolumeSpecName "kube-api-access-pk9kj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.127563 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f17c56b0-743a-4e97-9455-746034ea1df2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f17c56b0-743a-4e97-9455-746034ea1df2" (UID: "f17c56b0-743a-4e97-9455-746034ea1df2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.143263 4955 generic.go:334] "Generic (PLEG): container finished" podID="f17c56b0-743a-4e97-9455-746034ea1df2" containerID="6d6774c0f57b0cbcdb3df2c3e8efd46b66acd753069d6d28503baa6ad9ab112e" exitCode=0 Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.143293 4955 generic.go:334] "Generic (PLEG): container finished" podID="f17c56b0-743a-4e97-9455-746034ea1df2" containerID="7d11ab0a96376b29e0ca8fc1fd4bdd2b742428e7bc790d9317d46ffe72865c03" exitCode=2 Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.143300 4955 generic.go:334] "Generic (PLEG): container finished" podID="f17c56b0-743a-4e97-9455-746034ea1df2" containerID="377f4f6678ae5109e754ce319972e7be2aff5f0bfa80dee682db08a0bc10a9f9" exitCode=0 Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.143307 4955 generic.go:334] "Generic (PLEG): container finished" podID="f17c56b0-743a-4e97-9455-746034ea1df2" containerID="8be55cc8213bb18664b09a11fd1fe3bf8e06cdf2293d5aeb137df67e4a9a6b57" exitCode=0 Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.144216 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.144639 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f17c56b0-743a-4e97-9455-746034ea1df2","Type":"ContainerDied","Data":"6d6774c0f57b0cbcdb3df2c3e8efd46b66acd753069d6d28503baa6ad9ab112e"} Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.144904 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f17c56b0-743a-4e97-9455-746034ea1df2","Type":"ContainerDied","Data":"7d11ab0a96376b29e0ca8fc1fd4bdd2b742428e7bc790d9317d46ffe72865c03"} Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.144928 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f17c56b0-743a-4e97-9455-746034ea1df2","Type":"ContainerDied","Data":"377f4f6678ae5109e754ce319972e7be2aff5f0bfa80dee682db08a0bc10a9f9"} Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.144937 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f17c56b0-743a-4e97-9455-746034ea1df2","Type":"ContainerDied","Data":"8be55cc8213bb18664b09a11fd1fe3bf8e06cdf2293d5aeb137df67e4a9a6b57"} Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.144945 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f17c56b0-743a-4e97-9455-746034ea1df2","Type":"ContainerDied","Data":"7261edc80223418f10c9ee2a524b48e6101e2b77162c561674613b75a4cb43a8"} Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.144961 4955 scope.go:117] "RemoveContainer" containerID="6d6774c0f57b0cbcdb3df2c3e8efd46b66acd753069d6d28503baa6ad9ab112e" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.161468 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f17c56b0-743a-4e97-9455-746034ea1df2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f17c56b0-743a-4e97-9455-746034ea1df2" (UID: "f17c56b0-743a-4e97-9455-746034ea1df2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.168926 4955 scope.go:117] "RemoveContainer" containerID="7d11ab0a96376b29e0ca8fc1fd4bdd2b742428e7bc790d9317d46ffe72865c03" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.187703 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f17c56b0-743a-4e97-9455-746034ea1df2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.187736 4955 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f17c56b0-743a-4e97-9455-746034ea1df2-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.187748 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pk9kj\" (UniqueName: \"kubernetes.io/projected/f17c56b0-743a-4e97-9455-746034ea1df2-kube-api-access-pk9kj\") on node \"crc\" DevicePath \"\"" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.187758 4955 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f17c56b0-743a-4e97-9455-746034ea1df2-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.187768 4955 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f17c56b0-743a-4e97-9455-746034ea1df2-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.187777 4955 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f17c56b0-743a-4e97-9455-746034ea1df2-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.198130 4955 scope.go:117] "RemoveContainer" containerID="377f4f6678ae5109e754ce319972e7be2aff5f0bfa80dee682db08a0bc10a9f9" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.204524 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f17c56b0-743a-4e97-9455-746034ea1df2-config-data" (OuterVolumeSpecName: "config-data") pod "f17c56b0-743a-4e97-9455-746034ea1df2" (UID: "f17c56b0-743a-4e97-9455-746034ea1df2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.221776 4955 scope.go:117] "RemoveContainer" containerID="8be55cc8213bb18664b09a11fd1fe3bf8e06cdf2293d5aeb137df67e4a9a6b57" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.243520 4955 scope.go:117] "RemoveContainer" containerID="6d6774c0f57b0cbcdb3df2c3e8efd46b66acd753069d6d28503baa6ad9ab112e" Feb 17 13:24:11 crc kubenswrapper[4955]: E0217 13:24:11.243920 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d6774c0f57b0cbcdb3df2c3e8efd46b66acd753069d6d28503baa6ad9ab112e\": container with ID starting with 6d6774c0f57b0cbcdb3df2c3e8efd46b66acd753069d6d28503baa6ad9ab112e not found: ID does not exist" containerID="6d6774c0f57b0cbcdb3df2c3e8efd46b66acd753069d6d28503baa6ad9ab112e" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.243974 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d6774c0f57b0cbcdb3df2c3e8efd46b66acd753069d6d28503baa6ad9ab112e"} err="failed to get container status \"6d6774c0f57b0cbcdb3df2c3e8efd46b66acd753069d6d28503baa6ad9ab112e\": rpc error: code = NotFound desc = could not find container \"6d6774c0f57b0cbcdb3df2c3e8efd46b66acd753069d6d28503baa6ad9ab112e\": container with ID starting with 6d6774c0f57b0cbcdb3df2c3e8efd46b66acd753069d6d28503baa6ad9ab112e not found: ID does not exist" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.243998 4955 scope.go:117] "RemoveContainer" containerID="7d11ab0a96376b29e0ca8fc1fd4bdd2b742428e7bc790d9317d46ffe72865c03" Feb 17 13:24:11 crc kubenswrapper[4955]: E0217 13:24:11.244367 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d11ab0a96376b29e0ca8fc1fd4bdd2b742428e7bc790d9317d46ffe72865c03\": container with ID starting with 7d11ab0a96376b29e0ca8fc1fd4bdd2b742428e7bc790d9317d46ffe72865c03 not found: ID does not exist" containerID="7d11ab0a96376b29e0ca8fc1fd4bdd2b742428e7bc790d9317d46ffe72865c03" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.244388 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d11ab0a96376b29e0ca8fc1fd4bdd2b742428e7bc790d9317d46ffe72865c03"} err="failed to get container status \"7d11ab0a96376b29e0ca8fc1fd4bdd2b742428e7bc790d9317d46ffe72865c03\": rpc error: code = NotFound desc = could not find container \"7d11ab0a96376b29e0ca8fc1fd4bdd2b742428e7bc790d9317d46ffe72865c03\": container with ID starting with 7d11ab0a96376b29e0ca8fc1fd4bdd2b742428e7bc790d9317d46ffe72865c03 not found: ID does not exist" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.244539 4955 scope.go:117] "RemoveContainer" containerID="377f4f6678ae5109e754ce319972e7be2aff5f0bfa80dee682db08a0bc10a9f9" Feb 17 13:24:11 crc kubenswrapper[4955]: E0217 13:24:11.244939 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"377f4f6678ae5109e754ce319972e7be2aff5f0bfa80dee682db08a0bc10a9f9\": container with ID starting with 377f4f6678ae5109e754ce319972e7be2aff5f0bfa80dee682db08a0bc10a9f9 not found: ID does not exist" containerID="377f4f6678ae5109e754ce319972e7be2aff5f0bfa80dee682db08a0bc10a9f9" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.244961 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"377f4f6678ae5109e754ce319972e7be2aff5f0bfa80dee682db08a0bc10a9f9"} err="failed to get container status \"377f4f6678ae5109e754ce319972e7be2aff5f0bfa80dee682db08a0bc10a9f9\": rpc error: code = NotFound desc = could not find container \"377f4f6678ae5109e754ce319972e7be2aff5f0bfa80dee682db08a0bc10a9f9\": container with ID starting with 377f4f6678ae5109e754ce319972e7be2aff5f0bfa80dee682db08a0bc10a9f9 not found: ID does not exist" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.244981 4955 scope.go:117] "RemoveContainer" containerID="8be55cc8213bb18664b09a11fd1fe3bf8e06cdf2293d5aeb137df67e4a9a6b57" Feb 17 13:24:11 crc kubenswrapper[4955]: E0217 13:24:11.245421 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8be55cc8213bb18664b09a11fd1fe3bf8e06cdf2293d5aeb137df67e4a9a6b57\": container with ID starting with 8be55cc8213bb18664b09a11fd1fe3bf8e06cdf2293d5aeb137df67e4a9a6b57 not found: ID does not exist" containerID="8be55cc8213bb18664b09a11fd1fe3bf8e06cdf2293d5aeb137df67e4a9a6b57" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.245449 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8be55cc8213bb18664b09a11fd1fe3bf8e06cdf2293d5aeb137df67e4a9a6b57"} err="failed to get container status \"8be55cc8213bb18664b09a11fd1fe3bf8e06cdf2293d5aeb137df67e4a9a6b57\": rpc error: code = NotFound desc = could not find container \"8be55cc8213bb18664b09a11fd1fe3bf8e06cdf2293d5aeb137df67e4a9a6b57\": container with ID starting with 8be55cc8213bb18664b09a11fd1fe3bf8e06cdf2293d5aeb137df67e4a9a6b57 not found: ID does not exist" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.245464 4955 scope.go:117] "RemoveContainer" containerID="6d6774c0f57b0cbcdb3df2c3e8efd46b66acd753069d6d28503baa6ad9ab112e" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.245858 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d6774c0f57b0cbcdb3df2c3e8efd46b66acd753069d6d28503baa6ad9ab112e"} err="failed to get container status \"6d6774c0f57b0cbcdb3df2c3e8efd46b66acd753069d6d28503baa6ad9ab112e\": rpc error: code = NotFound desc = could not find container \"6d6774c0f57b0cbcdb3df2c3e8efd46b66acd753069d6d28503baa6ad9ab112e\": container with ID starting with 6d6774c0f57b0cbcdb3df2c3e8efd46b66acd753069d6d28503baa6ad9ab112e not found: ID does not exist" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.245884 4955 scope.go:117] "RemoveContainer" containerID="7d11ab0a96376b29e0ca8fc1fd4bdd2b742428e7bc790d9317d46ffe72865c03" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.246269 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d11ab0a96376b29e0ca8fc1fd4bdd2b742428e7bc790d9317d46ffe72865c03"} err="failed to get container status \"7d11ab0a96376b29e0ca8fc1fd4bdd2b742428e7bc790d9317d46ffe72865c03\": rpc error: code = NotFound desc = could not find container \"7d11ab0a96376b29e0ca8fc1fd4bdd2b742428e7bc790d9317d46ffe72865c03\": container with ID starting with 7d11ab0a96376b29e0ca8fc1fd4bdd2b742428e7bc790d9317d46ffe72865c03 not found: ID does not exist" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.246285 4955 scope.go:117] "RemoveContainer" containerID="377f4f6678ae5109e754ce319972e7be2aff5f0bfa80dee682db08a0bc10a9f9" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.246511 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"377f4f6678ae5109e754ce319972e7be2aff5f0bfa80dee682db08a0bc10a9f9"} err="failed to get container status \"377f4f6678ae5109e754ce319972e7be2aff5f0bfa80dee682db08a0bc10a9f9\": rpc error: code = NotFound desc = could not find container \"377f4f6678ae5109e754ce319972e7be2aff5f0bfa80dee682db08a0bc10a9f9\": container with ID starting with 377f4f6678ae5109e754ce319972e7be2aff5f0bfa80dee682db08a0bc10a9f9 not found: ID does not exist" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.246529 4955 scope.go:117] "RemoveContainer" containerID="8be55cc8213bb18664b09a11fd1fe3bf8e06cdf2293d5aeb137df67e4a9a6b57" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.246726 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8be55cc8213bb18664b09a11fd1fe3bf8e06cdf2293d5aeb137df67e4a9a6b57"} err="failed to get container status \"8be55cc8213bb18664b09a11fd1fe3bf8e06cdf2293d5aeb137df67e4a9a6b57\": rpc error: code = NotFound desc = could not find container \"8be55cc8213bb18664b09a11fd1fe3bf8e06cdf2293d5aeb137df67e4a9a6b57\": container with ID starting with 8be55cc8213bb18664b09a11fd1fe3bf8e06cdf2293d5aeb137df67e4a9a6b57 not found: ID does not exist" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.246740 4955 scope.go:117] "RemoveContainer" containerID="6d6774c0f57b0cbcdb3df2c3e8efd46b66acd753069d6d28503baa6ad9ab112e" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.246959 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d6774c0f57b0cbcdb3df2c3e8efd46b66acd753069d6d28503baa6ad9ab112e"} err="failed to get container status \"6d6774c0f57b0cbcdb3df2c3e8efd46b66acd753069d6d28503baa6ad9ab112e\": rpc error: code = NotFound desc = could not find container \"6d6774c0f57b0cbcdb3df2c3e8efd46b66acd753069d6d28503baa6ad9ab112e\": container with ID starting with 6d6774c0f57b0cbcdb3df2c3e8efd46b66acd753069d6d28503baa6ad9ab112e not found: ID does not exist" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.246979 4955 scope.go:117] "RemoveContainer" containerID="7d11ab0a96376b29e0ca8fc1fd4bdd2b742428e7bc790d9317d46ffe72865c03" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.247380 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d11ab0a96376b29e0ca8fc1fd4bdd2b742428e7bc790d9317d46ffe72865c03"} err="failed to get container status \"7d11ab0a96376b29e0ca8fc1fd4bdd2b742428e7bc790d9317d46ffe72865c03\": rpc error: code = NotFound desc = could not find container \"7d11ab0a96376b29e0ca8fc1fd4bdd2b742428e7bc790d9317d46ffe72865c03\": container with ID starting with 7d11ab0a96376b29e0ca8fc1fd4bdd2b742428e7bc790d9317d46ffe72865c03 not found: ID does not exist" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.247405 4955 scope.go:117] "RemoveContainer" containerID="377f4f6678ae5109e754ce319972e7be2aff5f0bfa80dee682db08a0bc10a9f9" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.247601 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"377f4f6678ae5109e754ce319972e7be2aff5f0bfa80dee682db08a0bc10a9f9"} err="failed to get container status \"377f4f6678ae5109e754ce319972e7be2aff5f0bfa80dee682db08a0bc10a9f9\": rpc error: code = NotFound desc = could not find container \"377f4f6678ae5109e754ce319972e7be2aff5f0bfa80dee682db08a0bc10a9f9\": container with ID starting with 377f4f6678ae5109e754ce319972e7be2aff5f0bfa80dee682db08a0bc10a9f9 not found: ID does not exist" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.247616 4955 scope.go:117] "RemoveContainer" containerID="8be55cc8213bb18664b09a11fd1fe3bf8e06cdf2293d5aeb137df67e4a9a6b57" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.248347 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8be55cc8213bb18664b09a11fd1fe3bf8e06cdf2293d5aeb137df67e4a9a6b57"} err="failed to get container status \"8be55cc8213bb18664b09a11fd1fe3bf8e06cdf2293d5aeb137df67e4a9a6b57\": rpc error: code = NotFound desc = could not find container \"8be55cc8213bb18664b09a11fd1fe3bf8e06cdf2293d5aeb137df67e4a9a6b57\": container with ID starting with 8be55cc8213bb18664b09a11fd1fe3bf8e06cdf2293d5aeb137df67e4a9a6b57 not found: ID does not exist" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.248377 4955 scope.go:117] "RemoveContainer" containerID="6d6774c0f57b0cbcdb3df2c3e8efd46b66acd753069d6d28503baa6ad9ab112e" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.248714 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d6774c0f57b0cbcdb3df2c3e8efd46b66acd753069d6d28503baa6ad9ab112e"} err="failed to get container status \"6d6774c0f57b0cbcdb3df2c3e8efd46b66acd753069d6d28503baa6ad9ab112e\": rpc error: code = NotFound desc = could not find container \"6d6774c0f57b0cbcdb3df2c3e8efd46b66acd753069d6d28503baa6ad9ab112e\": container with ID starting with 6d6774c0f57b0cbcdb3df2c3e8efd46b66acd753069d6d28503baa6ad9ab112e not found: ID does not exist" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.248735 4955 scope.go:117] "RemoveContainer" containerID="7d11ab0a96376b29e0ca8fc1fd4bdd2b742428e7bc790d9317d46ffe72865c03" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.249184 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d11ab0a96376b29e0ca8fc1fd4bdd2b742428e7bc790d9317d46ffe72865c03"} err="failed to get container status \"7d11ab0a96376b29e0ca8fc1fd4bdd2b742428e7bc790d9317d46ffe72865c03\": rpc error: code = NotFound desc = could not find container \"7d11ab0a96376b29e0ca8fc1fd4bdd2b742428e7bc790d9317d46ffe72865c03\": container with ID starting with 7d11ab0a96376b29e0ca8fc1fd4bdd2b742428e7bc790d9317d46ffe72865c03 not found: ID does not exist" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.249205 4955 scope.go:117] "RemoveContainer" containerID="377f4f6678ae5109e754ce319972e7be2aff5f0bfa80dee682db08a0bc10a9f9" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.249706 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"377f4f6678ae5109e754ce319972e7be2aff5f0bfa80dee682db08a0bc10a9f9"} err="failed to get container status \"377f4f6678ae5109e754ce319972e7be2aff5f0bfa80dee682db08a0bc10a9f9\": rpc error: code = NotFound desc = could not find container \"377f4f6678ae5109e754ce319972e7be2aff5f0bfa80dee682db08a0bc10a9f9\": container with ID starting with 377f4f6678ae5109e754ce319972e7be2aff5f0bfa80dee682db08a0bc10a9f9 not found: ID does not exist" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.249724 4955 scope.go:117] "RemoveContainer" containerID="8be55cc8213bb18664b09a11fd1fe3bf8e06cdf2293d5aeb137df67e4a9a6b57" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.250151 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8be55cc8213bb18664b09a11fd1fe3bf8e06cdf2293d5aeb137df67e4a9a6b57"} err="failed to get container status \"8be55cc8213bb18664b09a11fd1fe3bf8e06cdf2293d5aeb137df67e4a9a6b57\": rpc error: code = NotFound desc = could not find container \"8be55cc8213bb18664b09a11fd1fe3bf8e06cdf2293d5aeb137df67e4a9a6b57\": container with ID starting with 8be55cc8213bb18664b09a11fd1fe3bf8e06cdf2293d5aeb137df67e4a9a6b57 not found: ID does not exist" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.294346 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f17c56b0-743a-4e97-9455-746034ea1df2-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.477664 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.486504 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.506522 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:24:11 crc kubenswrapper[4955]: E0217 13:24:11.506935 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20d6bdbe-1dd5-4790-add1-2ea3f3496315" containerName="placement-api" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.506959 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="20d6bdbe-1dd5-4790-add1-2ea3f3496315" containerName="placement-api" Feb 17 13:24:11 crc kubenswrapper[4955]: E0217 13:24:11.506973 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20d6bdbe-1dd5-4790-add1-2ea3f3496315" containerName="placement-log" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.506982 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="20d6bdbe-1dd5-4790-add1-2ea3f3496315" containerName="placement-log" Feb 17 13:24:11 crc kubenswrapper[4955]: E0217 13:24:11.506999 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f17c56b0-743a-4e97-9455-746034ea1df2" containerName="ceilometer-central-agent" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.507007 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="f17c56b0-743a-4e97-9455-746034ea1df2" containerName="ceilometer-central-agent" Feb 17 13:24:11 crc kubenswrapper[4955]: E0217 13:24:11.507032 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f17c56b0-743a-4e97-9455-746034ea1df2" containerName="proxy-httpd" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.507042 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="f17c56b0-743a-4e97-9455-746034ea1df2" containerName="proxy-httpd" Feb 17 13:24:11 crc kubenswrapper[4955]: E0217 13:24:11.507058 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f17c56b0-743a-4e97-9455-746034ea1df2" containerName="ceilometer-notification-agent" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.507070 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="f17c56b0-743a-4e97-9455-746034ea1df2" containerName="ceilometer-notification-agent" Feb 17 13:24:11 crc kubenswrapper[4955]: E0217 13:24:11.507088 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f17c56b0-743a-4e97-9455-746034ea1df2" containerName="sg-core" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.507095 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="f17c56b0-743a-4e97-9455-746034ea1df2" containerName="sg-core" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.507323 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="f17c56b0-743a-4e97-9455-746034ea1df2" containerName="proxy-httpd" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.507346 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="f17c56b0-743a-4e97-9455-746034ea1df2" containerName="ceilometer-central-agent" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.507356 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="20d6bdbe-1dd5-4790-add1-2ea3f3496315" containerName="placement-log" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.507383 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="20d6bdbe-1dd5-4790-add1-2ea3f3496315" containerName="placement-api" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.507396 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="f17c56b0-743a-4e97-9455-746034ea1df2" containerName="ceilometer-notification-agent" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.507409 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="f17c56b0-743a-4e97-9455-746034ea1df2" containerName="sg-core" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.509112 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.511776 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.514933 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.523716 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.592116 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:24:11 crc kubenswrapper[4955]: E0217 13:24:11.592909 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle config-data kube-api-access-nlqkc log-httpd run-httpd scripts sg-core-conf-yaml], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/ceilometer-0" podUID="dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.599887 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff\") " pod="openstack/ceilometer-0" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.599978 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff-run-httpd\") pod \"ceilometer-0\" (UID: \"dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff\") " pod="openstack/ceilometer-0" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.600037 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff\") " pod="openstack/ceilometer-0" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.600073 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff-scripts\") pod \"ceilometer-0\" (UID: \"dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff\") " pod="openstack/ceilometer-0" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.600096 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff-log-httpd\") pod \"ceilometer-0\" (UID: \"dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff\") " pod="openstack/ceilometer-0" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.600139 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlqkc\" (UniqueName: \"kubernetes.io/projected/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff-kube-api-access-nlqkc\") pod \"ceilometer-0\" (UID: \"dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff\") " pod="openstack/ceilometer-0" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.600161 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff-config-data\") pod \"ceilometer-0\" (UID: \"dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff\") " pod="openstack/ceilometer-0" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.701675 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff\") " pod="openstack/ceilometer-0" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.701759 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff-run-httpd\") pod \"ceilometer-0\" (UID: \"dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff\") " pod="openstack/ceilometer-0" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.701866 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff\") " pod="openstack/ceilometer-0" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.701892 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff-scripts\") pod \"ceilometer-0\" (UID: \"dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff\") " pod="openstack/ceilometer-0" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.701912 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff-log-httpd\") pod \"ceilometer-0\" (UID: \"dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff\") " pod="openstack/ceilometer-0" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.701944 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlqkc\" (UniqueName: \"kubernetes.io/projected/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff-kube-api-access-nlqkc\") pod \"ceilometer-0\" (UID: \"dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff\") " pod="openstack/ceilometer-0" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.701963 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff-config-data\") pod \"ceilometer-0\" (UID: \"dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff\") " pod="openstack/ceilometer-0" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.702476 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff-log-httpd\") pod \"ceilometer-0\" (UID: \"dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff\") " pod="openstack/ceilometer-0" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.702645 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff-run-httpd\") pod \"ceilometer-0\" (UID: \"dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff\") " pod="openstack/ceilometer-0" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.705978 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff-scripts\") pod \"ceilometer-0\" (UID: \"dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff\") " pod="openstack/ceilometer-0" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.706157 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff\") " pod="openstack/ceilometer-0" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.707441 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff\") " pod="openstack/ceilometer-0" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.712091 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff-config-data\") pod \"ceilometer-0\" (UID: \"dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff\") " pod="openstack/ceilometer-0" Feb 17 13:24:11 crc kubenswrapper[4955]: I0217 13:24:11.719048 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlqkc\" (UniqueName: \"kubernetes.io/projected/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff-kube-api-access-nlqkc\") pod \"ceilometer-0\" (UID: \"dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff\") " pod="openstack/ceilometer-0" Feb 17 13:24:12 crc kubenswrapper[4955]: I0217 13:24:12.153412 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 17 13:24:12 crc kubenswrapper[4955]: I0217 13:24:12.165983 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 17 13:24:12 crc kubenswrapper[4955]: I0217 13:24:12.245192 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f17c56b0-743a-4e97-9455-746034ea1df2" path="/var/lib/kubelet/pods/f17c56b0-743a-4e97-9455-746034ea1df2/volumes" Feb 17 13:24:12 crc kubenswrapper[4955]: I0217 13:24:12.311893 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff-log-httpd\") pod \"dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff\" (UID: \"dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff\") " Feb 17 13:24:12 crc kubenswrapper[4955]: I0217 13:24:12.311962 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff-config-data\") pod \"dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff\" (UID: \"dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff\") " Feb 17 13:24:12 crc kubenswrapper[4955]: I0217 13:24:12.312031 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff-combined-ca-bundle\") pod \"dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff\" (UID: \"dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff\") " Feb 17 13:24:12 crc kubenswrapper[4955]: I0217 13:24:12.312101 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff-run-httpd\") pod \"dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff\" (UID: \"dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff\") " Feb 17 13:24:12 crc kubenswrapper[4955]: I0217 13:24:12.312164 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff-sg-core-conf-yaml\") pod \"dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff\" (UID: \"dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff\") " Feb 17 13:24:12 crc kubenswrapper[4955]: I0217 13:24:12.312184 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nlqkc\" (UniqueName: \"kubernetes.io/projected/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff-kube-api-access-nlqkc\") pod \"dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff\" (UID: \"dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff\") " Feb 17 13:24:12 crc kubenswrapper[4955]: I0217 13:24:12.312256 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff-scripts\") pod \"dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff\" (UID: \"dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff\") " Feb 17 13:24:12 crc kubenswrapper[4955]: I0217 13:24:12.312773 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff" (UID: "dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:24:12 crc kubenswrapper[4955]: I0217 13:24:12.312726 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff" (UID: "dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:24:12 crc kubenswrapper[4955]: I0217 13:24:12.317437 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff-config-data" (OuterVolumeSpecName: "config-data") pod "dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff" (UID: "dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:24:12 crc kubenswrapper[4955]: I0217 13:24:12.337315 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff" (UID: "dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:24:12 crc kubenswrapper[4955]: I0217 13:24:12.337607 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff-kube-api-access-nlqkc" (OuterVolumeSpecName: "kube-api-access-nlqkc") pod "dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff" (UID: "dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff"). InnerVolumeSpecName "kube-api-access-nlqkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:24:12 crc kubenswrapper[4955]: I0217 13:24:12.337630 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff-scripts" (OuterVolumeSpecName: "scripts") pod "dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff" (UID: "dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:24:12 crc kubenswrapper[4955]: I0217 13:24:12.337839 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff" (UID: "dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:24:12 crc kubenswrapper[4955]: I0217 13:24:12.415069 4955 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:24:12 crc kubenswrapper[4955]: I0217 13:24:12.415110 4955 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 17 13:24:12 crc kubenswrapper[4955]: I0217 13:24:12.415122 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:24:12 crc kubenswrapper[4955]: I0217 13:24:12.415131 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:24:12 crc kubenswrapper[4955]: I0217 13:24:12.415146 4955 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 17 13:24:12 crc kubenswrapper[4955]: I0217 13:24:12.415157 4955 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 17 13:24:12 crc kubenswrapper[4955]: I0217 13:24:12.415167 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nlqkc\" (UniqueName: \"kubernetes.io/projected/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff-kube-api-access-nlqkc\") on node \"crc\" DevicePath \"\"" Feb 17 13:24:13 crc kubenswrapper[4955]: I0217 13:24:13.160678 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 17 13:24:13 crc kubenswrapper[4955]: I0217 13:24:13.217129 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:24:13 crc kubenswrapper[4955]: I0217 13:24:13.236698 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:24:13 crc kubenswrapper[4955]: I0217 13:24:13.249378 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:24:13 crc kubenswrapper[4955]: I0217 13:24:13.253896 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 17 13:24:13 crc kubenswrapper[4955]: I0217 13:24:13.258382 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 17 13:24:13 crc kubenswrapper[4955]: I0217 13:24:13.258913 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 17 13:24:13 crc kubenswrapper[4955]: I0217 13:24:13.262313 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:24:13 crc kubenswrapper[4955]: I0217 13:24:13.433309 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f771ce9f-91d4-47c8-89fc-25758e4f907c-config-data\") pod \"ceilometer-0\" (UID: \"f771ce9f-91d4-47c8-89fc-25758e4f907c\") " pod="openstack/ceilometer-0" Feb 17 13:24:13 crc kubenswrapper[4955]: I0217 13:24:13.433513 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f771ce9f-91d4-47c8-89fc-25758e4f907c-scripts\") pod \"ceilometer-0\" (UID: \"f771ce9f-91d4-47c8-89fc-25758e4f907c\") " pod="openstack/ceilometer-0" Feb 17 13:24:13 crc kubenswrapper[4955]: I0217 13:24:13.433715 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f771ce9f-91d4-47c8-89fc-25758e4f907c-log-httpd\") pod \"ceilometer-0\" (UID: \"f771ce9f-91d4-47c8-89fc-25758e4f907c\") " pod="openstack/ceilometer-0" Feb 17 13:24:13 crc kubenswrapper[4955]: I0217 13:24:13.433755 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f771ce9f-91d4-47c8-89fc-25758e4f907c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f771ce9f-91d4-47c8-89fc-25758e4f907c\") " pod="openstack/ceilometer-0" Feb 17 13:24:13 crc kubenswrapper[4955]: I0217 13:24:13.433868 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwbg8\" (UniqueName: \"kubernetes.io/projected/f771ce9f-91d4-47c8-89fc-25758e4f907c-kube-api-access-vwbg8\") pod \"ceilometer-0\" (UID: \"f771ce9f-91d4-47c8-89fc-25758e4f907c\") " pod="openstack/ceilometer-0" Feb 17 13:24:13 crc kubenswrapper[4955]: I0217 13:24:13.433903 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f771ce9f-91d4-47c8-89fc-25758e4f907c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f771ce9f-91d4-47c8-89fc-25758e4f907c\") " pod="openstack/ceilometer-0" Feb 17 13:24:13 crc kubenswrapper[4955]: I0217 13:24:13.433926 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f771ce9f-91d4-47c8-89fc-25758e4f907c-run-httpd\") pod \"ceilometer-0\" (UID: \"f771ce9f-91d4-47c8-89fc-25758e4f907c\") " pod="openstack/ceilometer-0" Feb 17 13:24:13 crc kubenswrapper[4955]: I0217 13:24:13.535310 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f771ce9f-91d4-47c8-89fc-25758e4f907c-config-data\") pod \"ceilometer-0\" (UID: \"f771ce9f-91d4-47c8-89fc-25758e4f907c\") " pod="openstack/ceilometer-0" Feb 17 13:24:13 crc kubenswrapper[4955]: I0217 13:24:13.535451 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f771ce9f-91d4-47c8-89fc-25758e4f907c-scripts\") pod \"ceilometer-0\" (UID: \"f771ce9f-91d4-47c8-89fc-25758e4f907c\") " pod="openstack/ceilometer-0" Feb 17 13:24:13 crc kubenswrapper[4955]: I0217 13:24:13.535529 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f771ce9f-91d4-47c8-89fc-25758e4f907c-log-httpd\") pod \"ceilometer-0\" (UID: \"f771ce9f-91d4-47c8-89fc-25758e4f907c\") " pod="openstack/ceilometer-0" Feb 17 13:24:13 crc kubenswrapper[4955]: I0217 13:24:13.535558 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f771ce9f-91d4-47c8-89fc-25758e4f907c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f771ce9f-91d4-47c8-89fc-25758e4f907c\") " pod="openstack/ceilometer-0" Feb 17 13:24:13 crc kubenswrapper[4955]: I0217 13:24:13.535623 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwbg8\" (UniqueName: \"kubernetes.io/projected/f771ce9f-91d4-47c8-89fc-25758e4f907c-kube-api-access-vwbg8\") pod \"ceilometer-0\" (UID: \"f771ce9f-91d4-47c8-89fc-25758e4f907c\") " pod="openstack/ceilometer-0" Feb 17 13:24:13 crc kubenswrapper[4955]: I0217 13:24:13.535647 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f771ce9f-91d4-47c8-89fc-25758e4f907c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f771ce9f-91d4-47c8-89fc-25758e4f907c\") " pod="openstack/ceilometer-0" Feb 17 13:24:13 crc kubenswrapper[4955]: I0217 13:24:13.535670 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f771ce9f-91d4-47c8-89fc-25758e4f907c-run-httpd\") pod \"ceilometer-0\" (UID: \"f771ce9f-91d4-47c8-89fc-25758e4f907c\") " pod="openstack/ceilometer-0" Feb 17 13:24:13 crc kubenswrapper[4955]: I0217 13:24:13.536955 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f771ce9f-91d4-47c8-89fc-25758e4f907c-run-httpd\") pod \"ceilometer-0\" (UID: \"f771ce9f-91d4-47c8-89fc-25758e4f907c\") " pod="openstack/ceilometer-0" Feb 17 13:24:13 crc kubenswrapper[4955]: I0217 13:24:13.537143 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f771ce9f-91d4-47c8-89fc-25758e4f907c-log-httpd\") pod \"ceilometer-0\" (UID: \"f771ce9f-91d4-47c8-89fc-25758e4f907c\") " pod="openstack/ceilometer-0" Feb 17 13:24:13 crc kubenswrapper[4955]: I0217 13:24:13.541517 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f771ce9f-91d4-47c8-89fc-25758e4f907c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f771ce9f-91d4-47c8-89fc-25758e4f907c\") " pod="openstack/ceilometer-0" Feb 17 13:24:13 crc kubenswrapper[4955]: I0217 13:24:13.542884 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f771ce9f-91d4-47c8-89fc-25758e4f907c-config-data\") pod \"ceilometer-0\" (UID: \"f771ce9f-91d4-47c8-89fc-25758e4f907c\") " pod="openstack/ceilometer-0" Feb 17 13:24:13 crc kubenswrapper[4955]: I0217 13:24:13.547907 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f771ce9f-91d4-47c8-89fc-25758e4f907c-scripts\") pod \"ceilometer-0\" (UID: \"f771ce9f-91d4-47c8-89fc-25758e4f907c\") " pod="openstack/ceilometer-0" Feb 17 13:24:13 crc kubenswrapper[4955]: I0217 13:24:13.551303 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f771ce9f-91d4-47c8-89fc-25758e4f907c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f771ce9f-91d4-47c8-89fc-25758e4f907c\") " pod="openstack/ceilometer-0" Feb 17 13:24:13 crc kubenswrapper[4955]: I0217 13:24:13.563185 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwbg8\" (UniqueName: \"kubernetes.io/projected/f771ce9f-91d4-47c8-89fc-25758e4f907c-kube-api-access-vwbg8\") pod \"ceilometer-0\" (UID: \"f771ce9f-91d4-47c8-89fc-25758e4f907c\") " pod="openstack/ceilometer-0" Feb 17 13:24:13 crc kubenswrapper[4955]: I0217 13:24:13.578123 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 17 13:24:14 crc kubenswrapper[4955]: W0217 13:24:14.110195 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf771ce9f_91d4_47c8_89fc_25758e4f907c.slice/crio-bc82c442011807bbf16e641698169e943f13862b5af9857f6e60411a83d38c84 WatchSource:0}: Error finding container bc82c442011807bbf16e641698169e943f13862b5af9857f6e60411a83d38c84: Status 404 returned error can't find the container with id bc82c442011807bbf16e641698169e943f13862b5af9857f6e60411a83d38c84 Feb 17 13:24:14 crc kubenswrapper[4955]: I0217 13:24:14.112652 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:24:14 crc kubenswrapper[4955]: I0217 13:24:14.171002 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f771ce9f-91d4-47c8-89fc-25758e4f907c","Type":"ContainerStarted","Data":"bc82c442011807bbf16e641698169e943f13862b5af9857f6e60411a83d38c84"} Feb 17 13:24:14 crc kubenswrapper[4955]: I0217 13:24:14.232939 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff" path="/var/lib/kubelet/pods/dc5a60f2-8c88-47b0-8d14-dfbf1cbd37ff/volumes" Feb 17 13:24:16 crc kubenswrapper[4955]: I0217 13:24:16.190073 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f771ce9f-91d4-47c8-89fc-25758e4f907c","Type":"ContainerStarted","Data":"dba37fa8433346b7a443086df2bae36315d46326fd6f1fc468e8725b8b2309d2"} Feb 17 13:24:17 crc kubenswrapper[4955]: I0217 13:24:17.205616 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f771ce9f-91d4-47c8-89fc-25758e4f907c","Type":"ContainerStarted","Data":"10f706b94bc1b191bd842698869370d5c05c90b95c1b5b26d32a46a1182e8353"} Feb 17 13:24:19 crc kubenswrapper[4955]: I0217 13:24:19.238353 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f771ce9f-91d4-47c8-89fc-25758e4f907c","Type":"ContainerStarted","Data":"2e4cc3b50f918f5973e76643cdac48b19dea65ff101323e5b100098a9efbe7a6"} Feb 17 13:24:20 crc kubenswrapper[4955]: I0217 13:24:20.259933 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f771ce9f-91d4-47c8-89fc-25758e4f907c","Type":"ContainerStarted","Data":"64ab492073a6b823183a45013260c1fc2a41f543af85c160a98064f41f5f4142"} Feb 17 13:24:20 crc kubenswrapper[4955]: I0217 13:24:20.260536 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 17 13:24:20 crc kubenswrapper[4955]: I0217 13:24:20.283175 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.805198447 podStartE2EDuration="7.283153333s" podCreationTimestamp="2026-02-17 13:24:13 +0000 UTC" firstStartedPulling="2026-02-17 13:24:14.112280201 +0000 UTC m=+1192.635009744" lastFinishedPulling="2026-02-17 13:24:19.590235077 +0000 UTC m=+1198.112964630" observedRunningTime="2026-02-17 13:24:20.279359705 +0000 UTC m=+1198.802089268" watchObservedRunningTime="2026-02-17 13:24:20.283153333 +0000 UTC m=+1198.805882876" Feb 17 13:24:21 crc kubenswrapper[4955]: I0217 13:24:21.271335 4955 generic.go:334] "Generic (PLEG): container finished" podID="fbdb6999-85bf-4bd2-b7f4-7b7c8566941f" containerID="73d6d40c5cbbe29be944855ed25368d36864681d3acca1c2904cc68409ee2622" exitCode=0 Feb 17 13:24:21 crc kubenswrapper[4955]: I0217 13:24:21.271430 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-ffnlw" event={"ID":"fbdb6999-85bf-4bd2-b7f4-7b7c8566941f","Type":"ContainerDied","Data":"73d6d40c5cbbe29be944855ed25368d36864681d3acca1c2904cc68409ee2622"} Feb 17 13:24:22 crc kubenswrapper[4955]: I0217 13:24:22.648986 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-ffnlw" Feb 17 13:24:22 crc kubenswrapper[4955]: I0217 13:24:22.706400 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbdb6999-85bf-4bd2-b7f4-7b7c8566941f-combined-ca-bundle\") pod \"fbdb6999-85bf-4bd2-b7f4-7b7c8566941f\" (UID: \"fbdb6999-85bf-4bd2-b7f4-7b7c8566941f\") " Feb 17 13:24:22 crc kubenswrapper[4955]: I0217 13:24:22.706486 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbdb6999-85bf-4bd2-b7f4-7b7c8566941f-scripts\") pod \"fbdb6999-85bf-4bd2-b7f4-7b7c8566941f\" (UID: \"fbdb6999-85bf-4bd2-b7f4-7b7c8566941f\") " Feb 17 13:24:22 crc kubenswrapper[4955]: I0217 13:24:22.706527 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6zcm\" (UniqueName: \"kubernetes.io/projected/fbdb6999-85bf-4bd2-b7f4-7b7c8566941f-kube-api-access-x6zcm\") pod \"fbdb6999-85bf-4bd2-b7f4-7b7c8566941f\" (UID: \"fbdb6999-85bf-4bd2-b7f4-7b7c8566941f\") " Feb 17 13:24:22 crc kubenswrapper[4955]: I0217 13:24:22.706621 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbdb6999-85bf-4bd2-b7f4-7b7c8566941f-config-data\") pod \"fbdb6999-85bf-4bd2-b7f4-7b7c8566941f\" (UID: \"fbdb6999-85bf-4bd2-b7f4-7b7c8566941f\") " Feb 17 13:24:22 crc kubenswrapper[4955]: I0217 13:24:22.714289 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbdb6999-85bf-4bd2-b7f4-7b7c8566941f-kube-api-access-x6zcm" (OuterVolumeSpecName: "kube-api-access-x6zcm") pod "fbdb6999-85bf-4bd2-b7f4-7b7c8566941f" (UID: "fbdb6999-85bf-4bd2-b7f4-7b7c8566941f"). InnerVolumeSpecName "kube-api-access-x6zcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:24:22 crc kubenswrapper[4955]: I0217 13:24:22.718905 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbdb6999-85bf-4bd2-b7f4-7b7c8566941f-scripts" (OuterVolumeSpecName: "scripts") pod "fbdb6999-85bf-4bd2-b7f4-7b7c8566941f" (UID: "fbdb6999-85bf-4bd2-b7f4-7b7c8566941f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:24:22 crc kubenswrapper[4955]: I0217 13:24:22.734191 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbdb6999-85bf-4bd2-b7f4-7b7c8566941f-config-data" (OuterVolumeSpecName: "config-data") pod "fbdb6999-85bf-4bd2-b7f4-7b7c8566941f" (UID: "fbdb6999-85bf-4bd2-b7f4-7b7c8566941f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:24:22 crc kubenswrapper[4955]: I0217 13:24:22.741074 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbdb6999-85bf-4bd2-b7f4-7b7c8566941f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fbdb6999-85bf-4bd2-b7f4-7b7c8566941f" (UID: "fbdb6999-85bf-4bd2-b7f4-7b7c8566941f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:24:22 crc kubenswrapper[4955]: I0217 13:24:22.809599 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbdb6999-85bf-4bd2-b7f4-7b7c8566941f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:24:22 crc kubenswrapper[4955]: I0217 13:24:22.809638 4955 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbdb6999-85bf-4bd2-b7f4-7b7c8566941f-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:24:22 crc kubenswrapper[4955]: I0217 13:24:22.809648 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6zcm\" (UniqueName: \"kubernetes.io/projected/fbdb6999-85bf-4bd2-b7f4-7b7c8566941f-kube-api-access-x6zcm\") on node \"crc\" DevicePath \"\"" Feb 17 13:24:22 crc kubenswrapper[4955]: I0217 13:24:22.809658 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbdb6999-85bf-4bd2-b7f4-7b7c8566941f-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:24:23 crc kubenswrapper[4955]: I0217 13:24:23.296502 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-ffnlw" event={"ID":"fbdb6999-85bf-4bd2-b7f4-7b7c8566941f","Type":"ContainerDied","Data":"593388dc8214c4cc4b502c47af3a68e45c32c7f61a07184ffd6a6daa832a6168"} Feb 17 13:24:23 crc kubenswrapper[4955]: I0217 13:24:23.296559 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="593388dc8214c4cc4b502c47af3a68e45c32c7f61a07184ffd6a6daa832a6168" Feb 17 13:24:23 crc kubenswrapper[4955]: I0217 13:24:23.296707 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-ffnlw" Feb 17 13:24:23 crc kubenswrapper[4955]: I0217 13:24:23.400673 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 17 13:24:23 crc kubenswrapper[4955]: E0217 13:24:23.401196 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbdb6999-85bf-4bd2-b7f4-7b7c8566941f" containerName="nova-cell0-conductor-db-sync" Feb 17 13:24:23 crc kubenswrapper[4955]: I0217 13:24:23.401221 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbdb6999-85bf-4bd2-b7f4-7b7c8566941f" containerName="nova-cell0-conductor-db-sync" Feb 17 13:24:23 crc kubenswrapper[4955]: I0217 13:24:23.401445 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbdb6999-85bf-4bd2-b7f4-7b7c8566941f" containerName="nova-cell0-conductor-db-sync" Feb 17 13:24:23 crc kubenswrapper[4955]: I0217 13:24:23.402227 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 17 13:24:23 crc kubenswrapper[4955]: I0217 13:24:23.404194 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Feb 17 13:24:23 crc kubenswrapper[4955]: I0217 13:24:23.405219 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-rz62x" Feb 17 13:24:23 crc kubenswrapper[4955]: I0217 13:24:23.414401 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 17 13:24:23 crc kubenswrapper[4955]: I0217 13:24:23.522772 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1d0ee97-a796-4b24-a77c-3626ac31256f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e1d0ee97-a796-4b24-a77c-3626ac31256f\") " pod="openstack/nova-cell0-conductor-0" Feb 17 13:24:23 crc kubenswrapper[4955]: I0217 13:24:23.523200 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsb9z\" (UniqueName: \"kubernetes.io/projected/e1d0ee97-a796-4b24-a77c-3626ac31256f-kube-api-access-jsb9z\") pod \"nova-cell0-conductor-0\" (UID: \"e1d0ee97-a796-4b24-a77c-3626ac31256f\") " pod="openstack/nova-cell0-conductor-0" Feb 17 13:24:23 crc kubenswrapper[4955]: I0217 13:24:23.523385 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1d0ee97-a796-4b24-a77c-3626ac31256f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e1d0ee97-a796-4b24-a77c-3626ac31256f\") " pod="openstack/nova-cell0-conductor-0" Feb 17 13:24:23 crc kubenswrapper[4955]: I0217 13:24:23.624724 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1d0ee97-a796-4b24-a77c-3626ac31256f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e1d0ee97-a796-4b24-a77c-3626ac31256f\") " pod="openstack/nova-cell0-conductor-0" Feb 17 13:24:23 crc kubenswrapper[4955]: I0217 13:24:23.625154 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsb9z\" (UniqueName: \"kubernetes.io/projected/e1d0ee97-a796-4b24-a77c-3626ac31256f-kube-api-access-jsb9z\") pod \"nova-cell0-conductor-0\" (UID: \"e1d0ee97-a796-4b24-a77c-3626ac31256f\") " pod="openstack/nova-cell0-conductor-0" Feb 17 13:24:23 crc kubenswrapper[4955]: I0217 13:24:23.625540 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1d0ee97-a796-4b24-a77c-3626ac31256f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e1d0ee97-a796-4b24-a77c-3626ac31256f\") " pod="openstack/nova-cell0-conductor-0" Feb 17 13:24:23 crc kubenswrapper[4955]: I0217 13:24:23.630851 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1d0ee97-a796-4b24-a77c-3626ac31256f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e1d0ee97-a796-4b24-a77c-3626ac31256f\") " pod="openstack/nova-cell0-conductor-0" Feb 17 13:24:23 crc kubenswrapper[4955]: I0217 13:24:23.632528 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1d0ee97-a796-4b24-a77c-3626ac31256f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e1d0ee97-a796-4b24-a77c-3626ac31256f\") " pod="openstack/nova-cell0-conductor-0" Feb 17 13:24:23 crc kubenswrapper[4955]: I0217 13:24:23.642089 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsb9z\" (UniqueName: \"kubernetes.io/projected/e1d0ee97-a796-4b24-a77c-3626ac31256f-kube-api-access-jsb9z\") pod \"nova-cell0-conductor-0\" (UID: \"e1d0ee97-a796-4b24-a77c-3626ac31256f\") " pod="openstack/nova-cell0-conductor-0" Feb 17 13:24:23 crc kubenswrapper[4955]: I0217 13:24:23.716850 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 17 13:24:24 crc kubenswrapper[4955]: I0217 13:24:24.093120 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 17 13:24:24 crc kubenswrapper[4955]: W0217 13:24:24.184806 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1d0ee97_a796_4b24_a77c_3626ac31256f.slice/crio-0d1b4b29bdab20545b32eee77b2e8a8abd3054afcaad57a080f08382db5c0dd4 WatchSource:0}: Error finding container 0d1b4b29bdab20545b32eee77b2e8a8abd3054afcaad57a080f08382db5c0dd4: Status 404 returned error can't find the container with id 0d1b4b29bdab20545b32eee77b2e8a8abd3054afcaad57a080f08382db5c0dd4 Feb 17 13:24:24 crc kubenswrapper[4955]: I0217 13:24:24.194789 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 17 13:24:24 crc kubenswrapper[4955]: I0217 13:24:24.312616 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"e1d0ee97-a796-4b24-a77c-3626ac31256f","Type":"ContainerStarted","Data":"0d1b4b29bdab20545b32eee77b2e8a8abd3054afcaad57a080f08382db5c0dd4"} Feb 17 13:24:25 crc kubenswrapper[4955]: I0217 13:24:25.324771 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"e1d0ee97-a796-4b24-a77c-3626ac31256f","Type":"ContainerStarted","Data":"9b207aca56cd7a2e24f4a4822726c62323aa73716a862d511996c9f70ad4fe05"} Feb 17 13:24:25 crc kubenswrapper[4955]: I0217 13:24:25.325257 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="e1d0ee97-a796-4b24-a77c-3626ac31256f" containerName="nova-cell0-conductor-conductor" containerID="cri-o://9b207aca56cd7a2e24f4a4822726c62323aa73716a862d511996c9f70ad4fe05" gracePeriod=30 Feb 17 13:24:25 crc kubenswrapper[4955]: I0217 13:24:25.325415 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Feb 17 13:24:25 crc kubenswrapper[4955]: I0217 13:24:25.350938 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.350913687 podStartE2EDuration="2.350913687s" podCreationTimestamp="2026-02-17 13:24:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:24:25.344300548 +0000 UTC m=+1203.867030121" watchObservedRunningTime="2026-02-17 13:24:25.350913687 +0000 UTC m=+1203.873643230" Feb 17 13:24:25 crc kubenswrapper[4955]: I0217 13:24:25.870091 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:24:25 crc kubenswrapper[4955]: I0217 13:24:25.870374 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f771ce9f-91d4-47c8-89fc-25758e4f907c" containerName="ceilometer-central-agent" containerID="cri-o://dba37fa8433346b7a443086df2bae36315d46326fd6f1fc468e8725b8b2309d2" gracePeriod=30 Feb 17 13:24:25 crc kubenswrapper[4955]: I0217 13:24:25.870486 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f771ce9f-91d4-47c8-89fc-25758e4f907c" containerName="proxy-httpd" containerID="cri-o://64ab492073a6b823183a45013260c1fc2a41f543af85c160a98064f41f5f4142" gracePeriod=30 Feb 17 13:24:25 crc kubenswrapper[4955]: I0217 13:24:25.870528 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f771ce9f-91d4-47c8-89fc-25758e4f907c" containerName="sg-core" containerID="cri-o://2e4cc3b50f918f5973e76643cdac48b19dea65ff101323e5b100098a9efbe7a6" gracePeriod=30 Feb 17 13:24:25 crc kubenswrapper[4955]: I0217 13:24:25.870562 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f771ce9f-91d4-47c8-89fc-25758e4f907c" containerName="ceilometer-notification-agent" containerID="cri-o://10f706b94bc1b191bd842698869370d5c05c90b95c1b5b26d32a46a1182e8353" gracePeriod=30 Feb 17 13:24:26 crc kubenswrapper[4955]: I0217 13:24:26.349762 4955 generic.go:334] "Generic (PLEG): container finished" podID="f771ce9f-91d4-47c8-89fc-25758e4f907c" containerID="64ab492073a6b823183a45013260c1fc2a41f543af85c160a98064f41f5f4142" exitCode=0 Feb 17 13:24:26 crc kubenswrapper[4955]: I0217 13:24:26.349815 4955 generic.go:334] "Generic (PLEG): container finished" podID="f771ce9f-91d4-47c8-89fc-25758e4f907c" containerID="2e4cc3b50f918f5973e76643cdac48b19dea65ff101323e5b100098a9efbe7a6" exitCode=2 Feb 17 13:24:26 crc kubenswrapper[4955]: I0217 13:24:26.349825 4955 generic.go:334] "Generic (PLEG): container finished" podID="f771ce9f-91d4-47c8-89fc-25758e4f907c" containerID="dba37fa8433346b7a443086df2bae36315d46326fd6f1fc468e8725b8b2309d2" exitCode=0 Feb 17 13:24:26 crc kubenswrapper[4955]: I0217 13:24:26.349846 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f771ce9f-91d4-47c8-89fc-25758e4f907c","Type":"ContainerDied","Data":"64ab492073a6b823183a45013260c1fc2a41f543af85c160a98064f41f5f4142"} Feb 17 13:24:26 crc kubenswrapper[4955]: I0217 13:24:26.349873 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f771ce9f-91d4-47c8-89fc-25758e4f907c","Type":"ContainerDied","Data":"2e4cc3b50f918f5973e76643cdac48b19dea65ff101323e5b100098a9efbe7a6"} Feb 17 13:24:26 crc kubenswrapper[4955]: I0217 13:24:26.349890 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f771ce9f-91d4-47c8-89fc-25758e4f907c","Type":"ContainerDied","Data":"dba37fa8433346b7a443086df2bae36315d46326fd6f1fc468e8725b8b2309d2"} Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.025095 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.098093 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f771ce9f-91d4-47c8-89fc-25758e4f907c-run-httpd\") pod \"f771ce9f-91d4-47c8-89fc-25758e4f907c\" (UID: \"f771ce9f-91d4-47c8-89fc-25758e4f907c\") " Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.098206 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f771ce9f-91d4-47c8-89fc-25758e4f907c-scripts\") pod \"f771ce9f-91d4-47c8-89fc-25758e4f907c\" (UID: \"f771ce9f-91d4-47c8-89fc-25758e4f907c\") " Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.098277 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f771ce9f-91d4-47c8-89fc-25758e4f907c-config-data\") pod \"f771ce9f-91d4-47c8-89fc-25758e4f907c\" (UID: \"f771ce9f-91d4-47c8-89fc-25758e4f907c\") " Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.098311 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwbg8\" (UniqueName: \"kubernetes.io/projected/f771ce9f-91d4-47c8-89fc-25758e4f907c-kube-api-access-vwbg8\") pod \"f771ce9f-91d4-47c8-89fc-25758e4f907c\" (UID: \"f771ce9f-91d4-47c8-89fc-25758e4f907c\") " Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.098349 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f771ce9f-91d4-47c8-89fc-25758e4f907c-sg-core-conf-yaml\") pod \"f771ce9f-91d4-47c8-89fc-25758e4f907c\" (UID: \"f771ce9f-91d4-47c8-89fc-25758e4f907c\") " Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.098403 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f771ce9f-91d4-47c8-89fc-25758e4f907c-log-httpd\") pod \"f771ce9f-91d4-47c8-89fc-25758e4f907c\" (UID: \"f771ce9f-91d4-47c8-89fc-25758e4f907c\") " Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.098448 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f771ce9f-91d4-47c8-89fc-25758e4f907c-combined-ca-bundle\") pod \"f771ce9f-91d4-47c8-89fc-25758e4f907c\" (UID: \"f771ce9f-91d4-47c8-89fc-25758e4f907c\") " Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.098535 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f771ce9f-91d4-47c8-89fc-25758e4f907c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f771ce9f-91d4-47c8-89fc-25758e4f907c" (UID: "f771ce9f-91d4-47c8-89fc-25758e4f907c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.099582 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f771ce9f-91d4-47c8-89fc-25758e4f907c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f771ce9f-91d4-47c8-89fc-25758e4f907c" (UID: "f771ce9f-91d4-47c8-89fc-25758e4f907c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.099816 4955 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f771ce9f-91d4-47c8-89fc-25758e4f907c-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.099861 4955 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f771ce9f-91d4-47c8-89fc-25758e4f907c-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.107643 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f771ce9f-91d4-47c8-89fc-25758e4f907c-kube-api-access-vwbg8" (OuterVolumeSpecName: "kube-api-access-vwbg8") pod "f771ce9f-91d4-47c8-89fc-25758e4f907c" (UID: "f771ce9f-91d4-47c8-89fc-25758e4f907c"). InnerVolumeSpecName "kube-api-access-vwbg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.113436 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f771ce9f-91d4-47c8-89fc-25758e4f907c-scripts" (OuterVolumeSpecName: "scripts") pod "f771ce9f-91d4-47c8-89fc-25758e4f907c" (UID: "f771ce9f-91d4-47c8-89fc-25758e4f907c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.131636 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f771ce9f-91d4-47c8-89fc-25758e4f907c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f771ce9f-91d4-47c8-89fc-25758e4f907c" (UID: "f771ce9f-91d4-47c8-89fc-25758e4f907c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.178767 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f771ce9f-91d4-47c8-89fc-25758e4f907c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f771ce9f-91d4-47c8-89fc-25758e4f907c" (UID: "f771ce9f-91d4-47c8-89fc-25758e4f907c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.197192 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f771ce9f-91d4-47c8-89fc-25758e4f907c-config-data" (OuterVolumeSpecName: "config-data") pod "f771ce9f-91d4-47c8-89fc-25758e4f907c" (UID: "f771ce9f-91d4-47c8-89fc-25758e4f907c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.201520 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f771ce9f-91d4-47c8-89fc-25758e4f907c-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.201562 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwbg8\" (UniqueName: \"kubernetes.io/projected/f771ce9f-91d4-47c8-89fc-25758e4f907c-kube-api-access-vwbg8\") on node \"crc\" DevicePath \"\"" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.201579 4955 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f771ce9f-91d4-47c8-89fc-25758e4f907c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.201591 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f771ce9f-91d4-47c8-89fc-25758e4f907c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.201603 4955 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f771ce9f-91d4-47c8-89fc-25758e4f907c-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.359945 4955 generic.go:334] "Generic (PLEG): container finished" podID="f771ce9f-91d4-47c8-89fc-25758e4f907c" containerID="10f706b94bc1b191bd842698869370d5c05c90b95c1b5b26d32a46a1182e8353" exitCode=0 Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.360006 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f771ce9f-91d4-47c8-89fc-25758e4f907c","Type":"ContainerDied","Data":"10f706b94bc1b191bd842698869370d5c05c90b95c1b5b26d32a46a1182e8353"} Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.360062 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.360709 4955 scope.go:117] "RemoveContainer" containerID="64ab492073a6b823183a45013260c1fc2a41f543af85c160a98064f41f5f4142" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.360694 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f771ce9f-91d4-47c8-89fc-25758e4f907c","Type":"ContainerDied","Data":"bc82c442011807bbf16e641698169e943f13862b5af9857f6e60411a83d38c84"} Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.381101 4955 scope.go:117] "RemoveContainer" containerID="2e4cc3b50f918f5973e76643cdac48b19dea65ff101323e5b100098a9efbe7a6" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.419911 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.420459 4955 scope.go:117] "RemoveContainer" containerID="10f706b94bc1b191bd842698869370d5c05c90b95c1b5b26d32a46a1182e8353" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.431764 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.446209 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:24:27 crc kubenswrapper[4955]: E0217 13:24:27.446605 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f771ce9f-91d4-47c8-89fc-25758e4f907c" containerName="proxy-httpd" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.446622 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="f771ce9f-91d4-47c8-89fc-25758e4f907c" containerName="proxy-httpd" Feb 17 13:24:27 crc kubenswrapper[4955]: E0217 13:24:27.446639 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f771ce9f-91d4-47c8-89fc-25758e4f907c" containerName="ceilometer-central-agent" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.446647 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="f771ce9f-91d4-47c8-89fc-25758e4f907c" containerName="ceilometer-central-agent" Feb 17 13:24:27 crc kubenswrapper[4955]: E0217 13:24:27.446665 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f771ce9f-91d4-47c8-89fc-25758e4f907c" containerName="ceilometer-notification-agent" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.446672 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="f771ce9f-91d4-47c8-89fc-25758e4f907c" containerName="ceilometer-notification-agent" Feb 17 13:24:27 crc kubenswrapper[4955]: E0217 13:24:27.446687 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f771ce9f-91d4-47c8-89fc-25758e4f907c" containerName="sg-core" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.446693 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="f771ce9f-91d4-47c8-89fc-25758e4f907c" containerName="sg-core" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.446890 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="f771ce9f-91d4-47c8-89fc-25758e4f907c" containerName="ceilometer-central-agent" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.446905 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="f771ce9f-91d4-47c8-89fc-25758e4f907c" containerName="ceilometer-notification-agent" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.446917 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="f771ce9f-91d4-47c8-89fc-25758e4f907c" containerName="proxy-httpd" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.446929 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="f771ce9f-91d4-47c8-89fc-25758e4f907c" containerName="sg-core" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.448960 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.452247 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.452456 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.456857 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.465658 4955 scope.go:117] "RemoveContainer" containerID="dba37fa8433346b7a443086df2bae36315d46326fd6f1fc468e8725b8b2309d2" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.502708 4955 scope.go:117] "RemoveContainer" containerID="64ab492073a6b823183a45013260c1fc2a41f543af85c160a98064f41f5f4142" Feb 17 13:24:27 crc kubenswrapper[4955]: E0217 13:24:27.503593 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64ab492073a6b823183a45013260c1fc2a41f543af85c160a98064f41f5f4142\": container with ID starting with 64ab492073a6b823183a45013260c1fc2a41f543af85c160a98064f41f5f4142 not found: ID does not exist" containerID="64ab492073a6b823183a45013260c1fc2a41f543af85c160a98064f41f5f4142" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.503630 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64ab492073a6b823183a45013260c1fc2a41f543af85c160a98064f41f5f4142"} err="failed to get container status \"64ab492073a6b823183a45013260c1fc2a41f543af85c160a98064f41f5f4142\": rpc error: code = NotFound desc = could not find container \"64ab492073a6b823183a45013260c1fc2a41f543af85c160a98064f41f5f4142\": container with ID starting with 64ab492073a6b823183a45013260c1fc2a41f543af85c160a98064f41f5f4142 not found: ID does not exist" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.503652 4955 scope.go:117] "RemoveContainer" containerID="2e4cc3b50f918f5973e76643cdac48b19dea65ff101323e5b100098a9efbe7a6" Feb 17 13:24:27 crc kubenswrapper[4955]: E0217 13:24:27.503999 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e4cc3b50f918f5973e76643cdac48b19dea65ff101323e5b100098a9efbe7a6\": container with ID starting with 2e4cc3b50f918f5973e76643cdac48b19dea65ff101323e5b100098a9efbe7a6 not found: ID does not exist" containerID="2e4cc3b50f918f5973e76643cdac48b19dea65ff101323e5b100098a9efbe7a6" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.504020 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e4cc3b50f918f5973e76643cdac48b19dea65ff101323e5b100098a9efbe7a6"} err="failed to get container status \"2e4cc3b50f918f5973e76643cdac48b19dea65ff101323e5b100098a9efbe7a6\": rpc error: code = NotFound desc = could not find container \"2e4cc3b50f918f5973e76643cdac48b19dea65ff101323e5b100098a9efbe7a6\": container with ID starting with 2e4cc3b50f918f5973e76643cdac48b19dea65ff101323e5b100098a9efbe7a6 not found: ID does not exist" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.504033 4955 scope.go:117] "RemoveContainer" containerID="10f706b94bc1b191bd842698869370d5c05c90b95c1b5b26d32a46a1182e8353" Feb 17 13:24:27 crc kubenswrapper[4955]: E0217 13:24:27.504257 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10f706b94bc1b191bd842698869370d5c05c90b95c1b5b26d32a46a1182e8353\": container with ID starting with 10f706b94bc1b191bd842698869370d5c05c90b95c1b5b26d32a46a1182e8353 not found: ID does not exist" containerID="10f706b94bc1b191bd842698869370d5c05c90b95c1b5b26d32a46a1182e8353" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.504278 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10f706b94bc1b191bd842698869370d5c05c90b95c1b5b26d32a46a1182e8353"} err="failed to get container status \"10f706b94bc1b191bd842698869370d5c05c90b95c1b5b26d32a46a1182e8353\": rpc error: code = NotFound desc = could not find container \"10f706b94bc1b191bd842698869370d5c05c90b95c1b5b26d32a46a1182e8353\": container with ID starting with 10f706b94bc1b191bd842698869370d5c05c90b95c1b5b26d32a46a1182e8353 not found: ID does not exist" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.504292 4955 scope.go:117] "RemoveContainer" containerID="dba37fa8433346b7a443086df2bae36315d46326fd6f1fc468e8725b8b2309d2" Feb 17 13:24:27 crc kubenswrapper[4955]: E0217 13:24:27.504478 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dba37fa8433346b7a443086df2bae36315d46326fd6f1fc468e8725b8b2309d2\": container with ID starting with dba37fa8433346b7a443086df2bae36315d46326fd6f1fc468e8725b8b2309d2 not found: ID does not exist" containerID="dba37fa8433346b7a443086df2bae36315d46326fd6f1fc468e8725b8b2309d2" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.504497 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dba37fa8433346b7a443086df2bae36315d46326fd6f1fc468e8725b8b2309d2"} err="failed to get container status \"dba37fa8433346b7a443086df2bae36315d46326fd6f1fc468e8725b8b2309d2\": rpc error: code = NotFound desc = could not find container \"dba37fa8433346b7a443086df2bae36315d46326fd6f1fc468e8725b8b2309d2\": container with ID starting with dba37fa8433346b7a443086df2bae36315d46326fd6f1fc468e8725b8b2309d2 not found: ID does not exist" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.511545 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhm99\" (UniqueName: \"kubernetes.io/projected/7389a7d7-83cb-4c0d-bb69-030c3331caf5-kube-api-access-vhm99\") pod \"ceilometer-0\" (UID: \"7389a7d7-83cb-4c0d-bb69-030c3331caf5\") " pod="openstack/ceilometer-0" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.511678 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7389a7d7-83cb-4c0d-bb69-030c3331caf5-scripts\") pod \"ceilometer-0\" (UID: \"7389a7d7-83cb-4c0d-bb69-030c3331caf5\") " pod="openstack/ceilometer-0" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.511699 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7389a7d7-83cb-4c0d-bb69-030c3331caf5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7389a7d7-83cb-4c0d-bb69-030c3331caf5\") " pod="openstack/ceilometer-0" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.511859 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7389a7d7-83cb-4c0d-bb69-030c3331caf5-config-data\") pod \"ceilometer-0\" (UID: \"7389a7d7-83cb-4c0d-bb69-030c3331caf5\") " pod="openstack/ceilometer-0" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.511929 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7389a7d7-83cb-4c0d-bb69-030c3331caf5-run-httpd\") pod \"ceilometer-0\" (UID: \"7389a7d7-83cb-4c0d-bb69-030c3331caf5\") " pod="openstack/ceilometer-0" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.512064 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7389a7d7-83cb-4c0d-bb69-030c3331caf5-log-httpd\") pod \"ceilometer-0\" (UID: \"7389a7d7-83cb-4c0d-bb69-030c3331caf5\") " pod="openstack/ceilometer-0" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.512092 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7389a7d7-83cb-4c0d-bb69-030c3331caf5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7389a7d7-83cb-4c0d-bb69-030c3331caf5\") " pod="openstack/ceilometer-0" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.613275 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7389a7d7-83cb-4c0d-bb69-030c3331caf5-config-data\") pod \"ceilometer-0\" (UID: \"7389a7d7-83cb-4c0d-bb69-030c3331caf5\") " pod="openstack/ceilometer-0" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.613325 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7389a7d7-83cb-4c0d-bb69-030c3331caf5-run-httpd\") pod \"ceilometer-0\" (UID: \"7389a7d7-83cb-4c0d-bb69-030c3331caf5\") " pod="openstack/ceilometer-0" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.613382 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7389a7d7-83cb-4c0d-bb69-030c3331caf5-log-httpd\") pod \"ceilometer-0\" (UID: \"7389a7d7-83cb-4c0d-bb69-030c3331caf5\") " pod="openstack/ceilometer-0" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.613402 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7389a7d7-83cb-4c0d-bb69-030c3331caf5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7389a7d7-83cb-4c0d-bb69-030c3331caf5\") " pod="openstack/ceilometer-0" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.613865 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7389a7d7-83cb-4c0d-bb69-030c3331caf5-log-httpd\") pod \"ceilometer-0\" (UID: \"7389a7d7-83cb-4c0d-bb69-030c3331caf5\") " pod="openstack/ceilometer-0" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.613996 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7389a7d7-83cb-4c0d-bb69-030c3331caf5-run-httpd\") pod \"ceilometer-0\" (UID: \"7389a7d7-83cb-4c0d-bb69-030c3331caf5\") " pod="openstack/ceilometer-0" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.614190 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhm99\" (UniqueName: \"kubernetes.io/projected/7389a7d7-83cb-4c0d-bb69-030c3331caf5-kube-api-access-vhm99\") pod \"ceilometer-0\" (UID: \"7389a7d7-83cb-4c0d-bb69-030c3331caf5\") " pod="openstack/ceilometer-0" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.614313 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7389a7d7-83cb-4c0d-bb69-030c3331caf5-scripts\") pod \"ceilometer-0\" (UID: \"7389a7d7-83cb-4c0d-bb69-030c3331caf5\") " pod="openstack/ceilometer-0" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.614338 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7389a7d7-83cb-4c0d-bb69-030c3331caf5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7389a7d7-83cb-4c0d-bb69-030c3331caf5\") " pod="openstack/ceilometer-0" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.618767 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7389a7d7-83cb-4c0d-bb69-030c3331caf5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7389a7d7-83cb-4c0d-bb69-030c3331caf5\") " pod="openstack/ceilometer-0" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.619211 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7389a7d7-83cb-4c0d-bb69-030c3331caf5-config-data\") pod \"ceilometer-0\" (UID: \"7389a7d7-83cb-4c0d-bb69-030c3331caf5\") " pod="openstack/ceilometer-0" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.619477 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7389a7d7-83cb-4c0d-bb69-030c3331caf5-scripts\") pod \"ceilometer-0\" (UID: \"7389a7d7-83cb-4c0d-bb69-030c3331caf5\") " pod="openstack/ceilometer-0" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.620571 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7389a7d7-83cb-4c0d-bb69-030c3331caf5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7389a7d7-83cb-4c0d-bb69-030c3331caf5\") " pod="openstack/ceilometer-0" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.631387 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhm99\" (UniqueName: \"kubernetes.io/projected/7389a7d7-83cb-4c0d-bb69-030c3331caf5-kube-api-access-vhm99\") pod \"ceilometer-0\" (UID: \"7389a7d7-83cb-4c0d-bb69-030c3331caf5\") " pod="openstack/ceilometer-0" Feb 17 13:24:27 crc kubenswrapper[4955]: I0217 13:24:27.766678 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 17 13:24:28 crc kubenswrapper[4955]: I0217 13:24:28.082897 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:24:28 crc kubenswrapper[4955]: I0217 13:24:28.237535 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f771ce9f-91d4-47c8-89fc-25758e4f907c" path="/var/lib/kubelet/pods/f771ce9f-91d4-47c8-89fc-25758e4f907c/volumes" Feb 17 13:24:28 crc kubenswrapper[4955]: I0217 13:24:28.373673 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7389a7d7-83cb-4c0d-bb69-030c3331caf5","Type":"ContainerStarted","Data":"c929e36b4ca9fb82be6838f34032ea4e6cea8b4f48381fd07e51ea2352306457"} Feb 17 13:24:29 crc kubenswrapper[4955]: I0217 13:24:29.393481 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7389a7d7-83cb-4c0d-bb69-030c3331caf5","Type":"ContainerStarted","Data":"abc6b8d601853195caab008f3b8e84b0ccbedfad1fc886969c8c98835f077d7f"} Feb 17 13:24:30 crc kubenswrapper[4955]: I0217 13:24:30.407835 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7389a7d7-83cb-4c0d-bb69-030c3331caf5","Type":"ContainerStarted","Data":"9052f1507175292366564a8ac671e92809a029127e38cbd05a33f28841be19d9"} Feb 17 13:24:30 crc kubenswrapper[4955]: I0217 13:24:30.408262 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7389a7d7-83cb-4c0d-bb69-030c3331caf5","Type":"ContainerStarted","Data":"a5fa3a4ed0a0ff395c5c44ad2e45b52e16a7527b29765548b1e51df776873468"} Feb 17 13:24:32 crc kubenswrapper[4955]: I0217 13:24:32.429872 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7389a7d7-83cb-4c0d-bb69-030c3331caf5","Type":"ContainerStarted","Data":"33c70cbe6a16ffe2997b72bfe5ffe17ac36b209090875f6c32429fbdde279ca7"} Feb 17 13:24:32 crc kubenswrapper[4955]: I0217 13:24:32.430321 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 17 13:24:32 crc kubenswrapper[4955]: I0217 13:24:32.468504 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.024110962 podStartE2EDuration="5.468480042s" podCreationTimestamp="2026-02-17 13:24:27 +0000 UTC" firstStartedPulling="2026-02-17 13:24:28.088036832 +0000 UTC m=+1206.610766375" lastFinishedPulling="2026-02-17 13:24:31.532405912 +0000 UTC m=+1210.055135455" observedRunningTime="2026-02-17 13:24:32.462990855 +0000 UTC m=+1210.985720398" watchObservedRunningTime="2026-02-17 13:24:32.468480042 +0000 UTC m=+1210.991209585" Feb 17 13:24:33 crc kubenswrapper[4955]: E0217 13:24:33.722519 4955 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9b207aca56cd7a2e24f4a4822726c62323aa73716a862d511996c9f70ad4fe05" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 17 13:24:33 crc kubenswrapper[4955]: E0217 13:24:33.724614 4955 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9b207aca56cd7a2e24f4a4822726c62323aa73716a862d511996c9f70ad4fe05" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 17 13:24:33 crc kubenswrapper[4955]: E0217 13:24:33.726350 4955 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9b207aca56cd7a2e24f4a4822726c62323aa73716a862d511996c9f70ad4fe05" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 17 13:24:33 crc kubenswrapper[4955]: E0217 13:24:33.726383 4955 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="e1d0ee97-a796-4b24-a77c-3626ac31256f" containerName="nova-cell0-conductor-conductor" Feb 17 13:24:38 crc kubenswrapper[4955]: E0217 13:24:38.720602 4955 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9b207aca56cd7a2e24f4a4822726c62323aa73716a862d511996c9f70ad4fe05" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 17 13:24:38 crc kubenswrapper[4955]: E0217 13:24:38.723004 4955 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9b207aca56cd7a2e24f4a4822726c62323aa73716a862d511996c9f70ad4fe05" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 17 13:24:38 crc kubenswrapper[4955]: E0217 13:24:38.740261 4955 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9b207aca56cd7a2e24f4a4822726c62323aa73716a862d511996c9f70ad4fe05" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 17 13:24:38 crc kubenswrapper[4955]: E0217 13:24:38.740347 4955 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="e1d0ee97-a796-4b24-a77c-3626ac31256f" containerName="nova-cell0-conductor-conductor" Feb 17 13:24:43 crc kubenswrapper[4955]: E0217 13:24:43.720022 4955 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9b207aca56cd7a2e24f4a4822726c62323aa73716a862d511996c9f70ad4fe05" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 17 13:24:43 crc kubenswrapper[4955]: E0217 13:24:43.722955 4955 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9b207aca56cd7a2e24f4a4822726c62323aa73716a862d511996c9f70ad4fe05" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 17 13:24:43 crc kubenswrapper[4955]: E0217 13:24:43.725261 4955 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9b207aca56cd7a2e24f4a4822726c62323aa73716a862d511996c9f70ad4fe05" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 17 13:24:43 crc kubenswrapper[4955]: E0217 13:24:43.725327 4955 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="e1d0ee97-a796-4b24-a77c-3626ac31256f" containerName="nova-cell0-conductor-conductor" Feb 17 13:24:48 crc kubenswrapper[4955]: E0217 13:24:48.720364 4955 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9b207aca56cd7a2e24f4a4822726c62323aa73716a862d511996c9f70ad4fe05" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 17 13:24:48 crc kubenswrapper[4955]: E0217 13:24:48.722915 4955 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9b207aca56cd7a2e24f4a4822726c62323aa73716a862d511996c9f70ad4fe05" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 17 13:24:48 crc kubenswrapper[4955]: E0217 13:24:48.724495 4955 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9b207aca56cd7a2e24f4a4822726c62323aa73716a862d511996c9f70ad4fe05" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 17 13:24:48 crc kubenswrapper[4955]: E0217 13:24:48.724539 4955 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="e1d0ee97-a796-4b24-a77c-3626ac31256f" containerName="nova-cell0-conductor-conductor" Feb 17 13:24:53 crc kubenswrapper[4955]: E0217 13:24:53.721365 4955 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9b207aca56cd7a2e24f4a4822726c62323aa73716a862d511996c9f70ad4fe05" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 17 13:24:53 crc kubenswrapper[4955]: E0217 13:24:53.724057 4955 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9b207aca56cd7a2e24f4a4822726c62323aa73716a862d511996c9f70ad4fe05" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 17 13:24:53 crc kubenswrapper[4955]: E0217 13:24:53.725717 4955 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9b207aca56cd7a2e24f4a4822726c62323aa73716a862d511996c9f70ad4fe05" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 17 13:24:53 crc kubenswrapper[4955]: E0217 13:24:53.725754 4955 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="e1d0ee97-a796-4b24-a77c-3626ac31256f" containerName="nova-cell0-conductor-conductor" Feb 17 13:24:55 crc kubenswrapper[4955]: I0217 13:24:55.667543 4955 generic.go:334] "Generic (PLEG): container finished" podID="e1d0ee97-a796-4b24-a77c-3626ac31256f" containerID="9b207aca56cd7a2e24f4a4822726c62323aa73716a862d511996c9f70ad4fe05" exitCode=137 Feb 17 13:24:55 crc kubenswrapper[4955]: I0217 13:24:55.667635 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"e1d0ee97-a796-4b24-a77c-3626ac31256f","Type":"ContainerDied","Data":"9b207aca56cd7a2e24f4a4822726c62323aa73716a862d511996c9f70ad4fe05"} Feb 17 13:24:55 crc kubenswrapper[4955]: I0217 13:24:55.667939 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"e1d0ee97-a796-4b24-a77c-3626ac31256f","Type":"ContainerDied","Data":"0d1b4b29bdab20545b32eee77b2e8a8abd3054afcaad57a080f08382db5c0dd4"} Feb 17 13:24:55 crc kubenswrapper[4955]: I0217 13:24:55.667958 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d1b4b29bdab20545b32eee77b2e8a8abd3054afcaad57a080f08382db5c0dd4" Feb 17 13:24:55 crc kubenswrapper[4955]: I0217 13:24:55.716368 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 17 13:24:55 crc kubenswrapper[4955]: I0217 13:24:55.879331 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jsb9z\" (UniqueName: \"kubernetes.io/projected/e1d0ee97-a796-4b24-a77c-3626ac31256f-kube-api-access-jsb9z\") pod \"e1d0ee97-a796-4b24-a77c-3626ac31256f\" (UID: \"e1d0ee97-a796-4b24-a77c-3626ac31256f\") " Feb 17 13:24:55 crc kubenswrapper[4955]: I0217 13:24:55.879434 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1d0ee97-a796-4b24-a77c-3626ac31256f-combined-ca-bundle\") pod \"e1d0ee97-a796-4b24-a77c-3626ac31256f\" (UID: \"e1d0ee97-a796-4b24-a77c-3626ac31256f\") " Feb 17 13:24:55 crc kubenswrapper[4955]: I0217 13:24:55.879473 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1d0ee97-a796-4b24-a77c-3626ac31256f-config-data\") pod \"e1d0ee97-a796-4b24-a77c-3626ac31256f\" (UID: \"e1d0ee97-a796-4b24-a77c-3626ac31256f\") " Feb 17 13:24:55 crc kubenswrapper[4955]: I0217 13:24:55.884614 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1d0ee97-a796-4b24-a77c-3626ac31256f-kube-api-access-jsb9z" (OuterVolumeSpecName: "kube-api-access-jsb9z") pod "e1d0ee97-a796-4b24-a77c-3626ac31256f" (UID: "e1d0ee97-a796-4b24-a77c-3626ac31256f"). InnerVolumeSpecName "kube-api-access-jsb9z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:24:55 crc kubenswrapper[4955]: I0217 13:24:55.908151 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1d0ee97-a796-4b24-a77c-3626ac31256f-config-data" (OuterVolumeSpecName: "config-data") pod "e1d0ee97-a796-4b24-a77c-3626ac31256f" (UID: "e1d0ee97-a796-4b24-a77c-3626ac31256f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:24:55 crc kubenswrapper[4955]: I0217 13:24:55.910662 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1d0ee97-a796-4b24-a77c-3626ac31256f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e1d0ee97-a796-4b24-a77c-3626ac31256f" (UID: "e1d0ee97-a796-4b24-a77c-3626ac31256f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:24:55 crc kubenswrapper[4955]: I0217 13:24:55.981211 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1d0ee97-a796-4b24-a77c-3626ac31256f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:24:55 crc kubenswrapper[4955]: I0217 13:24:55.981244 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1d0ee97-a796-4b24-a77c-3626ac31256f-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:24:55 crc kubenswrapper[4955]: I0217 13:24:55.981256 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jsb9z\" (UniqueName: \"kubernetes.io/projected/e1d0ee97-a796-4b24-a77c-3626ac31256f-kube-api-access-jsb9z\") on node \"crc\" DevicePath \"\"" Feb 17 13:24:56 crc kubenswrapper[4955]: I0217 13:24:56.681851 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 17 13:24:56 crc kubenswrapper[4955]: I0217 13:24:56.713167 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 17 13:24:56 crc kubenswrapper[4955]: I0217 13:24:56.734336 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 17 13:24:56 crc kubenswrapper[4955]: I0217 13:24:56.742223 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 17 13:24:56 crc kubenswrapper[4955]: E0217 13:24:56.742710 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1d0ee97-a796-4b24-a77c-3626ac31256f" containerName="nova-cell0-conductor-conductor" Feb 17 13:24:56 crc kubenswrapper[4955]: I0217 13:24:56.742735 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1d0ee97-a796-4b24-a77c-3626ac31256f" containerName="nova-cell0-conductor-conductor" Feb 17 13:24:56 crc kubenswrapper[4955]: I0217 13:24:56.742979 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1d0ee97-a796-4b24-a77c-3626ac31256f" containerName="nova-cell0-conductor-conductor" Feb 17 13:24:56 crc kubenswrapper[4955]: I0217 13:24:56.744658 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 17 13:24:56 crc kubenswrapper[4955]: I0217 13:24:56.750651 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-rz62x" Feb 17 13:24:56 crc kubenswrapper[4955]: I0217 13:24:56.750955 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Feb 17 13:24:56 crc kubenswrapper[4955]: I0217 13:24:56.751388 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 17 13:24:56 crc kubenswrapper[4955]: I0217 13:24:56.904412 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nr8z\" (UniqueName: \"kubernetes.io/projected/115dedcd-9ffe-408e-9b6b-878f53c6c252-kube-api-access-9nr8z\") pod \"nova-cell0-conductor-0\" (UID: \"115dedcd-9ffe-408e-9b6b-878f53c6c252\") " pod="openstack/nova-cell0-conductor-0" Feb 17 13:24:56 crc kubenswrapper[4955]: I0217 13:24:56.904649 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/115dedcd-9ffe-408e-9b6b-878f53c6c252-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"115dedcd-9ffe-408e-9b6b-878f53c6c252\") " pod="openstack/nova-cell0-conductor-0" Feb 17 13:24:56 crc kubenswrapper[4955]: I0217 13:24:56.904821 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/115dedcd-9ffe-408e-9b6b-878f53c6c252-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"115dedcd-9ffe-408e-9b6b-878f53c6c252\") " pod="openstack/nova-cell0-conductor-0" Feb 17 13:24:57 crc kubenswrapper[4955]: I0217 13:24:57.006753 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/115dedcd-9ffe-408e-9b6b-878f53c6c252-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"115dedcd-9ffe-408e-9b6b-878f53c6c252\") " pod="openstack/nova-cell0-conductor-0" Feb 17 13:24:57 crc kubenswrapper[4955]: I0217 13:24:57.007101 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nr8z\" (UniqueName: \"kubernetes.io/projected/115dedcd-9ffe-408e-9b6b-878f53c6c252-kube-api-access-9nr8z\") pod \"nova-cell0-conductor-0\" (UID: \"115dedcd-9ffe-408e-9b6b-878f53c6c252\") " pod="openstack/nova-cell0-conductor-0" Feb 17 13:24:57 crc kubenswrapper[4955]: I0217 13:24:57.007216 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/115dedcd-9ffe-408e-9b6b-878f53c6c252-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"115dedcd-9ffe-408e-9b6b-878f53c6c252\") " pod="openstack/nova-cell0-conductor-0" Feb 17 13:24:57 crc kubenswrapper[4955]: I0217 13:24:57.011576 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/115dedcd-9ffe-408e-9b6b-878f53c6c252-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"115dedcd-9ffe-408e-9b6b-878f53c6c252\") " pod="openstack/nova-cell0-conductor-0" Feb 17 13:24:57 crc kubenswrapper[4955]: I0217 13:24:57.012336 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/115dedcd-9ffe-408e-9b6b-878f53c6c252-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"115dedcd-9ffe-408e-9b6b-878f53c6c252\") " pod="openstack/nova-cell0-conductor-0" Feb 17 13:24:57 crc kubenswrapper[4955]: I0217 13:24:57.022876 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nr8z\" (UniqueName: \"kubernetes.io/projected/115dedcd-9ffe-408e-9b6b-878f53c6c252-kube-api-access-9nr8z\") pod \"nova-cell0-conductor-0\" (UID: \"115dedcd-9ffe-408e-9b6b-878f53c6c252\") " pod="openstack/nova-cell0-conductor-0" Feb 17 13:24:57 crc kubenswrapper[4955]: I0217 13:24:57.064062 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 17 13:24:57 crc kubenswrapper[4955]: I0217 13:24:57.496575 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 17 13:24:57 crc kubenswrapper[4955]: I0217 13:24:57.689442 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"115dedcd-9ffe-408e-9b6b-878f53c6c252","Type":"ContainerStarted","Data":"d1bc8358b1ef1671b70b5aae2261d56a3f40067fbd948a81eccabc5cee2435e9"} Feb 17 13:24:57 crc kubenswrapper[4955]: I0217 13:24:57.689485 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"115dedcd-9ffe-408e-9b6b-878f53c6c252","Type":"ContainerStarted","Data":"f279b65f4827b977324e1ca1fa151c8b6704e16da821d4c862656fbfd24578bb"} Feb 17 13:24:57 crc kubenswrapper[4955]: I0217 13:24:57.689773 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Feb 17 13:24:57 crc kubenswrapper[4955]: I0217 13:24:57.705949 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=1.705921874 podStartE2EDuration="1.705921874s" podCreationTimestamp="2026-02-17 13:24:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:24:57.70299218 +0000 UTC m=+1236.225721743" watchObservedRunningTime="2026-02-17 13:24:57.705921874 +0000 UTC m=+1236.228651427" Feb 17 13:24:57 crc kubenswrapper[4955]: I0217 13:24:57.783935 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 17 13:24:58 crc kubenswrapper[4955]: I0217 13:24:58.245970 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1d0ee97-a796-4b24-a77c-3626ac31256f" path="/var/lib/kubelet/pods/e1d0ee97-a796-4b24-a77c-3626ac31256f/volumes" Feb 17 13:25:01 crc kubenswrapper[4955]: I0217 13:25:01.046024 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 17 13:25:01 crc kubenswrapper[4955]: I0217 13:25:01.046463 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="c631bd4b-26ac-44e3-87aa-8878c42e6ce2" containerName="kube-state-metrics" containerID="cri-o://4a097e5724bf1f20c0f6234932ca8c7b9bcd1c3f18c46c0839ea88f05514bba0" gracePeriod=30 Feb 17 13:25:01 crc kubenswrapper[4955]: I0217 13:25:01.510401 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 17 13:25:01 crc kubenswrapper[4955]: I0217 13:25:01.599876 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9lnjg\" (UniqueName: \"kubernetes.io/projected/c631bd4b-26ac-44e3-87aa-8878c42e6ce2-kube-api-access-9lnjg\") pod \"c631bd4b-26ac-44e3-87aa-8878c42e6ce2\" (UID: \"c631bd4b-26ac-44e3-87aa-8878c42e6ce2\") " Feb 17 13:25:01 crc kubenswrapper[4955]: I0217 13:25:01.620515 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c631bd4b-26ac-44e3-87aa-8878c42e6ce2-kube-api-access-9lnjg" (OuterVolumeSpecName: "kube-api-access-9lnjg") pod "c631bd4b-26ac-44e3-87aa-8878c42e6ce2" (UID: "c631bd4b-26ac-44e3-87aa-8878c42e6ce2"). InnerVolumeSpecName "kube-api-access-9lnjg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:25:01 crc kubenswrapper[4955]: I0217 13:25:01.701462 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9lnjg\" (UniqueName: \"kubernetes.io/projected/c631bd4b-26ac-44e3-87aa-8878c42e6ce2-kube-api-access-9lnjg\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:01 crc kubenswrapper[4955]: I0217 13:25:01.742659 4955 generic.go:334] "Generic (PLEG): container finished" podID="c631bd4b-26ac-44e3-87aa-8878c42e6ce2" containerID="4a097e5724bf1f20c0f6234932ca8c7b9bcd1c3f18c46c0839ea88f05514bba0" exitCode=2 Feb 17 13:25:01 crc kubenswrapper[4955]: I0217 13:25:01.742706 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 17 13:25:01 crc kubenswrapper[4955]: I0217 13:25:01.742705 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c631bd4b-26ac-44e3-87aa-8878c42e6ce2","Type":"ContainerDied","Data":"4a097e5724bf1f20c0f6234932ca8c7b9bcd1c3f18c46c0839ea88f05514bba0"} Feb 17 13:25:01 crc kubenswrapper[4955]: I0217 13:25:01.743041 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c631bd4b-26ac-44e3-87aa-8878c42e6ce2","Type":"ContainerDied","Data":"a3bbf1e6807283776874cd38f2f3ab5172faa85f27dd4f7335fb13c561776de0"} Feb 17 13:25:01 crc kubenswrapper[4955]: I0217 13:25:01.743070 4955 scope.go:117] "RemoveContainer" containerID="4a097e5724bf1f20c0f6234932ca8c7b9bcd1c3f18c46c0839ea88f05514bba0" Feb 17 13:25:01 crc kubenswrapper[4955]: I0217 13:25:01.779486 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 17 13:25:01 crc kubenswrapper[4955]: I0217 13:25:01.805118 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 17 13:25:01 crc kubenswrapper[4955]: I0217 13:25:01.811112 4955 scope.go:117] "RemoveContainer" containerID="4a097e5724bf1f20c0f6234932ca8c7b9bcd1c3f18c46c0839ea88f05514bba0" Feb 17 13:25:01 crc kubenswrapper[4955]: E0217 13:25:01.811748 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a097e5724bf1f20c0f6234932ca8c7b9bcd1c3f18c46c0839ea88f05514bba0\": container with ID starting with 4a097e5724bf1f20c0f6234932ca8c7b9bcd1c3f18c46c0839ea88f05514bba0 not found: ID does not exist" containerID="4a097e5724bf1f20c0f6234932ca8c7b9bcd1c3f18c46c0839ea88f05514bba0" Feb 17 13:25:01 crc kubenswrapper[4955]: I0217 13:25:01.811793 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a097e5724bf1f20c0f6234932ca8c7b9bcd1c3f18c46c0839ea88f05514bba0"} err="failed to get container status \"4a097e5724bf1f20c0f6234932ca8c7b9bcd1c3f18c46c0839ea88f05514bba0\": rpc error: code = NotFound desc = could not find container \"4a097e5724bf1f20c0f6234932ca8c7b9bcd1c3f18c46c0839ea88f05514bba0\": container with ID starting with 4a097e5724bf1f20c0f6234932ca8c7b9bcd1c3f18c46c0839ea88f05514bba0 not found: ID does not exist" Feb 17 13:25:01 crc kubenswrapper[4955]: I0217 13:25:01.823900 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Feb 17 13:25:01 crc kubenswrapper[4955]: E0217 13:25:01.824698 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c631bd4b-26ac-44e3-87aa-8878c42e6ce2" containerName="kube-state-metrics" Feb 17 13:25:01 crc kubenswrapper[4955]: I0217 13:25:01.824724 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="c631bd4b-26ac-44e3-87aa-8878c42e6ce2" containerName="kube-state-metrics" Feb 17 13:25:01 crc kubenswrapper[4955]: I0217 13:25:01.824967 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="c631bd4b-26ac-44e3-87aa-8878c42e6ce2" containerName="kube-state-metrics" Feb 17 13:25:01 crc kubenswrapper[4955]: I0217 13:25:01.825555 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 17 13:25:01 crc kubenswrapper[4955]: I0217 13:25:01.827643 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Feb 17 13:25:01 crc kubenswrapper[4955]: I0217 13:25:01.827835 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Feb 17 13:25:01 crc kubenswrapper[4955]: I0217 13:25:01.846208 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 17 13:25:01 crc kubenswrapper[4955]: I0217 13:25:01.907609 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjnst\" (UniqueName: \"kubernetes.io/projected/3e1af961-6b57-4a77-ae1c-dcfacea5c8c6-kube-api-access-kjnst\") pod \"kube-state-metrics-0\" (UID: \"3e1af961-6b57-4a77-ae1c-dcfacea5c8c6\") " pod="openstack/kube-state-metrics-0" Feb 17 13:25:01 crc kubenswrapper[4955]: I0217 13:25:01.907742 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/3e1af961-6b57-4a77-ae1c-dcfacea5c8c6-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"3e1af961-6b57-4a77-ae1c-dcfacea5c8c6\") " pod="openstack/kube-state-metrics-0" Feb 17 13:25:01 crc kubenswrapper[4955]: I0217 13:25:01.907774 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/3e1af961-6b57-4a77-ae1c-dcfacea5c8c6-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"3e1af961-6b57-4a77-ae1c-dcfacea5c8c6\") " pod="openstack/kube-state-metrics-0" Feb 17 13:25:01 crc kubenswrapper[4955]: I0217 13:25:01.907925 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e1af961-6b57-4a77-ae1c-dcfacea5c8c6-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"3e1af961-6b57-4a77-ae1c-dcfacea5c8c6\") " pod="openstack/kube-state-metrics-0" Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.009110 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjnst\" (UniqueName: \"kubernetes.io/projected/3e1af961-6b57-4a77-ae1c-dcfacea5c8c6-kube-api-access-kjnst\") pod \"kube-state-metrics-0\" (UID: \"3e1af961-6b57-4a77-ae1c-dcfacea5c8c6\") " pod="openstack/kube-state-metrics-0" Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.009257 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/3e1af961-6b57-4a77-ae1c-dcfacea5c8c6-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"3e1af961-6b57-4a77-ae1c-dcfacea5c8c6\") " pod="openstack/kube-state-metrics-0" Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.009284 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/3e1af961-6b57-4a77-ae1c-dcfacea5c8c6-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"3e1af961-6b57-4a77-ae1c-dcfacea5c8c6\") " pod="openstack/kube-state-metrics-0" Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.009325 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e1af961-6b57-4a77-ae1c-dcfacea5c8c6-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"3e1af961-6b57-4a77-ae1c-dcfacea5c8c6\") " pod="openstack/kube-state-metrics-0" Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.014235 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e1af961-6b57-4a77-ae1c-dcfacea5c8c6-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"3e1af961-6b57-4a77-ae1c-dcfacea5c8c6\") " pod="openstack/kube-state-metrics-0" Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.014387 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/3e1af961-6b57-4a77-ae1c-dcfacea5c8c6-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"3e1af961-6b57-4a77-ae1c-dcfacea5c8c6\") " pod="openstack/kube-state-metrics-0" Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.015485 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/3e1af961-6b57-4a77-ae1c-dcfacea5c8c6-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"3e1af961-6b57-4a77-ae1c-dcfacea5c8c6\") " pod="openstack/kube-state-metrics-0" Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.029387 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjnst\" (UniqueName: \"kubernetes.io/projected/3e1af961-6b57-4a77-ae1c-dcfacea5c8c6-kube-api-access-kjnst\") pod \"kube-state-metrics-0\" (UID: \"3e1af961-6b57-4a77-ae1c-dcfacea5c8c6\") " pod="openstack/kube-state-metrics-0" Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.095325 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.146680 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.247064 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c631bd4b-26ac-44e3-87aa-8878c42e6ce2" path="/var/lib/kubelet/pods/c631bd4b-26ac-44e3-87aa-8878c42e6ce2/volumes" Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.571824 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-4kqpw"] Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.573015 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-4kqpw" Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.578258 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.578563 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.586083 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-4kqpw"] Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.621595 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41b91660-a2ac-4f9e-8315-5de36243f161-config-data\") pod \"nova-cell0-cell-mapping-4kqpw\" (UID: \"41b91660-a2ac-4f9e-8315-5de36243f161\") " pod="openstack/nova-cell0-cell-mapping-4kqpw" Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.621967 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41b91660-a2ac-4f9e-8315-5de36243f161-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-4kqpw\" (UID: \"41b91660-a2ac-4f9e-8315-5de36243f161\") " pod="openstack/nova-cell0-cell-mapping-4kqpw" Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.622043 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mjpr\" (UniqueName: \"kubernetes.io/projected/41b91660-a2ac-4f9e-8315-5de36243f161-kube-api-access-4mjpr\") pod \"nova-cell0-cell-mapping-4kqpw\" (UID: \"41b91660-a2ac-4f9e-8315-5de36243f161\") " pod="openstack/nova-cell0-cell-mapping-4kqpw" Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.622271 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41b91660-a2ac-4f9e-8315-5de36243f161-scripts\") pod \"nova-cell0-cell-mapping-4kqpw\" (UID: \"41b91660-a2ac-4f9e-8315-5de36243f161\") " pod="openstack/nova-cell0-cell-mapping-4kqpw" Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.642909 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.724023 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mjpr\" (UniqueName: \"kubernetes.io/projected/41b91660-a2ac-4f9e-8315-5de36243f161-kube-api-access-4mjpr\") pod \"nova-cell0-cell-mapping-4kqpw\" (UID: \"41b91660-a2ac-4f9e-8315-5de36243f161\") " pod="openstack/nova-cell0-cell-mapping-4kqpw" Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.724274 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41b91660-a2ac-4f9e-8315-5de36243f161-scripts\") pod \"nova-cell0-cell-mapping-4kqpw\" (UID: \"41b91660-a2ac-4f9e-8315-5de36243f161\") " pod="openstack/nova-cell0-cell-mapping-4kqpw" Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.724451 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41b91660-a2ac-4f9e-8315-5de36243f161-config-data\") pod \"nova-cell0-cell-mapping-4kqpw\" (UID: \"41b91660-a2ac-4f9e-8315-5de36243f161\") " pod="openstack/nova-cell0-cell-mapping-4kqpw" Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.724494 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41b91660-a2ac-4f9e-8315-5de36243f161-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-4kqpw\" (UID: \"41b91660-a2ac-4f9e-8315-5de36243f161\") " pod="openstack/nova-cell0-cell-mapping-4kqpw" Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.738949 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41b91660-a2ac-4f9e-8315-5de36243f161-config-data\") pod \"nova-cell0-cell-mapping-4kqpw\" (UID: \"41b91660-a2ac-4f9e-8315-5de36243f161\") " pod="openstack/nova-cell0-cell-mapping-4kqpw" Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.742172 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41b91660-a2ac-4f9e-8315-5de36243f161-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-4kqpw\" (UID: \"41b91660-a2ac-4f9e-8315-5de36243f161\") " pod="openstack/nova-cell0-cell-mapping-4kqpw" Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.744133 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41b91660-a2ac-4f9e-8315-5de36243f161-scripts\") pod \"nova-cell0-cell-mapping-4kqpw\" (UID: \"41b91660-a2ac-4f9e-8315-5de36243f161\") " pod="openstack/nova-cell0-cell-mapping-4kqpw" Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.772088 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3e1af961-6b57-4a77-ae1c-dcfacea5c8c6","Type":"ContainerStarted","Data":"bd1b0b2b64e68a42227a59ccf0541f9bf1fe1db8fd41f2717c7501b1ec385e93"} Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.802054 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.803411 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.806684 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.814413 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mjpr\" (UniqueName: \"kubernetes.io/projected/41b91660-a2ac-4f9e-8315-5de36243f161-kube-api-access-4mjpr\") pod \"nova-cell0-cell-mapping-4kqpw\" (UID: \"41b91660-a2ac-4f9e-8315-5de36243f161\") " pod="openstack/nova-cell0-cell-mapping-4kqpw" Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.828343 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5\") " pod="openstack/nova-scheduler-0" Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.828392 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5-config-data\") pod \"nova-scheduler-0\" (UID: \"01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5\") " pod="openstack/nova-scheduler-0" Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.828480 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tktw7\" (UniqueName: \"kubernetes.io/projected/01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5-kube-api-access-tktw7\") pod \"nova-scheduler-0\" (UID: \"01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5\") " pod="openstack/nova-scheduler-0" Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.839385 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.899658 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-4kqpw" Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.933659 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tktw7\" (UniqueName: \"kubernetes.io/projected/01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5-kube-api-access-tktw7\") pod \"nova-scheduler-0\" (UID: \"01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5\") " pod="openstack/nova-scheduler-0" Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.934067 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5\") " pod="openstack/nova-scheduler-0" Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.934179 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5-config-data\") pod \"nova-scheduler-0\" (UID: \"01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5\") " pod="openstack/nova-scheduler-0" Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.953805 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.961996 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.973240 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.982063 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.982610 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tktw7\" (UniqueName: \"kubernetes.io/projected/01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5-kube-api-access-tktw7\") pod \"nova-scheduler-0\" (UID: \"01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5\") " pod="openstack/nova-scheduler-0" Feb 17 13:25:02 crc kubenswrapper[4955]: I0217 13:25:02.993395 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5\") " pod="openstack/nova-scheduler-0" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.020681 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.035854 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7a1ed55-8add-4f6d-8210-76ccf42a58ab-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c7a1ed55-8add-4f6d-8210-76ccf42a58ab\") " pod="openstack/nova-metadata-0" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.035926 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cp66t\" (UniqueName: \"kubernetes.io/projected/c7a1ed55-8add-4f6d-8210-76ccf42a58ab-kube-api-access-cp66t\") pod \"nova-metadata-0\" (UID: \"c7a1ed55-8add-4f6d-8210-76ccf42a58ab\") " pod="openstack/nova-metadata-0" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.035980 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7a1ed55-8add-4f6d-8210-76ccf42a58ab-config-data\") pod \"nova-metadata-0\" (UID: \"c7a1ed55-8add-4f6d-8210-76ccf42a58ab\") " pod="openstack/nova-metadata-0" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.036030 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c7a1ed55-8add-4f6d-8210-76ccf42a58ab-logs\") pod \"nova-metadata-0\" (UID: \"c7a1ed55-8add-4f6d-8210-76ccf42a58ab\") " pod="openstack/nova-metadata-0" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.038270 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5-config-data\") pod \"nova-scheduler-0\" (UID: \"01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5\") " pod="openstack/nova-scheduler-0" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.047765 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.051395 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.065536 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.089558 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-b6cbh"] Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.091089 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-b6cbh" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.104628 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.106702 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.114482 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.137326 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7a1ed55-8add-4f6d-8210-76ccf42a58ab-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c7a1ed55-8add-4f6d-8210-76ccf42a58ab\") " pod="openstack/nova-metadata-0" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.137387 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/242f0cab-eded-4013-8ed0-f9ddd4de5fd0-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"242f0cab-eded-4013-8ed0-f9ddd4de5fd0\") " pod="openstack/nova-cell1-novncproxy-0" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.137473 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cp66t\" (UniqueName: \"kubernetes.io/projected/c7a1ed55-8add-4f6d-8210-76ccf42a58ab-kube-api-access-cp66t\") pod \"nova-metadata-0\" (UID: \"c7a1ed55-8add-4f6d-8210-76ccf42a58ab\") " pod="openstack/nova-metadata-0" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.137562 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7a1ed55-8add-4f6d-8210-76ccf42a58ab-config-data\") pod \"nova-metadata-0\" (UID: \"c7a1ed55-8add-4f6d-8210-76ccf42a58ab\") " pod="openstack/nova-metadata-0" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.137632 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c7a1ed55-8add-4f6d-8210-76ccf42a58ab-logs\") pod \"nova-metadata-0\" (UID: \"c7a1ed55-8add-4f6d-8210-76ccf42a58ab\") " pod="openstack/nova-metadata-0" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.137674 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/242f0cab-eded-4013-8ed0-f9ddd4de5fd0-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"242f0cab-eded-4013-8ed0-f9ddd4de5fd0\") " pod="openstack/nova-cell1-novncproxy-0" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.137705 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrtph\" (UniqueName: \"kubernetes.io/projected/242f0cab-eded-4013-8ed0-f9ddd4de5fd0-kube-api-access-nrtph\") pod \"nova-cell1-novncproxy-0\" (UID: \"242f0cab-eded-4013-8ed0-f9ddd4de5fd0\") " pod="openstack/nova-cell1-novncproxy-0" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.138241 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c7a1ed55-8add-4f6d-8210-76ccf42a58ab-logs\") pod \"nova-metadata-0\" (UID: \"c7a1ed55-8add-4f6d-8210-76ccf42a58ab\") " pod="openstack/nova-metadata-0" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.140722 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.145439 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7a1ed55-8add-4f6d-8210-76ccf42a58ab-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c7a1ed55-8add-4f6d-8210-76ccf42a58ab\") " pod="openstack/nova-metadata-0" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.154204 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-b6cbh"] Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.163882 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cp66t\" (UniqueName: \"kubernetes.io/projected/c7a1ed55-8add-4f6d-8210-76ccf42a58ab-kube-api-access-cp66t\") pod \"nova-metadata-0\" (UID: \"c7a1ed55-8add-4f6d-8210-76ccf42a58ab\") " pod="openstack/nova-metadata-0" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.166752 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7a1ed55-8add-4f6d-8210-76ccf42a58ab-config-data\") pod \"nova-metadata-0\" (UID: \"c7a1ed55-8add-4f6d-8210-76ccf42a58ab\") " pod="openstack/nova-metadata-0" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.195236 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.243037 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrtph\" (UniqueName: \"kubernetes.io/projected/242f0cab-eded-4013-8ed0-f9ddd4de5fd0-kube-api-access-nrtph\") pod \"nova-cell1-novncproxy-0\" (UID: \"242f0cab-eded-4013-8ed0-f9ddd4de5fd0\") " pod="openstack/nova-cell1-novncproxy-0" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.243105 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnh22\" (UniqueName: \"kubernetes.io/projected/987ec26b-c18e-47ec-be15-24059d88e961-kube-api-access-qnh22\") pod \"dnsmasq-dns-bccf8f775-b6cbh\" (UID: \"987ec26b-c18e-47ec-be15-24059d88e961\") " pod="openstack/dnsmasq-dns-bccf8f775-b6cbh" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.243139 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f69d681-27e5-4fb9-a199-3c6847c80dc9-config-data\") pod \"nova-api-0\" (UID: \"4f69d681-27e5-4fb9-a199-3c6847c80dc9\") " pod="openstack/nova-api-0" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.243158 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/987ec26b-c18e-47ec-be15-24059d88e961-config\") pod \"dnsmasq-dns-bccf8f775-b6cbh\" (UID: \"987ec26b-c18e-47ec-be15-24059d88e961\") " pod="openstack/dnsmasq-dns-bccf8f775-b6cbh" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.243173 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/987ec26b-c18e-47ec-be15-24059d88e961-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-b6cbh\" (UID: \"987ec26b-c18e-47ec-be15-24059d88e961\") " pod="openstack/dnsmasq-dns-bccf8f775-b6cbh" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.243192 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/242f0cab-eded-4013-8ed0-f9ddd4de5fd0-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"242f0cab-eded-4013-8ed0-f9ddd4de5fd0\") " pod="openstack/nova-cell1-novncproxy-0" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.243323 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/987ec26b-c18e-47ec-be15-24059d88e961-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-b6cbh\" (UID: \"987ec26b-c18e-47ec-be15-24059d88e961\") " pod="openstack/dnsmasq-dns-bccf8f775-b6cbh" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.243442 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f69d681-27e5-4fb9-a199-3c6847c80dc9-logs\") pod \"nova-api-0\" (UID: \"4f69d681-27e5-4fb9-a199-3c6847c80dc9\") " pod="openstack/nova-api-0" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.243563 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/987ec26b-c18e-47ec-be15-24059d88e961-dns-svc\") pod \"dnsmasq-dns-bccf8f775-b6cbh\" (UID: \"987ec26b-c18e-47ec-be15-24059d88e961\") " pod="openstack/dnsmasq-dns-bccf8f775-b6cbh" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.243683 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/987ec26b-c18e-47ec-be15-24059d88e961-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-b6cbh\" (UID: \"987ec26b-c18e-47ec-be15-24059d88e961\") " pod="openstack/dnsmasq-dns-bccf8f775-b6cbh" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.243820 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f69d681-27e5-4fb9-a199-3c6847c80dc9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4f69d681-27e5-4fb9-a199-3c6847c80dc9\") " pod="openstack/nova-api-0" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.243845 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxmm7\" (UniqueName: \"kubernetes.io/projected/4f69d681-27e5-4fb9-a199-3c6847c80dc9-kube-api-access-fxmm7\") pod \"nova-api-0\" (UID: \"4f69d681-27e5-4fb9-a199-3c6847c80dc9\") " pod="openstack/nova-api-0" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.243895 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/242f0cab-eded-4013-8ed0-f9ddd4de5fd0-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"242f0cab-eded-4013-8ed0-f9ddd4de5fd0\") " pod="openstack/nova-cell1-novncproxy-0" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.251054 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/242f0cab-eded-4013-8ed0-f9ddd4de5fd0-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"242f0cab-eded-4013-8ed0-f9ddd4de5fd0\") " pod="openstack/nova-cell1-novncproxy-0" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.261652 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/242f0cab-eded-4013-8ed0-f9ddd4de5fd0-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"242f0cab-eded-4013-8ed0-f9ddd4de5fd0\") " pod="openstack/nova-cell1-novncproxy-0" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.279568 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrtph\" (UniqueName: \"kubernetes.io/projected/242f0cab-eded-4013-8ed0-f9ddd4de5fd0-kube-api-access-nrtph\") pod \"nova-cell1-novncproxy-0\" (UID: \"242f0cab-eded-4013-8ed0-f9ddd4de5fd0\") " pod="openstack/nova-cell1-novncproxy-0" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.328240 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.329838 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.330114 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7389a7d7-83cb-4c0d-bb69-030c3331caf5" containerName="ceilometer-central-agent" containerID="cri-o://abc6b8d601853195caab008f3b8e84b0ccbedfad1fc886969c8c98835f077d7f" gracePeriod=30 Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.330564 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7389a7d7-83cb-4c0d-bb69-030c3331caf5" containerName="proxy-httpd" containerID="cri-o://33c70cbe6a16ffe2997b72bfe5ffe17ac36b209090875f6c32429fbdde279ca7" gracePeriod=30 Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.330615 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7389a7d7-83cb-4c0d-bb69-030c3331caf5" containerName="sg-core" containerID="cri-o://9052f1507175292366564a8ac671e92809a029127e38cbd05a33f28841be19d9" gracePeriod=30 Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.330667 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7389a7d7-83cb-4c0d-bb69-030c3331caf5" containerName="ceilometer-notification-agent" containerID="cri-o://a5fa3a4ed0a0ff395c5c44ad2e45b52e16a7527b29765548b1e51df776873468" gracePeriod=30 Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.350144 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/987ec26b-c18e-47ec-be15-24059d88e961-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-b6cbh\" (UID: \"987ec26b-c18e-47ec-be15-24059d88e961\") " pod="openstack/dnsmasq-dns-bccf8f775-b6cbh" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.350220 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f69d681-27e5-4fb9-a199-3c6847c80dc9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4f69d681-27e5-4fb9-a199-3c6847c80dc9\") " pod="openstack/nova-api-0" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.350242 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxmm7\" (UniqueName: \"kubernetes.io/projected/4f69d681-27e5-4fb9-a199-3c6847c80dc9-kube-api-access-fxmm7\") pod \"nova-api-0\" (UID: \"4f69d681-27e5-4fb9-a199-3c6847c80dc9\") " pod="openstack/nova-api-0" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.350320 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnh22\" (UniqueName: \"kubernetes.io/projected/987ec26b-c18e-47ec-be15-24059d88e961-kube-api-access-qnh22\") pod \"dnsmasq-dns-bccf8f775-b6cbh\" (UID: \"987ec26b-c18e-47ec-be15-24059d88e961\") " pod="openstack/dnsmasq-dns-bccf8f775-b6cbh" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.350354 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f69d681-27e5-4fb9-a199-3c6847c80dc9-config-data\") pod \"nova-api-0\" (UID: \"4f69d681-27e5-4fb9-a199-3c6847c80dc9\") " pod="openstack/nova-api-0" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.350380 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/987ec26b-c18e-47ec-be15-24059d88e961-config\") pod \"dnsmasq-dns-bccf8f775-b6cbh\" (UID: \"987ec26b-c18e-47ec-be15-24059d88e961\") " pod="openstack/dnsmasq-dns-bccf8f775-b6cbh" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.350400 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/987ec26b-c18e-47ec-be15-24059d88e961-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-b6cbh\" (UID: \"987ec26b-c18e-47ec-be15-24059d88e961\") " pod="openstack/dnsmasq-dns-bccf8f775-b6cbh" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.350451 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/987ec26b-c18e-47ec-be15-24059d88e961-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-b6cbh\" (UID: \"987ec26b-c18e-47ec-be15-24059d88e961\") " pod="openstack/dnsmasq-dns-bccf8f775-b6cbh" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.350499 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f69d681-27e5-4fb9-a199-3c6847c80dc9-logs\") pod \"nova-api-0\" (UID: \"4f69d681-27e5-4fb9-a199-3c6847c80dc9\") " pod="openstack/nova-api-0" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.350528 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/987ec26b-c18e-47ec-be15-24059d88e961-dns-svc\") pod \"dnsmasq-dns-bccf8f775-b6cbh\" (UID: \"987ec26b-c18e-47ec-be15-24059d88e961\") " pod="openstack/dnsmasq-dns-bccf8f775-b6cbh" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.350726 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.353066 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/987ec26b-c18e-47ec-be15-24059d88e961-config\") pod \"dnsmasq-dns-bccf8f775-b6cbh\" (UID: \"987ec26b-c18e-47ec-be15-24059d88e961\") " pod="openstack/dnsmasq-dns-bccf8f775-b6cbh" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.353152 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f69d681-27e5-4fb9-a199-3c6847c80dc9-logs\") pod \"nova-api-0\" (UID: \"4f69d681-27e5-4fb9-a199-3c6847c80dc9\") " pod="openstack/nova-api-0" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.353320 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/987ec26b-c18e-47ec-be15-24059d88e961-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-b6cbh\" (UID: \"987ec26b-c18e-47ec-be15-24059d88e961\") " pod="openstack/dnsmasq-dns-bccf8f775-b6cbh" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.353367 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/987ec26b-c18e-47ec-be15-24059d88e961-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-b6cbh\" (UID: \"987ec26b-c18e-47ec-be15-24059d88e961\") " pod="openstack/dnsmasq-dns-bccf8f775-b6cbh" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.354565 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/987ec26b-c18e-47ec-be15-24059d88e961-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-b6cbh\" (UID: \"987ec26b-c18e-47ec-be15-24059d88e961\") " pod="openstack/dnsmasq-dns-bccf8f775-b6cbh" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.366079 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/987ec26b-c18e-47ec-be15-24059d88e961-dns-svc\") pod \"dnsmasq-dns-bccf8f775-b6cbh\" (UID: \"987ec26b-c18e-47ec-be15-24059d88e961\") " pod="openstack/dnsmasq-dns-bccf8f775-b6cbh" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.371521 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f69d681-27e5-4fb9-a199-3c6847c80dc9-config-data\") pod \"nova-api-0\" (UID: \"4f69d681-27e5-4fb9-a199-3c6847c80dc9\") " pod="openstack/nova-api-0" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.375541 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f69d681-27e5-4fb9-a199-3c6847c80dc9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4f69d681-27e5-4fb9-a199-3c6847c80dc9\") " pod="openstack/nova-api-0" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.386032 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxmm7\" (UniqueName: \"kubernetes.io/projected/4f69d681-27e5-4fb9-a199-3c6847c80dc9-kube-api-access-fxmm7\") pod \"nova-api-0\" (UID: \"4f69d681-27e5-4fb9-a199-3c6847c80dc9\") " pod="openstack/nova-api-0" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.392339 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnh22\" (UniqueName: \"kubernetes.io/projected/987ec26b-c18e-47ec-be15-24059d88e961-kube-api-access-qnh22\") pod \"dnsmasq-dns-bccf8f775-b6cbh\" (UID: \"987ec26b-c18e-47ec-be15-24059d88e961\") " pod="openstack/dnsmasq-dns-bccf8f775-b6cbh" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.430285 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-b6cbh" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.455335 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.545772 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-4kqpw"] Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.660634 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-mzzkq"] Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.666913 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-mzzkq" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.673509 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.674539 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.689462 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-mzzkq"] Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.757841 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c40c5ab-a929-4850-9a70-b98106008500-config-data\") pod \"nova-cell1-conductor-db-sync-mzzkq\" (UID: \"0c40c5ab-a929-4850-9a70-b98106008500\") " pod="openstack/nova-cell1-conductor-db-sync-mzzkq" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.758214 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c40c5ab-a929-4850-9a70-b98106008500-scripts\") pod \"nova-cell1-conductor-db-sync-mzzkq\" (UID: \"0c40c5ab-a929-4850-9a70-b98106008500\") " pod="openstack/nova-cell1-conductor-db-sync-mzzkq" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.758275 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4nnq\" (UniqueName: \"kubernetes.io/projected/0c40c5ab-a929-4850-9a70-b98106008500-kube-api-access-c4nnq\") pod \"nova-cell1-conductor-db-sync-mzzkq\" (UID: \"0c40c5ab-a929-4850-9a70-b98106008500\") " pod="openstack/nova-cell1-conductor-db-sync-mzzkq" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.758326 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c40c5ab-a929-4850-9a70-b98106008500-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-mzzkq\" (UID: \"0c40c5ab-a929-4850-9a70-b98106008500\") " pod="openstack/nova-cell1-conductor-db-sync-mzzkq" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.791141 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 17 13:25:03 crc kubenswrapper[4955]: W0217 13:25:03.798241 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01100c9a_bf1b_4d2a_bc47_c6f0090ebaa5.slice/crio-a2e27ba89ac3e2b41e320af3aada6f6083193b336df082aa0219d559bdedafbd WatchSource:0}: Error finding container a2e27ba89ac3e2b41e320af3aada6f6083193b336df082aa0219d559bdedafbd: Status 404 returned error can't find the container with id a2e27ba89ac3e2b41e320af3aada6f6083193b336df082aa0219d559bdedafbd Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.805820 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-4kqpw" event={"ID":"41b91660-a2ac-4f9e-8315-5de36243f161","Type":"ContainerStarted","Data":"056c3f45c2a20fe4fc964fcad05f11f1ac3f6c60aa5c9a4c3a46aeccfb0bc0eb"} Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.810798 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3e1af961-6b57-4a77-ae1c-dcfacea5c8c6","Type":"ContainerStarted","Data":"29e99de2213a347546cf0858c807a76cbdae6dabd985cdda457628bdfd9cddd3"} Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.811000 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.818379 4955 generic.go:334] "Generic (PLEG): container finished" podID="7389a7d7-83cb-4c0d-bb69-030c3331caf5" containerID="33c70cbe6a16ffe2997b72bfe5ffe17ac36b209090875f6c32429fbdde279ca7" exitCode=0 Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.818420 4955 generic.go:334] "Generic (PLEG): container finished" podID="7389a7d7-83cb-4c0d-bb69-030c3331caf5" containerID="9052f1507175292366564a8ac671e92809a029127e38cbd05a33f28841be19d9" exitCode=2 Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.818425 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7389a7d7-83cb-4c0d-bb69-030c3331caf5","Type":"ContainerDied","Data":"33c70cbe6a16ffe2997b72bfe5ffe17ac36b209090875f6c32429fbdde279ca7"} Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.818458 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7389a7d7-83cb-4c0d-bb69-030c3331caf5","Type":"ContainerDied","Data":"9052f1507175292366564a8ac671e92809a029127e38cbd05a33f28841be19d9"} Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.843172 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.34806832 podStartE2EDuration="2.843156249s" podCreationTimestamp="2026-02-17 13:25:01 +0000 UTC" firstStartedPulling="2026-02-17 13:25:02.646229513 +0000 UTC m=+1241.168959056" lastFinishedPulling="2026-02-17 13:25:03.141317442 +0000 UTC m=+1241.664046985" observedRunningTime="2026-02-17 13:25:03.829663482 +0000 UTC m=+1242.352393035" watchObservedRunningTime="2026-02-17 13:25:03.843156249 +0000 UTC m=+1242.365885782" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.863121 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c40c5ab-a929-4850-9a70-b98106008500-config-data\") pod \"nova-cell1-conductor-db-sync-mzzkq\" (UID: \"0c40c5ab-a929-4850-9a70-b98106008500\") " pod="openstack/nova-cell1-conductor-db-sync-mzzkq" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.863174 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c40c5ab-a929-4850-9a70-b98106008500-scripts\") pod \"nova-cell1-conductor-db-sync-mzzkq\" (UID: \"0c40c5ab-a929-4850-9a70-b98106008500\") " pod="openstack/nova-cell1-conductor-db-sync-mzzkq" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.863237 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4nnq\" (UniqueName: \"kubernetes.io/projected/0c40c5ab-a929-4850-9a70-b98106008500-kube-api-access-c4nnq\") pod \"nova-cell1-conductor-db-sync-mzzkq\" (UID: \"0c40c5ab-a929-4850-9a70-b98106008500\") " pod="openstack/nova-cell1-conductor-db-sync-mzzkq" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.863283 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c40c5ab-a929-4850-9a70-b98106008500-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-mzzkq\" (UID: \"0c40c5ab-a929-4850-9a70-b98106008500\") " pod="openstack/nova-cell1-conductor-db-sync-mzzkq" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.879485 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c40c5ab-a929-4850-9a70-b98106008500-scripts\") pod \"nova-cell1-conductor-db-sync-mzzkq\" (UID: \"0c40c5ab-a929-4850-9a70-b98106008500\") " pod="openstack/nova-cell1-conductor-db-sync-mzzkq" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.879986 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c40c5ab-a929-4850-9a70-b98106008500-config-data\") pod \"nova-cell1-conductor-db-sync-mzzkq\" (UID: \"0c40c5ab-a929-4850-9a70-b98106008500\") " pod="openstack/nova-cell1-conductor-db-sync-mzzkq" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.882613 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c40c5ab-a929-4850-9a70-b98106008500-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-mzzkq\" (UID: \"0c40c5ab-a929-4850-9a70-b98106008500\") " pod="openstack/nova-cell1-conductor-db-sync-mzzkq" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.883061 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4nnq\" (UniqueName: \"kubernetes.io/projected/0c40c5ab-a929-4850-9a70-b98106008500-kube-api-access-c4nnq\") pod \"nova-cell1-conductor-db-sync-mzzkq\" (UID: \"0c40c5ab-a929-4850-9a70-b98106008500\") " pod="openstack/nova-cell1-conductor-db-sync-mzzkq" Feb 17 13:25:03 crc kubenswrapper[4955]: I0217 13:25:03.961388 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 17 13:25:03 crc kubenswrapper[4955]: W0217 13:25:03.991119 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod242f0cab_eded_4013_8ed0_f9ddd4de5fd0.slice/crio-73b19a13056adeea2860efe3a1fa16447b7dcb9ffd294f69964d6b214a0db361 WatchSource:0}: Error finding container 73b19a13056adeea2860efe3a1fa16447b7dcb9ffd294f69964d6b214a0db361: Status 404 returned error can't find the container with id 73b19a13056adeea2860efe3a1fa16447b7dcb9ffd294f69964d6b214a0db361 Feb 17 13:25:04 crc kubenswrapper[4955]: I0217 13:25:03.998592 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-mzzkq" Feb 17 13:25:04 crc kubenswrapper[4955]: I0217 13:25:04.009427 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 17 13:25:04 crc kubenswrapper[4955]: I0217 13:25:04.133860 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-b6cbh"] Feb 17 13:25:04 crc kubenswrapper[4955]: I0217 13:25:04.273417 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 17 13:25:04 crc kubenswrapper[4955]: I0217 13:25:04.622513 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-mzzkq"] Feb 17 13:25:04 crc kubenswrapper[4955]: I0217 13:25:04.842748 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-4kqpw" event={"ID":"41b91660-a2ac-4f9e-8315-5de36243f161","Type":"ContainerStarted","Data":"53bb653afbe4c0a45289c67de05edbd4e34d900f6ef49832c55c7b0d6452eeeb"} Feb 17 13:25:04 crc kubenswrapper[4955]: I0217 13:25:04.847549 4955 generic.go:334] "Generic (PLEG): container finished" podID="987ec26b-c18e-47ec-be15-24059d88e961" containerID="d0e0c44eab047bf10331abf17b10ea10e7863a066569ee44f2da418f0d8edb45" exitCode=0 Feb 17 13:25:04 crc kubenswrapper[4955]: I0217 13:25:04.847624 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-b6cbh" event={"ID":"987ec26b-c18e-47ec-be15-24059d88e961","Type":"ContainerDied","Data":"d0e0c44eab047bf10331abf17b10ea10e7863a066569ee44f2da418f0d8edb45"} Feb 17 13:25:04 crc kubenswrapper[4955]: I0217 13:25:04.847653 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-b6cbh" event={"ID":"987ec26b-c18e-47ec-be15-24059d88e961","Type":"ContainerStarted","Data":"0fe99668d6bc86a35848edfc3878bb6e957baf6ddd91c7e5be544e3d77f889bd"} Feb 17 13:25:04 crc kubenswrapper[4955]: I0217 13:25:04.856032 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-mzzkq" event={"ID":"0c40c5ab-a929-4850-9a70-b98106008500","Type":"ContainerStarted","Data":"caec95bd40473db1d3da799ea2fdbaecfbec1549504dd90087cb326c218fac76"} Feb 17 13:25:04 crc kubenswrapper[4955]: I0217 13:25:04.863633 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-4kqpw" podStartSLOduration=2.863614064 podStartE2EDuration="2.863614064s" podCreationTimestamp="2026-02-17 13:25:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:25:04.860867195 +0000 UTC m=+1243.383596738" watchObservedRunningTime="2026-02-17 13:25:04.863614064 +0000 UTC m=+1243.386343607" Feb 17 13:25:04 crc kubenswrapper[4955]: I0217 13:25:04.865657 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5","Type":"ContainerStarted","Data":"a2e27ba89ac3e2b41e320af3aada6f6083193b336df082aa0219d559bdedafbd"} Feb 17 13:25:04 crc kubenswrapper[4955]: I0217 13:25:04.872070 4955 generic.go:334] "Generic (PLEG): container finished" podID="7389a7d7-83cb-4c0d-bb69-030c3331caf5" containerID="abc6b8d601853195caab008f3b8e84b0ccbedfad1fc886969c8c98835f077d7f" exitCode=0 Feb 17 13:25:04 crc kubenswrapper[4955]: I0217 13:25:04.872134 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7389a7d7-83cb-4c0d-bb69-030c3331caf5","Type":"ContainerDied","Data":"abc6b8d601853195caab008f3b8e84b0ccbedfad1fc886969c8c98835f077d7f"} Feb 17 13:25:04 crc kubenswrapper[4955]: I0217 13:25:04.876580 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4f69d681-27e5-4fb9-a199-3c6847c80dc9","Type":"ContainerStarted","Data":"51bd1d076f26064ebc8811e6d84ef38c06c2ee7b90434a8de2ffaeb25929b45b"} Feb 17 13:25:04 crc kubenswrapper[4955]: I0217 13:25:04.891469 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"242f0cab-eded-4013-8ed0-f9ddd4de5fd0","Type":"ContainerStarted","Data":"73b19a13056adeea2860efe3a1fa16447b7dcb9ffd294f69964d6b214a0db361"} Feb 17 13:25:04 crc kubenswrapper[4955]: I0217 13:25:04.895737 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c7a1ed55-8add-4f6d-8210-76ccf42a58ab","Type":"ContainerStarted","Data":"281d1febd28a5c3d81367f3c198ee3b0972633283889f94f1d27281caf057e02"} Feb 17 13:25:05 crc kubenswrapper[4955]: I0217 13:25:05.912321 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-b6cbh" event={"ID":"987ec26b-c18e-47ec-be15-24059d88e961","Type":"ContainerStarted","Data":"4513616fbc9ee9a8b58471bb35d22e68a4a066e423e406eada0ad59580bc24f2"} Feb 17 13:25:05 crc kubenswrapper[4955]: I0217 13:25:05.912657 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-bccf8f775-b6cbh" Feb 17 13:25:05 crc kubenswrapper[4955]: I0217 13:25:05.915473 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-mzzkq" event={"ID":"0c40c5ab-a929-4850-9a70-b98106008500","Type":"ContainerStarted","Data":"7fbd0bf75c3a27db6ce7417c4a97396f46f96609a3d031db30a4adcab8f30fd3"} Feb 17 13:25:05 crc kubenswrapper[4955]: I0217 13:25:05.939357 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-bccf8f775-b6cbh" podStartSLOduration=3.939337365 podStartE2EDuration="3.939337365s" podCreationTimestamp="2026-02-17 13:25:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:25:05.933177518 +0000 UTC m=+1244.455907081" watchObservedRunningTime="2026-02-17 13:25:05.939337365 +0000 UTC m=+1244.462066908" Feb 17 13:25:05 crc kubenswrapper[4955]: I0217 13:25:05.956656 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-mzzkq" podStartSLOduration=2.956635721 podStartE2EDuration="2.956635721s" podCreationTimestamp="2026-02-17 13:25:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:25:05.948729694 +0000 UTC m=+1244.471459237" watchObservedRunningTime="2026-02-17 13:25:05.956635721 +0000 UTC m=+1244.479365264" Feb 17 13:25:05 crc kubenswrapper[4955]: I0217 13:25:05.957251 4955 generic.go:334] "Generic (PLEG): container finished" podID="7389a7d7-83cb-4c0d-bb69-030c3331caf5" containerID="a5fa3a4ed0a0ff395c5c44ad2e45b52e16a7527b29765548b1e51df776873468" exitCode=0 Feb 17 13:25:05 crc kubenswrapper[4955]: I0217 13:25:05.958047 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7389a7d7-83cb-4c0d-bb69-030c3331caf5","Type":"ContainerDied","Data":"a5fa3a4ed0a0ff395c5c44ad2e45b52e16a7527b29765548b1e51df776873468"} Feb 17 13:25:06 crc kubenswrapper[4955]: I0217 13:25:06.063423 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 17 13:25:06 crc kubenswrapper[4955]: I0217 13:25:06.085553 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 17 13:25:07 crc kubenswrapper[4955]: I0217 13:25:07.103496 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 17 13:25:07 crc kubenswrapper[4955]: I0217 13:25:07.150909 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7389a7d7-83cb-4c0d-bb69-030c3331caf5-sg-core-conf-yaml\") pod \"7389a7d7-83cb-4c0d-bb69-030c3331caf5\" (UID: \"7389a7d7-83cb-4c0d-bb69-030c3331caf5\") " Feb 17 13:25:07 crc kubenswrapper[4955]: I0217 13:25:07.150961 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhm99\" (UniqueName: \"kubernetes.io/projected/7389a7d7-83cb-4c0d-bb69-030c3331caf5-kube-api-access-vhm99\") pod \"7389a7d7-83cb-4c0d-bb69-030c3331caf5\" (UID: \"7389a7d7-83cb-4c0d-bb69-030c3331caf5\") " Feb 17 13:25:07 crc kubenswrapper[4955]: I0217 13:25:07.151000 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7389a7d7-83cb-4c0d-bb69-030c3331caf5-run-httpd\") pod \"7389a7d7-83cb-4c0d-bb69-030c3331caf5\" (UID: \"7389a7d7-83cb-4c0d-bb69-030c3331caf5\") " Feb 17 13:25:07 crc kubenswrapper[4955]: I0217 13:25:07.151018 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7389a7d7-83cb-4c0d-bb69-030c3331caf5-scripts\") pod \"7389a7d7-83cb-4c0d-bb69-030c3331caf5\" (UID: \"7389a7d7-83cb-4c0d-bb69-030c3331caf5\") " Feb 17 13:25:07 crc kubenswrapper[4955]: I0217 13:25:07.151101 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7389a7d7-83cb-4c0d-bb69-030c3331caf5-log-httpd\") pod \"7389a7d7-83cb-4c0d-bb69-030c3331caf5\" (UID: \"7389a7d7-83cb-4c0d-bb69-030c3331caf5\") " Feb 17 13:25:07 crc kubenswrapper[4955]: I0217 13:25:07.151214 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7389a7d7-83cb-4c0d-bb69-030c3331caf5-config-data\") pod \"7389a7d7-83cb-4c0d-bb69-030c3331caf5\" (UID: \"7389a7d7-83cb-4c0d-bb69-030c3331caf5\") " Feb 17 13:25:07 crc kubenswrapper[4955]: I0217 13:25:07.151246 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7389a7d7-83cb-4c0d-bb69-030c3331caf5-combined-ca-bundle\") pod \"7389a7d7-83cb-4c0d-bb69-030c3331caf5\" (UID: \"7389a7d7-83cb-4c0d-bb69-030c3331caf5\") " Feb 17 13:25:07 crc kubenswrapper[4955]: I0217 13:25:07.152968 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7389a7d7-83cb-4c0d-bb69-030c3331caf5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "7389a7d7-83cb-4c0d-bb69-030c3331caf5" (UID: "7389a7d7-83cb-4c0d-bb69-030c3331caf5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:25:07 crc kubenswrapper[4955]: I0217 13:25:07.153232 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7389a7d7-83cb-4c0d-bb69-030c3331caf5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "7389a7d7-83cb-4c0d-bb69-030c3331caf5" (UID: "7389a7d7-83cb-4c0d-bb69-030c3331caf5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:25:07 crc kubenswrapper[4955]: I0217 13:25:07.157640 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7389a7d7-83cb-4c0d-bb69-030c3331caf5-kube-api-access-vhm99" (OuterVolumeSpecName: "kube-api-access-vhm99") pod "7389a7d7-83cb-4c0d-bb69-030c3331caf5" (UID: "7389a7d7-83cb-4c0d-bb69-030c3331caf5"). InnerVolumeSpecName "kube-api-access-vhm99". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:25:07 crc kubenswrapper[4955]: I0217 13:25:07.158036 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7389a7d7-83cb-4c0d-bb69-030c3331caf5-scripts" (OuterVolumeSpecName: "scripts") pod "7389a7d7-83cb-4c0d-bb69-030c3331caf5" (UID: "7389a7d7-83cb-4c0d-bb69-030c3331caf5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:25:07 crc kubenswrapper[4955]: I0217 13:25:07.195359 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7389a7d7-83cb-4c0d-bb69-030c3331caf5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "7389a7d7-83cb-4c0d-bb69-030c3331caf5" (UID: "7389a7d7-83cb-4c0d-bb69-030c3331caf5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:25:07 crc kubenswrapper[4955]: I0217 13:25:07.248578 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7389a7d7-83cb-4c0d-bb69-030c3331caf5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7389a7d7-83cb-4c0d-bb69-030c3331caf5" (UID: "7389a7d7-83cb-4c0d-bb69-030c3331caf5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:25:07 crc kubenswrapper[4955]: I0217 13:25:07.253983 4955 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7389a7d7-83cb-4c0d-bb69-030c3331caf5-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:07 crc kubenswrapper[4955]: I0217 13:25:07.254020 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7389a7d7-83cb-4c0d-bb69-030c3331caf5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:07 crc kubenswrapper[4955]: I0217 13:25:07.254038 4955 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7389a7d7-83cb-4c0d-bb69-030c3331caf5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:07 crc kubenswrapper[4955]: I0217 13:25:07.254049 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhm99\" (UniqueName: \"kubernetes.io/projected/7389a7d7-83cb-4c0d-bb69-030c3331caf5-kube-api-access-vhm99\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:07 crc kubenswrapper[4955]: I0217 13:25:07.254060 4955 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7389a7d7-83cb-4c0d-bb69-030c3331caf5-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:07 crc kubenswrapper[4955]: I0217 13:25:07.254070 4955 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7389a7d7-83cb-4c0d-bb69-030c3331caf5-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:07 crc kubenswrapper[4955]: I0217 13:25:07.314493 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7389a7d7-83cb-4c0d-bb69-030c3331caf5-config-data" (OuterVolumeSpecName: "config-data") pod "7389a7d7-83cb-4c0d-bb69-030c3331caf5" (UID: "7389a7d7-83cb-4c0d-bb69-030c3331caf5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:25:07 crc kubenswrapper[4955]: I0217 13:25:07.356000 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7389a7d7-83cb-4c0d-bb69-030c3331caf5-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:07 crc kubenswrapper[4955]: I0217 13:25:07.988900 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c7a1ed55-8add-4f6d-8210-76ccf42a58ab","Type":"ContainerStarted","Data":"4a08ccf38d2598943d136d7704d1b698199baa699a28452a09e8febea77a5314"} Feb 17 13:25:07 crc kubenswrapper[4955]: I0217 13:25:07.989157 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c7a1ed55-8add-4f6d-8210-76ccf42a58ab","Type":"ContainerStarted","Data":"52440ed4e221c2435cc3b96478a7d6c2a4fa2acac333904752ec3fc71ae352ff"} Feb 17 13:25:07 crc kubenswrapper[4955]: I0217 13:25:07.989050 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c7a1ed55-8add-4f6d-8210-76ccf42a58ab" containerName="nova-metadata-metadata" containerID="cri-o://4a08ccf38d2598943d136d7704d1b698199baa699a28452a09e8febea77a5314" gracePeriod=30 Feb 17 13:25:07 crc kubenswrapper[4955]: I0217 13:25:07.988966 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c7a1ed55-8add-4f6d-8210-76ccf42a58ab" containerName="nova-metadata-log" containerID="cri-o://52440ed4e221c2435cc3b96478a7d6c2a4fa2acac333904752ec3fc71ae352ff" gracePeriod=30 Feb 17 13:25:07 crc kubenswrapper[4955]: I0217 13:25:07.993042 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5","Type":"ContainerStarted","Data":"97759fe437604ffabfe0c43677a2a1cbd3cfc9e1bafbc9cf1352d1e5a5980e17"} Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:07.999542 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7389a7d7-83cb-4c0d-bb69-030c3331caf5","Type":"ContainerDied","Data":"c929e36b4ca9fb82be6838f34032ea4e6cea8b4f48381fd07e51ea2352306457"} Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:07.999587 4955 scope.go:117] "RemoveContainer" containerID="33c70cbe6a16ffe2997b72bfe5ffe17ac36b209090875f6c32429fbdde279ca7" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:07.999735 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.004638 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4f69d681-27e5-4fb9-a199-3c6847c80dc9","Type":"ContainerStarted","Data":"d0c2a1648e356808973156b6f18094775d928e7b2eea74ad79b7853145dd8060"} Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.004689 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4f69d681-27e5-4fb9-a199-3c6847c80dc9","Type":"ContainerStarted","Data":"39b9c23a8dd3329c82edc3d7482ed05f79bff1def6d7c17d6fdb9c742ff68c8d"} Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.011915 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"242f0cab-eded-4013-8ed0-f9ddd4de5fd0","Type":"ContainerStarted","Data":"142e9ee4cc1286ea2bc083be3162bcbf9edc08dc9caa0a0b95ec9437e6c3b941"} Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.012087 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="242f0cab-eded-4013-8ed0-f9ddd4de5fd0" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://142e9ee4cc1286ea2bc083be3162bcbf9edc08dc9caa0a0b95ec9437e6c3b941" gracePeriod=30 Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.025145 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.797405754 podStartE2EDuration="6.025124311s" podCreationTimestamp="2026-02-17 13:25:02 +0000 UTC" firstStartedPulling="2026-02-17 13:25:03.971706465 +0000 UTC m=+1242.494436008" lastFinishedPulling="2026-02-17 13:25:07.199425022 +0000 UTC m=+1245.722154565" observedRunningTime="2026-02-17 13:25:08.012287083 +0000 UTC m=+1246.535016646" watchObservedRunningTime="2026-02-17 13:25:08.025124311 +0000 UTC m=+1246.547853854" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.029508 4955 scope.go:117] "RemoveContainer" containerID="9052f1507175292366564a8ac671e92809a029127e38cbd05a33f28841be19d9" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.046864 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.267309253 podStartE2EDuration="6.046846094s" podCreationTimestamp="2026-02-17 13:25:02 +0000 UTC" firstStartedPulling="2026-02-17 13:25:03.806060585 +0000 UTC m=+1242.328790138" lastFinishedPulling="2026-02-17 13:25:07.585597436 +0000 UTC m=+1246.108326979" observedRunningTime="2026-02-17 13:25:08.038972518 +0000 UTC m=+1246.561702061" watchObservedRunningTime="2026-02-17 13:25:08.046846094 +0000 UTC m=+1246.569575637" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.058942 4955 scope.go:117] "RemoveContainer" containerID="a5fa3a4ed0a0ff395c5c44ad2e45b52e16a7527b29765548b1e51df776873468" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.060649 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.916825628 podStartE2EDuration="6.060625839s" podCreationTimestamp="2026-02-17 13:25:02 +0000 UTC" firstStartedPulling="2026-02-17 13:25:04.04405648 +0000 UTC m=+1242.566786023" lastFinishedPulling="2026-02-17 13:25:07.187856691 +0000 UTC m=+1245.710586234" observedRunningTime="2026-02-17 13:25:08.054597726 +0000 UTC m=+1246.577327279" watchObservedRunningTime="2026-02-17 13:25:08.060625839 +0000 UTC m=+1246.583355382" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.089411 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.17695384 podStartE2EDuration="6.089371494s" podCreationTimestamp="2026-02-17 13:25:02 +0000 UTC" firstStartedPulling="2026-02-17 13:25:04.273999095 +0000 UTC m=+1242.796728638" lastFinishedPulling="2026-02-17 13:25:07.186416749 +0000 UTC m=+1245.709146292" observedRunningTime="2026-02-17 13:25:08.081031325 +0000 UTC m=+1246.603760868" watchObservedRunningTime="2026-02-17 13:25:08.089371494 +0000 UTC m=+1246.612101037" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.119561 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.126970 4955 scope.go:117] "RemoveContainer" containerID="abc6b8d601853195caab008f3b8e84b0ccbedfad1fc886969c8c98835f077d7f" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.133453 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.142267 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.152219 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:25:08 crc kubenswrapper[4955]: E0217 13:25:08.152624 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7389a7d7-83cb-4c0d-bb69-030c3331caf5" containerName="ceilometer-central-agent" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.152640 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="7389a7d7-83cb-4c0d-bb69-030c3331caf5" containerName="ceilometer-central-agent" Feb 17 13:25:08 crc kubenswrapper[4955]: E0217 13:25:08.152657 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7389a7d7-83cb-4c0d-bb69-030c3331caf5" containerName="proxy-httpd" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.152664 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="7389a7d7-83cb-4c0d-bb69-030c3331caf5" containerName="proxy-httpd" Feb 17 13:25:08 crc kubenswrapper[4955]: E0217 13:25:08.152675 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7389a7d7-83cb-4c0d-bb69-030c3331caf5" containerName="ceilometer-notification-agent" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.152681 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="7389a7d7-83cb-4c0d-bb69-030c3331caf5" containerName="ceilometer-notification-agent" Feb 17 13:25:08 crc kubenswrapper[4955]: E0217 13:25:08.152695 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7389a7d7-83cb-4c0d-bb69-030c3331caf5" containerName="sg-core" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.152701 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="7389a7d7-83cb-4c0d-bb69-030c3331caf5" containerName="sg-core" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.152877 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="7389a7d7-83cb-4c0d-bb69-030c3331caf5" containerName="proxy-httpd" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.152888 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="7389a7d7-83cb-4c0d-bb69-030c3331caf5" containerName="ceilometer-notification-agent" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.152901 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="7389a7d7-83cb-4c0d-bb69-030c3331caf5" containerName="sg-core" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.152913 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="7389a7d7-83cb-4c0d-bb69-030c3331caf5" containerName="ceilometer-central-agent" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.154439 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.159391 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.159645 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.159806 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.177905 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.236120 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7389a7d7-83cb-4c0d-bb69-030c3331caf5" path="/var/lib/kubelet/pods/7389a7d7-83cb-4c0d-bb69-030c3331caf5/volumes" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.276803 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e1c521d2-55b0-4386-9375-7c7b64d1610f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e1c521d2-55b0-4386-9375-7c7b64d1610f\") " pod="openstack/ceilometer-0" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.276859 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1c521d2-55b0-4386-9375-7c7b64d1610f-config-data\") pod \"ceilometer-0\" (UID: \"e1c521d2-55b0-4386-9375-7c7b64d1610f\") " pod="openstack/ceilometer-0" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.276881 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1c521d2-55b0-4386-9375-7c7b64d1610f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"e1c521d2-55b0-4386-9375-7c7b64d1610f\") " pod="openstack/ceilometer-0" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.276937 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1c521d2-55b0-4386-9375-7c7b64d1610f-log-httpd\") pod \"ceilometer-0\" (UID: \"e1c521d2-55b0-4386-9375-7c7b64d1610f\") " pod="openstack/ceilometer-0" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.276952 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1c521d2-55b0-4386-9375-7c7b64d1610f-scripts\") pod \"ceilometer-0\" (UID: \"e1c521d2-55b0-4386-9375-7c7b64d1610f\") " pod="openstack/ceilometer-0" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.277001 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1c521d2-55b0-4386-9375-7c7b64d1610f-run-httpd\") pod \"ceilometer-0\" (UID: \"e1c521d2-55b0-4386-9375-7c7b64d1610f\") " pod="openstack/ceilometer-0" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.277027 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kccmt\" (UniqueName: \"kubernetes.io/projected/e1c521d2-55b0-4386-9375-7c7b64d1610f-kube-api-access-kccmt\") pod \"ceilometer-0\" (UID: \"e1c521d2-55b0-4386-9375-7c7b64d1610f\") " pod="openstack/ceilometer-0" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.277048 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1c521d2-55b0-4386-9375-7c7b64d1610f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e1c521d2-55b0-4386-9375-7c7b64d1610f\") " pod="openstack/ceilometer-0" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.329576 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.329625 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.354013 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.379212 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1c521d2-55b0-4386-9375-7c7b64d1610f-log-httpd\") pod \"ceilometer-0\" (UID: \"e1c521d2-55b0-4386-9375-7c7b64d1610f\") " pod="openstack/ceilometer-0" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.379269 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1c521d2-55b0-4386-9375-7c7b64d1610f-scripts\") pod \"ceilometer-0\" (UID: \"e1c521d2-55b0-4386-9375-7c7b64d1610f\") " pod="openstack/ceilometer-0" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.379356 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1c521d2-55b0-4386-9375-7c7b64d1610f-run-httpd\") pod \"ceilometer-0\" (UID: \"e1c521d2-55b0-4386-9375-7c7b64d1610f\") " pod="openstack/ceilometer-0" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.379393 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kccmt\" (UniqueName: \"kubernetes.io/projected/e1c521d2-55b0-4386-9375-7c7b64d1610f-kube-api-access-kccmt\") pod \"ceilometer-0\" (UID: \"e1c521d2-55b0-4386-9375-7c7b64d1610f\") " pod="openstack/ceilometer-0" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.379428 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1c521d2-55b0-4386-9375-7c7b64d1610f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e1c521d2-55b0-4386-9375-7c7b64d1610f\") " pod="openstack/ceilometer-0" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.379502 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e1c521d2-55b0-4386-9375-7c7b64d1610f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e1c521d2-55b0-4386-9375-7c7b64d1610f\") " pod="openstack/ceilometer-0" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.379533 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1c521d2-55b0-4386-9375-7c7b64d1610f-config-data\") pod \"ceilometer-0\" (UID: \"e1c521d2-55b0-4386-9375-7c7b64d1610f\") " pod="openstack/ceilometer-0" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.379557 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1c521d2-55b0-4386-9375-7c7b64d1610f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"e1c521d2-55b0-4386-9375-7c7b64d1610f\") " pod="openstack/ceilometer-0" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.380444 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1c521d2-55b0-4386-9375-7c7b64d1610f-log-httpd\") pod \"ceilometer-0\" (UID: \"e1c521d2-55b0-4386-9375-7c7b64d1610f\") " pod="openstack/ceilometer-0" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.381986 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1c521d2-55b0-4386-9375-7c7b64d1610f-run-httpd\") pod \"ceilometer-0\" (UID: \"e1c521d2-55b0-4386-9375-7c7b64d1610f\") " pod="openstack/ceilometer-0" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.386353 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e1c521d2-55b0-4386-9375-7c7b64d1610f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e1c521d2-55b0-4386-9375-7c7b64d1610f\") " pod="openstack/ceilometer-0" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.386425 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1c521d2-55b0-4386-9375-7c7b64d1610f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e1c521d2-55b0-4386-9375-7c7b64d1610f\") " pod="openstack/ceilometer-0" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.386545 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1c521d2-55b0-4386-9375-7c7b64d1610f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"e1c521d2-55b0-4386-9375-7c7b64d1610f\") " pod="openstack/ceilometer-0" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.387266 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1c521d2-55b0-4386-9375-7c7b64d1610f-config-data\") pod \"ceilometer-0\" (UID: \"e1c521d2-55b0-4386-9375-7c7b64d1610f\") " pod="openstack/ceilometer-0" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.388314 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1c521d2-55b0-4386-9375-7c7b64d1610f-scripts\") pod \"ceilometer-0\" (UID: \"e1c521d2-55b0-4386-9375-7c7b64d1610f\") " pod="openstack/ceilometer-0" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.404308 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kccmt\" (UniqueName: \"kubernetes.io/projected/e1c521d2-55b0-4386-9375-7c7b64d1610f-kube-api-access-kccmt\") pod \"ceilometer-0\" (UID: \"e1c521d2-55b0-4386-9375-7c7b64d1610f\") " pod="openstack/ceilometer-0" Feb 17 13:25:08 crc kubenswrapper[4955]: I0217 13:25:08.508868 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 17 13:25:09 crc kubenswrapper[4955]: I0217 13:25:08.999856 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:25:09 crc kubenswrapper[4955]: I0217 13:25:09.024347 4955 generic.go:334] "Generic (PLEG): container finished" podID="c7a1ed55-8add-4f6d-8210-76ccf42a58ab" containerID="52440ed4e221c2435cc3b96478a7d6c2a4fa2acac333904752ec3fc71ae352ff" exitCode=143 Feb 17 13:25:09 crc kubenswrapper[4955]: I0217 13:25:09.024465 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c7a1ed55-8add-4f6d-8210-76ccf42a58ab","Type":"ContainerDied","Data":"52440ed4e221c2435cc3b96478a7d6c2a4fa2acac333904752ec3fc71ae352ff"} Feb 17 13:25:09 crc kubenswrapper[4955]: W0217 13:25:09.055246 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1c521d2_55b0_4386_9375_7c7b64d1610f.slice/crio-a129372e16752f4b32b48de60cd95e470f523f4ab09f370e483ead6fd71565b9 WatchSource:0}: Error finding container a129372e16752f4b32b48de60cd95e470f523f4ab09f370e483ead6fd71565b9: Status 404 returned error can't find the container with id a129372e16752f4b32b48de60cd95e470f523f4ab09f370e483ead6fd71565b9 Feb 17 13:25:10 crc kubenswrapper[4955]: I0217 13:25:10.058733 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1c521d2-55b0-4386-9375-7c7b64d1610f","Type":"ContainerStarted","Data":"089fe3dff83ab164a4663da073016319402cb07ed7779e22c97e9d51bd600efa"} Feb 17 13:25:10 crc kubenswrapper[4955]: I0217 13:25:10.059125 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1c521d2-55b0-4386-9375-7c7b64d1610f","Type":"ContainerStarted","Data":"a129372e16752f4b32b48de60cd95e470f523f4ab09f370e483ead6fd71565b9"} Feb 17 13:25:11 crc kubenswrapper[4955]: I0217 13:25:11.074940 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1c521d2-55b0-4386-9375-7c7b64d1610f","Type":"ContainerStarted","Data":"744b17f1d9b104c8caa4129128acc8dc79daba9c743ce68b87f254b512b4d903"} Feb 17 13:25:12 crc kubenswrapper[4955]: I0217 13:25:12.090176 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1c521d2-55b0-4386-9375-7c7b64d1610f","Type":"ContainerStarted","Data":"1ecfa2946eb9cafc9cad6f71311771dcc302de5797b49660b697cd2042a0fff9"} Feb 17 13:25:12 crc kubenswrapper[4955]: I0217 13:25:12.158444 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Feb 17 13:25:13 crc kubenswrapper[4955]: I0217 13:25:13.109419 4955 generic.go:334] "Generic (PLEG): container finished" podID="41b91660-a2ac-4f9e-8315-5de36243f161" containerID="53bb653afbe4c0a45289c67de05edbd4e34d900f6ef49832c55c7b0d6452eeeb" exitCode=0 Feb 17 13:25:13 crc kubenswrapper[4955]: I0217 13:25:13.109796 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-4kqpw" event={"ID":"41b91660-a2ac-4f9e-8315-5de36243f161","Type":"ContainerDied","Data":"53bb653afbe4c0a45289c67de05edbd4e34d900f6ef49832c55c7b0d6452eeeb"} Feb 17 13:25:13 crc kubenswrapper[4955]: I0217 13:25:13.141616 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 17 13:25:13 crc kubenswrapper[4955]: I0217 13:25:13.178537 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 17 13:25:13 crc kubenswrapper[4955]: I0217 13:25:13.431996 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-bccf8f775-b6cbh" Feb 17 13:25:13 crc kubenswrapper[4955]: I0217 13:25:13.456567 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 17 13:25:13 crc kubenswrapper[4955]: I0217 13:25:13.456621 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 17 13:25:13 crc kubenswrapper[4955]: I0217 13:25:13.497152 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-hvvzs"] Feb 17 13:25:13 crc kubenswrapper[4955]: I0217 13:25:13.497465 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6578955fd5-hvvzs" podUID="29f9dab6-c497-4a2b-a39c-0c58adae2aa1" containerName="dnsmasq-dns" containerID="cri-o://ed9ae269161b452111acb8fe7e3cac0476b8ea07308d6f3e4c8d0072aacdca7e" gracePeriod=10 Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.124479 4955 generic.go:334] "Generic (PLEG): container finished" podID="29f9dab6-c497-4a2b-a39c-0c58adae2aa1" containerID="ed9ae269161b452111acb8fe7e3cac0476b8ea07308d6f3e4c8d0072aacdca7e" exitCode=0 Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.124679 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-hvvzs" event={"ID":"29f9dab6-c497-4a2b-a39c-0c58adae2aa1","Type":"ContainerDied","Data":"ed9ae269161b452111acb8fe7e3cac0476b8ea07308d6f3e4c8d0072aacdca7e"} Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.124942 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-hvvzs" event={"ID":"29f9dab6-c497-4a2b-a39c-0c58adae2aa1","Type":"ContainerDied","Data":"f90ddae94e8e106c7edb0893caf8923a733f2b267bba1313ad0a500fee111ac1"} Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.124958 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f90ddae94e8e106c7edb0893caf8923a733f2b267bba1313ad0a500fee111ac1" Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.128042 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1c521d2-55b0-4386-9375-7c7b64d1610f","Type":"ContainerStarted","Data":"61150ce92008ba120f051cc2960e790a0f85e45b49f99bc87e4d000b7f6f0af5"} Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.129351 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.131815 4955 generic.go:334] "Generic (PLEG): container finished" podID="0c40c5ab-a929-4850-9a70-b98106008500" containerID="7fbd0bf75c3a27db6ce7417c4a97396f46f96609a3d031db30a4adcab8f30fd3" exitCode=0 Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.131993 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-mzzkq" event={"ID":"0c40c5ab-a929-4850-9a70-b98106008500","Type":"ContainerDied","Data":"7fbd0bf75c3a27db6ce7417c4a97396f46f96609a3d031db30a4adcab8f30fd3"} Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.166238 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-hvvzs" Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.171432 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.178589 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.350355726 podStartE2EDuration="6.178568513s" podCreationTimestamp="2026-02-17 13:25:08 +0000 UTC" firstStartedPulling="2026-02-17 13:25:09.058467096 +0000 UTC m=+1247.581196639" lastFinishedPulling="2026-02-17 13:25:12.886679883 +0000 UTC m=+1251.409409426" observedRunningTime="2026-02-17 13:25:14.157469097 +0000 UTC m=+1252.680198660" watchObservedRunningTime="2026-02-17 13:25:14.178568513 +0000 UTC m=+1252.701298056" Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.302450 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/29f9dab6-c497-4a2b-a39c-0c58adae2aa1-ovsdbserver-sb\") pod \"29f9dab6-c497-4a2b-a39c-0c58adae2aa1\" (UID: \"29f9dab6-c497-4a2b-a39c-0c58adae2aa1\") " Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.302568 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/29f9dab6-c497-4a2b-a39c-0c58adae2aa1-dns-swift-storage-0\") pod \"29f9dab6-c497-4a2b-a39c-0c58adae2aa1\" (UID: \"29f9dab6-c497-4a2b-a39c-0c58adae2aa1\") " Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.302645 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/29f9dab6-c497-4a2b-a39c-0c58adae2aa1-ovsdbserver-nb\") pod \"29f9dab6-c497-4a2b-a39c-0c58adae2aa1\" (UID: \"29f9dab6-c497-4a2b-a39c-0c58adae2aa1\") " Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.302700 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rfgpf\" (UniqueName: \"kubernetes.io/projected/29f9dab6-c497-4a2b-a39c-0c58adae2aa1-kube-api-access-rfgpf\") pod \"29f9dab6-c497-4a2b-a39c-0c58adae2aa1\" (UID: \"29f9dab6-c497-4a2b-a39c-0c58adae2aa1\") " Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.302750 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29f9dab6-c497-4a2b-a39c-0c58adae2aa1-config\") pod \"29f9dab6-c497-4a2b-a39c-0c58adae2aa1\" (UID: \"29f9dab6-c497-4a2b-a39c-0c58adae2aa1\") " Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.302851 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/29f9dab6-c497-4a2b-a39c-0c58adae2aa1-dns-svc\") pod \"29f9dab6-c497-4a2b-a39c-0c58adae2aa1\" (UID: \"29f9dab6-c497-4a2b-a39c-0c58adae2aa1\") " Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.323471 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29f9dab6-c497-4a2b-a39c-0c58adae2aa1-kube-api-access-rfgpf" (OuterVolumeSpecName: "kube-api-access-rfgpf") pod "29f9dab6-c497-4a2b-a39c-0c58adae2aa1" (UID: "29f9dab6-c497-4a2b-a39c-0c58adae2aa1"). InnerVolumeSpecName "kube-api-access-rfgpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.409031 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rfgpf\" (UniqueName: \"kubernetes.io/projected/29f9dab6-c497-4a2b-a39c-0c58adae2aa1-kube-api-access-rfgpf\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.421521 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/29f9dab6-c497-4a2b-a39c-0c58adae2aa1-config" (OuterVolumeSpecName: "config") pod "29f9dab6-c497-4a2b-a39c-0c58adae2aa1" (UID: "29f9dab6-c497-4a2b-a39c-0c58adae2aa1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.450827 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/29f9dab6-c497-4a2b-a39c-0c58adae2aa1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "29f9dab6-c497-4a2b-a39c-0c58adae2aa1" (UID: "29f9dab6-c497-4a2b-a39c-0c58adae2aa1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.456404 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/29f9dab6-c497-4a2b-a39c-0c58adae2aa1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "29f9dab6-c497-4a2b-a39c-0c58adae2aa1" (UID: "29f9dab6-c497-4a2b-a39c-0c58adae2aa1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.475108 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/29f9dab6-c497-4a2b-a39c-0c58adae2aa1-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "29f9dab6-c497-4a2b-a39c-0c58adae2aa1" (UID: "29f9dab6-c497-4a2b-a39c-0c58adae2aa1"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.515633 4955 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/29f9dab6-c497-4a2b-a39c-0c58adae2aa1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.515667 4955 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/29f9dab6-c497-4a2b-a39c-0c58adae2aa1-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.515687 4955 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/29f9dab6-c497-4a2b-a39c-0c58adae2aa1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.515697 4955 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29f9dab6-c497-4a2b-a39c-0c58adae2aa1-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.528889 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/29f9dab6-c497-4a2b-a39c-0c58adae2aa1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "29f9dab6-c497-4a2b-a39c-0c58adae2aa1" (UID: "29f9dab6-c497-4a2b-a39c-0c58adae2aa1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.535420 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-4kqpw" Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.539979 4955 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="4f69d681-27e5-4fb9-a199-3c6847c80dc9" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.196:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.541160 4955 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="4f69d681-27e5-4fb9-a199-3c6847c80dc9" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.196:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.617200 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4mjpr\" (UniqueName: \"kubernetes.io/projected/41b91660-a2ac-4f9e-8315-5de36243f161-kube-api-access-4mjpr\") pod \"41b91660-a2ac-4f9e-8315-5de36243f161\" (UID: \"41b91660-a2ac-4f9e-8315-5de36243f161\") " Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.617371 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41b91660-a2ac-4f9e-8315-5de36243f161-combined-ca-bundle\") pod \"41b91660-a2ac-4f9e-8315-5de36243f161\" (UID: \"41b91660-a2ac-4f9e-8315-5de36243f161\") " Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.617435 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41b91660-a2ac-4f9e-8315-5de36243f161-scripts\") pod \"41b91660-a2ac-4f9e-8315-5de36243f161\" (UID: \"41b91660-a2ac-4f9e-8315-5de36243f161\") " Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.617475 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41b91660-a2ac-4f9e-8315-5de36243f161-config-data\") pod \"41b91660-a2ac-4f9e-8315-5de36243f161\" (UID: \"41b91660-a2ac-4f9e-8315-5de36243f161\") " Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.618145 4955 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/29f9dab6-c497-4a2b-a39c-0c58adae2aa1-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.624913 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41b91660-a2ac-4f9e-8315-5de36243f161-kube-api-access-4mjpr" (OuterVolumeSpecName: "kube-api-access-4mjpr") pod "41b91660-a2ac-4f9e-8315-5de36243f161" (UID: "41b91660-a2ac-4f9e-8315-5de36243f161"). InnerVolumeSpecName "kube-api-access-4mjpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.624997 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41b91660-a2ac-4f9e-8315-5de36243f161-scripts" (OuterVolumeSpecName: "scripts") pod "41b91660-a2ac-4f9e-8315-5de36243f161" (UID: "41b91660-a2ac-4f9e-8315-5de36243f161"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.653486 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41b91660-a2ac-4f9e-8315-5de36243f161-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "41b91660-a2ac-4f9e-8315-5de36243f161" (UID: "41b91660-a2ac-4f9e-8315-5de36243f161"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.658099 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41b91660-a2ac-4f9e-8315-5de36243f161-config-data" (OuterVolumeSpecName: "config-data") pod "41b91660-a2ac-4f9e-8315-5de36243f161" (UID: "41b91660-a2ac-4f9e-8315-5de36243f161"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.719878 4955 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41b91660-a2ac-4f9e-8315-5de36243f161-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.719922 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41b91660-a2ac-4f9e-8315-5de36243f161-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.719934 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4mjpr\" (UniqueName: \"kubernetes.io/projected/41b91660-a2ac-4f9e-8315-5de36243f161-kube-api-access-4mjpr\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:14 crc kubenswrapper[4955]: I0217 13:25:14.719947 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41b91660-a2ac-4f9e-8315-5de36243f161-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:15 crc kubenswrapper[4955]: I0217 13:25:15.145750 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-4kqpw" event={"ID":"41b91660-a2ac-4f9e-8315-5de36243f161","Type":"ContainerDied","Data":"056c3f45c2a20fe4fc964fcad05f11f1ac3f6c60aa5c9a4c3a46aeccfb0bc0eb"} Feb 17 13:25:15 crc kubenswrapper[4955]: I0217 13:25:15.145821 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="056c3f45c2a20fe4fc964fcad05f11f1ac3f6c60aa5c9a4c3a46aeccfb0bc0eb" Feb 17 13:25:15 crc kubenswrapper[4955]: I0217 13:25:15.145907 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-4kqpw" Feb 17 13:25:15 crc kubenswrapper[4955]: I0217 13:25:15.145913 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-hvvzs" Feb 17 13:25:15 crc kubenswrapper[4955]: I0217 13:25:15.215158 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-hvvzs"] Feb 17 13:25:15 crc kubenswrapper[4955]: I0217 13:25:15.242289 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-hvvzs"] Feb 17 13:25:15 crc kubenswrapper[4955]: I0217 13:25:15.313516 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 17 13:25:15 crc kubenswrapper[4955]: I0217 13:25:15.313757 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="4f69d681-27e5-4fb9-a199-3c6847c80dc9" containerName="nova-api-log" containerID="cri-o://39b9c23a8dd3329c82edc3d7482ed05f79bff1def6d7c17d6fdb9c742ff68c8d" gracePeriod=30 Feb 17 13:25:15 crc kubenswrapper[4955]: I0217 13:25:15.314037 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="4f69d681-27e5-4fb9-a199-3c6847c80dc9" containerName="nova-api-api" containerID="cri-o://d0c2a1648e356808973156b6f18094775d928e7b2eea74ad79b7853145dd8060" gracePeriod=30 Feb 17 13:25:15 crc kubenswrapper[4955]: I0217 13:25:15.352793 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 17 13:25:15 crc kubenswrapper[4955]: I0217 13:25:15.613249 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-mzzkq" Feb 17 13:25:15 crc kubenswrapper[4955]: I0217 13:25:15.747148 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c40c5ab-a929-4850-9a70-b98106008500-config-data\") pod \"0c40c5ab-a929-4850-9a70-b98106008500\" (UID: \"0c40c5ab-a929-4850-9a70-b98106008500\") " Feb 17 13:25:15 crc kubenswrapper[4955]: I0217 13:25:15.747254 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c40c5ab-a929-4850-9a70-b98106008500-combined-ca-bundle\") pod \"0c40c5ab-a929-4850-9a70-b98106008500\" (UID: \"0c40c5ab-a929-4850-9a70-b98106008500\") " Feb 17 13:25:15 crc kubenswrapper[4955]: I0217 13:25:15.747275 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4nnq\" (UniqueName: \"kubernetes.io/projected/0c40c5ab-a929-4850-9a70-b98106008500-kube-api-access-c4nnq\") pod \"0c40c5ab-a929-4850-9a70-b98106008500\" (UID: \"0c40c5ab-a929-4850-9a70-b98106008500\") " Feb 17 13:25:15 crc kubenswrapper[4955]: I0217 13:25:15.747388 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c40c5ab-a929-4850-9a70-b98106008500-scripts\") pod \"0c40c5ab-a929-4850-9a70-b98106008500\" (UID: \"0c40c5ab-a929-4850-9a70-b98106008500\") " Feb 17 13:25:15 crc kubenswrapper[4955]: I0217 13:25:15.788246 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c40c5ab-a929-4850-9a70-b98106008500-scripts" (OuterVolumeSpecName: "scripts") pod "0c40c5ab-a929-4850-9a70-b98106008500" (UID: "0c40c5ab-a929-4850-9a70-b98106008500"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:25:15 crc kubenswrapper[4955]: I0217 13:25:15.789954 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c40c5ab-a929-4850-9a70-b98106008500-kube-api-access-c4nnq" (OuterVolumeSpecName: "kube-api-access-c4nnq") pod "0c40c5ab-a929-4850-9a70-b98106008500" (UID: "0c40c5ab-a929-4850-9a70-b98106008500"). InnerVolumeSpecName "kube-api-access-c4nnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:25:15 crc kubenswrapper[4955]: I0217 13:25:15.814952 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c40c5ab-a929-4850-9a70-b98106008500-config-data" (OuterVolumeSpecName: "config-data") pod "0c40c5ab-a929-4850-9a70-b98106008500" (UID: "0c40c5ab-a929-4850-9a70-b98106008500"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:25:15 crc kubenswrapper[4955]: I0217 13:25:15.816019 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c40c5ab-a929-4850-9a70-b98106008500-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0c40c5ab-a929-4850-9a70-b98106008500" (UID: "0c40c5ab-a929-4850-9a70-b98106008500"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:25:15 crc kubenswrapper[4955]: I0217 13:25:15.849696 4955 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c40c5ab-a929-4850-9a70-b98106008500-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:15 crc kubenswrapper[4955]: I0217 13:25:15.849740 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c40c5ab-a929-4850-9a70-b98106008500-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:15 crc kubenswrapper[4955]: I0217 13:25:15.849753 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c40c5ab-a929-4850-9a70-b98106008500-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:15 crc kubenswrapper[4955]: I0217 13:25:15.849764 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4nnq\" (UniqueName: \"kubernetes.io/projected/0c40c5ab-a929-4850-9a70-b98106008500-kube-api-access-c4nnq\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:16 crc kubenswrapper[4955]: I0217 13:25:16.156853 4955 generic.go:334] "Generic (PLEG): container finished" podID="4f69d681-27e5-4fb9-a199-3c6847c80dc9" containerID="39b9c23a8dd3329c82edc3d7482ed05f79bff1def6d7c17d6fdb9c742ff68c8d" exitCode=143 Feb 17 13:25:16 crc kubenswrapper[4955]: I0217 13:25:16.156967 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4f69d681-27e5-4fb9-a199-3c6847c80dc9","Type":"ContainerDied","Data":"39b9c23a8dd3329c82edc3d7482ed05f79bff1def6d7c17d6fdb9c742ff68c8d"} Feb 17 13:25:16 crc kubenswrapper[4955]: I0217 13:25:16.159010 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-mzzkq" event={"ID":"0c40c5ab-a929-4850-9a70-b98106008500","Type":"ContainerDied","Data":"caec95bd40473db1d3da799ea2fdbaecfbec1549504dd90087cb326c218fac76"} Feb 17 13:25:16 crc kubenswrapper[4955]: I0217 13:25:16.159056 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="caec95bd40473db1d3da799ea2fdbaecfbec1549504dd90087cb326c218fac76" Feb 17 13:25:16 crc kubenswrapper[4955]: I0217 13:25:16.159089 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-mzzkq" Feb 17 13:25:16 crc kubenswrapper[4955]: I0217 13:25:16.159590 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5" containerName="nova-scheduler-scheduler" containerID="cri-o://97759fe437604ffabfe0c43677a2a1cbd3cfc9e1bafbc9cf1352d1e5a5980e17" gracePeriod=30 Feb 17 13:25:16 crc kubenswrapper[4955]: I0217 13:25:16.235058 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29f9dab6-c497-4a2b-a39c-0c58adae2aa1" path="/var/lib/kubelet/pods/29f9dab6-c497-4a2b-a39c-0c58adae2aa1/volumes" Feb 17 13:25:16 crc kubenswrapper[4955]: I0217 13:25:16.254355 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 17 13:25:16 crc kubenswrapper[4955]: E0217 13:25:16.258084 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29f9dab6-c497-4a2b-a39c-0c58adae2aa1" containerName="init" Feb 17 13:25:16 crc kubenswrapper[4955]: I0217 13:25:16.258112 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="29f9dab6-c497-4a2b-a39c-0c58adae2aa1" containerName="init" Feb 17 13:25:16 crc kubenswrapper[4955]: E0217 13:25:16.258136 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29f9dab6-c497-4a2b-a39c-0c58adae2aa1" containerName="dnsmasq-dns" Feb 17 13:25:16 crc kubenswrapper[4955]: I0217 13:25:16.258142 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="29f9dab6-c497-4a2b-a39c-0c58adae2aa1" containerName="dnsmasq-dns" Feb 17 13:25:16 crc kubenswrapper[4955]: E0217 13:25:16.258152 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c40c5ab-a929-4850-9a70-b98106008500" containerName="nova-cell1-conductor-db-sync" Feb 17 13:25:16 crc kubenswrapper[4955]: I0217 13:25:16.258157 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c40c5ab-a929-4850-9a70-b98106008500" containerName="nova-cell1-conductor-db-sync" Feb 17 13:25:16 crc kubenswrapper[4955]: E0217 13:25:16.258179 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41b91660-a2ac-4f9e-8315-5de36243f161" containerName="nova-manage" Feb 17 13:25:16 crc kubenswrapper[4955]: I0217 13:25:16.258185 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="41b91660-a2ac-4f9e-8315-5de36243f161" containerName="nova-manage" Feb 17 13:25:16 crc kubenswrapper[4955]: I0217 13:25:16.258457 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c40c5ab-a929-4850-9a70-b98106008500" containerName="nova-cell1-conductor-db-sync" Feb 17 13:25:16 crc kubenswrapper[4955]: I0217 13:25:16.258470 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="29f9dab6-c497-4a2b-a39c-0c58adae2aa1" containerName="dnsmasq-dns" Feb 17 13:25:16 crc kubenswrapper[4955]: I0217 13:25:16.258489 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="41b91660-a2ac-4f9e-8315-5de36243f161" containerName="nova-manage" Feb 17 13:25:16 crc kubenswrapper[4955]: I0217 13:25:16.259117 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 17 13:25:16 crc kubenswrapper[4955]: I0217 13:25:16.261222 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Feb 17 13:25:16 crc kubenswrapper[4955]: I0217 13:25:16.272916 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 17 13:25:16 crc kubenswrapper[4955]: I0217 13:25:16.359023 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2cc025ff-ae43-4877-940e-dab395505735-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"2cc025ff-ae43-4877-940e-dab395505735\") " pod="openstack/nova-cell1-conductor-0" Feb 17 13:25:16 crc kubenswrapper[4955]: I0217 13:25:16.359347 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjst8\" (UniqueName: \"kubernetes.io/projected/2cc025ff-ae43-4877-940e-dab395505735-kube-api-access-bjst8\") pod \"nova-cell1-conductor-0\" (UID: \"2cc025ff-ae43-4877-940e-dab395505735\") " pod="openstack/nova-cell1-conductor-0" Feb 17 13:25:16 crc kubenswrapper[4955]: I0217 13:25:16.359486 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cc025ff-ae43-4877-940e-dab395505735-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"2cc025ff-ae43-4877-940e-dab395505735\") " pod="openstack/nova-cell1-conductor-0" Feb 17 13:25:16 crc kubenswrapper[4955]: I0217 13:25:16.460943 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2cc025ff-ae43-4877-940e-dab395505735-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"2cc025ff-ae43-4877-940e-dab395505735\") " pod="openstack/nova-cell1-conductor-0" Feb 17 13:25:16 crc kubenswrapper[4955]: I0217 13:25:16.461900 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjst8\" (UniqueName: \"kubernetes.io/projected/2cc025ff-ae43-4877-940e-dab395505735-kube-api-access-bjst8\") pod \"nova-cell1-conductor-0\" (UID: \"2cc025ff-ae43-4877-940e-dab395505735\") " pod="openstack/nova-cell1-conductor-0" Feb 17 13:25:16 crc kubenswrapper[4955]: I0217 13:25:16.462250 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cc025ff-ae43-4877-940e-dab395505735-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"2cc025ff-ae43-4877-940e-dab395505735\") " pod="openstack/nova-cell1-conductor-0" Feb 17 13:25:16 crc kubenswrapper[4955]: I0217 13:25:16.464740 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2cc025ff-ae43-4877-940e-dab395505735-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"2cc025ff-ae43-4877-940e-dab395505735\") " pod="openstack/nova-cell1-conductor-0" Feb 17 13:25:16 crc kubenswrapper[4955]: I0217 13:25:16.468738 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cc025ff-ae43-4877-940e-dab395505735-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"2cc025ff-ae43-4877-940e-dab395505735\") " pod="openstack/nova-cell1-conductor-0" Feb 17 13:25:16 crc kubenswrapper[4955]: I0217 13:25:16.490524 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjst8\" (UniqueName: \"kubernetes.io/projected/2cc025ff-ae43-4877-940e-dab395505735-kube-api-access-bjst8\") pod \"nova-cell1-conductor-0\" (UID: \"2cc025ff-ae43-4877-940e-dab395505735\") " pod="openstack/nova-cell1-conductor-0" Feb 17 13:25:16 crc kubenswrapper[4955]: I0217 13:25:16.577913 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 17 13:25:17 crc kubenswrapper[4955]: I0217 13:25:17.040219 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 17 13:25:17 crc kubenswrapper[4955]: I0217 13:25:17.170905 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"2cc025ff-ae43-4877-940e-dab395505735","Type":"ContainerStarted","Data":"d6cf5315f6d2ffd90e9bb51e8261bacb8bdd615516b318a20d73f8357cdacc0d"} Feb 17 13:25:18 crc kubenswrapper[4955]: E0217 13:25:18.143870 4955 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="97759fe437604ffabfe0c43677a2a1cbd3cfc9e1bafbc9cf1352d1e5a5980e17" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 17 13:25:18 crc kubenswrapper[4955]: E0217 13:25:18.145563 4955 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="97759fe437604ffabfe0c43677a2a1cbd3cfc9e1bafbc9cf1352d1e5a5980e17" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 17 13:25:18 crc kubenswrapper[4955]: E0217 13:25:18.146539 4955 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="97759fe437604ffabfe0c43677a2a1cbd3cfc9e1bafbc9cf1352d1e5a5980e17" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 17 13:25:18 crc kubenswrapper[4955]: E0217 13:25:18.146600 4955 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5" containerName="nova-scheduler-scheduler" Feb 17 13:25:18 crc kubenswrapper[4955]: I0217 13:25:18.179765 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"2cc025ff-ae43-4877-940e-dab395505735","Type":"ContainerStarted","Data":"8af3aac5c53ac3032c71de1cf7bfd38fcc92a579b27495b551ef0d561faa4bf0"} Feb 17 13:25:18 crc kubenswrapper[4955]: I0217 13:25:18.179916 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Feb 17 13:25:18 crc kubenswrapper[4955]: I0217 13:25:18.198092 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.198077254 podStartE2EDuration="2.198077254s" podCreationTimestamp="2026-02-17 13:25:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:25:18.193075461 +0000 UTC m=+1256.715805004" watchObservedRunningTime="2026-02-17 13:25:18.198077254 +0000 UTC m=+1256.720806787" Feb 17 13:25:20 crc kubenswrapper[4955]: I0217 13:25:20.157644 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 17 13:25:20 crc kubenswrapper[4955]: I0217 13:25:20.199854 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 17 13:25:20 crc kubenswrapper[4955]: I0217 13:25:20.199770 4955 generic.go:334] "Generic (PLEG): container finished" podID="01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5" containerID="97759fe437604ffabfe0c43677a2a1cbd3cfc9e1bafbc9cf1352d1e5a5980e17" exitCode=0 Feb 17 13:25:20 crc kubenswrapper[4955]: I0217 13:25:20.199898 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5","Type":"ContainerDied","Data":"97759fe437604ffabfe0c43677a2a1cbd3cfc9e1bafbc9cf1352d1e5a5980e17"} Feb 17 13:25:20 crc kubenswrapper[4955]: I0217 13:25:20.199961 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5","Type":"ContainerDied","Data":"a2e27ba89ac3e2b41e320af3aada6f6083193b336df082aa0219d559bdedafbd"} Feb 17 13:25:20 crc kubenswrapper[4955]: I0217 13:25:20.199988 4955 scope.go:117] "RemoveContainer" containerID="97759fe437604ffabfe0c43677a2a1cbd3cfc9e1bafbc9cf1352d1e5a5980e17" Feb 17 13:25:20 crc kubenswrapper[4955]: I0217 13:25:20.236729 4955 scope.go:117] "RemoveContainer" containerID="97759fe437604ffabfe0c43677a2a1cbd3cfc9e1bafbc9cf1352d1e5a5980e17" Feb 17 13:25:20 crc kubenswrapper[4955]: E0217 13:25:20.237421 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97759fe437604ffabfe0c43677a2a1cbd3cfc9e1bafbc9cf1352d1e5a5980e17\": container with ID starting with 97759fe437604ffabfe0c43677a2a1cbd3cfc9e1bafbc9cf1352d1e5a5980e17 not found: ID does not exist" containerID="97759fe437604ffabfe0c43677a2a1cbd3cfc9e1bafbc9cf1352d1e5a5980e17" Feb 17 13:25:20 crc kubenswrapper[4955]: I0217 13:25:20.237470 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97759fe437604ffabfe0c43677a2a1cbd3cfc9e1bafbc9cf1352d1e5a5980e17"} err="failed to get container status \"97759fe437604ffabfe0c43677a2a1cbd3cfc9e1bafbc9cf1352d1e5a5980e17\": rpc error: code = NotFound desc = could not find container \"97759fe437604ffabfe0c43677a2a1cbd3cfc9e1bafbc9cf1352d1e5a5980e17\": container with ID starting with 97759fe437604ffabfe0c43677a2a1cbd3cfc9e1bafbc9cf1352d1e5a5980e17 not found: ID does not exist" Feb 17 13:25:20 crc kubenswrapper[4955]: I0217 13:25:20.269829 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5-combined-ca-bundle\") pod \"01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5\" (UID: \"01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5\") " Feb 17 13:25:20 crc kubenswrapper[4955]: I0217 13:25:20.269959 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5-config-data\") pod \"01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5\" (UID: \"01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5\") " Feb 17 13:25:20 crc kubenswrapper[4955]: I0217 13:25:20.269998 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tktw7\" (UniqueName: \"kubernetes.io/projected/01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5-kube-api-access-tktw7\") pod \"01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5\" (UID: \"01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5\") " Feb 17 13:25:20 crc kubenswrapper[4955]: I0217 13:25:20.275122 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5-kube-api-access-tktw7" (OuterVolumeSpecName: "kube-api-access-tktw7") pod "01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5" (UID: "01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5"). InnerVolumeSpecName "kube-api-access-tktw7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:25:20 crc kubenswrapper[4955]: E0217 13:25:20.292837 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5-config-data podName:01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5 nodeName:}" failed. No retries permitted until 2026-02-17 13:25:20.792803828 +0000 UTC m=+1259.315533381 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config-data" (UniqueName: "kubernetes.io/secret/01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5-config-data") pod "01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5" (UID: "01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5") : error deleting /var/lib/kubelet/pods/01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5/volume-subpaths: remove /var/lib/kubelet/pods/01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5/volume-subpaths: no such file or directory Feb 17 13:25:20 crc kubenswrapper[4955]: I0217 13:25:20.295812 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5" (UID: "01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:25:20 crc kubenswrapper[4955]: I0217 13:25:20.372328 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:20 crc kubenswrapper[4955]: I0217 13:25:20.372374 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tktw7\" (UniqueName: \"kubernetes.io/projected/01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5-kube-api-access-tktw7\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:20 crc kubenswrapper[4955]: I0217 13:25:20.880513 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5-config-data\") pod \"01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5\" (UID: \"01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5\") " Feb 17 13:25:20 crc kubenswrapper[4955]: I0217 13:25:20.887342 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5-config-data" (OuterVolumeSpecName: "config-data") pod "01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5" (UID: "01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:25:20 crc kubenswrapper[4955]: I0217 13:25:20.982530 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.146875 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.163408 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.173816 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.192342 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 17 13:25:21 crc kubenswrapper[4955]: E0217 13:25:21.192852 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f69d681-27e5-4fb9-a199-3c6847c80dc9" containerName="nova-api-api" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.192875 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f69d681-27e5-4fb9-a199-3c6847c80dc9" containerName="nova-api-api" Feb 17 13:25:21 crc kubenswrapper[4955]: E0217 13:25:21.192889 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f69d681-27e5-4fb9-a199-3c6847c80dc9" containerName="nova-api-log" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.192897 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f69d681-27e5-4fb9-a199-3c6847c80dc9" containerName="nova-api-log" Feb 17 13:25:21 crc kubenswrapper[4955]: E0217 13:25:21.192914 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5" containerName="nova-scheduler-scheduler" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.192924 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5" containerName="nova-scheduler-scheduler" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.193108 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f69d681-27e5-4fb9-a199-3c6847c80dc9" containerName="nova-api-api" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.193135 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f69d681-27e5-4fb9-a199-3c6847c80dc9" containerName="nova-api-log" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.193153 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5" containerName="nova-scheduler-scheduler" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.193830 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.196282 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.219165 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.220009 4955 generic.go:334] "Generic (PLEG): container finished" podID="4f69d681-27e5-4fb9-a199-3c6847c80dc9" containerID="d0c2a1648e356808973156b6f18094775d928e7b2eea74ad79b7853145dd8060" exitCode=0 Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.220047 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4f69d681-27e5-4fb9-a199-3c6847c80dc9","Type":"ContainerDied","Data":"d0c2a1648e356808973156b6f18094775d928e7b2eea74ad79b7853145dd8060"} Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.220071 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4f69d681-27e5-4fb9-a199-3c6847c80dc9","Type":"ContainerDied","Data":"51bd1d076f26064ebc8811e6d84ef38c06c2ee7b90434a8de2ffaeb25929b45b"} Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.220087 4955 scope.go:117] "RemoveContainer" containerID="d0c2a1648e356808973156b6f18094775d928e7b2eea74ad79b7853145dd8060" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.220189 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.241155 4955 scope.go:117] "RemoveContainer" containerID="39b9c23a8dd3329c82edc3d7482ed05f79bff1def6d7c17d6fdb9c742ff68c8d" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.258146 4955 scope.go:117] "RemoveContainer" containerID="d0c2a1648e356808973156b6f18094775d928e7b2eea74ad79b7853145dd8060" Feb 17 13:25:21 crc kubenswrapper[4955]: E0217 13:25:21.258564 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0c2a1648e356808973156b6f18094775d928e7b2eea74ad79b7853145dd8060\": container with ID starting with d0c2a1648e356808973156b6f18094775d928e7b2eea74ad79b7853145dd8060 not found: ID does not exist" containerID="d0c2a1648e356808973156b6f18094775d928e7b2eea74ad79b7853145dd8060" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.258589 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0c2a1648e356808973156b6f18094775d928e7b2eea74ad79b7853145dd8060"} err="failed to get container status \"d0c2a1648e356808973156b6f18094775d928e7b2eea74ad79b7853145dd8060\": rpc error: code = NotFound desc = could not find container \"d0c2a1648e356808973156b6f18094775d928e7b2eea74ad79b7853145dd8060\": container with ID starting with d0c2a1648e356808973156b6f18094775d928e7b2eea74ad79b7853145dd8060 not found: ID does not exist" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.258608 4955 scope.go:117] "RemoveContainer" containerID="39b9c23a8dd3329c82edc3d7482ed05f79bff1def6d7c17d6fdb9c742ff68c8d" Feb 17 13:25:21 crc kubenswrapper[4955]: E0217 13:25:21.259051 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39b9c23a8dd3329c82edc3d7482ed05f79bff1def6d7c17d6fdb9c742ff68c8d\": container with ID starting with 39b9c23a8dd3329c82edc3d7482ed05f79bff1def6d7c17d6fdb9c742ff68c8d not found: ID does not exist" containerID="39b9c23a8dd3329c82edc3d7482ed05f79bff1def6d7c17d6fdb9c742ff68c8d" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.259069 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39b9c23a8dd3329c82edc3d7482ed05f79bff1def6d7c17d6fdb9c742ff68c8d"} err="failed to get container status \"39b9c23a8dd3329c82edc3d7482ed05f79bff1def6d7c17d6fdb9c742ff68c8d\": rpc error: code = NotFound desc = could not find container \"39b9c23a8dd3329c82edc3d7482ed05f79bff1def6d7c17d6fdb9c742ff68c8d\": container with ID starting with 39b9c23a8dd3329c82edc3d7482ed05f79bff1def6d7c17d6fdb9c742ff68c8d not found: ID does not exist" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.289051 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxmm7\" (UniqueName: \"kubernetes.io/projected/4f69d681-27e5-4fb9-a199-3c6847c80dc9-kube-api-access-fxmm7\") pod \"4f69d681-27e5-4fb9-a199-3c6847c80dc9\" (UID: \"4f69d681-27e5-4fb9-a199-3c6847c80dc9\") " Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.289185 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f69d681-27e5-4fb9-a199-3c6847c80dc9-logs\") pod \"4f69d681-27e5-4fb9-a199-3c6847c80dc9\" (UID: \"4f69d681-27e5-4fb9-a199-3c6847c80dc9\") " Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.289210 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f69d681-27e5-4fb9-a199-3c6847c80dc9-combined-ca-bundle\") pod \"4f69d681-27e5-4fb9-a199-3c6847c80dc9\" (UID: \"4f69d681-27e5-4fb9-a199-3c6847c80dc9\") " Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.289497 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f69d681-27e5-4fb9-a199-3c6847c80dc9-config-data\") pod \"4f69d681-27e5-4fb9-a199-3c6847c80dc9\" (UID: \"4f69d681-27e5-4fb9-a199-3c6847c80dc9\") " Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.289905 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f69d681-27e5-4fb9-a199-3c6847c80dc9-logs" (OuterVolumeSpecName: "logs") pod "4f69d681-27e5-4fb9-a199-3c6847c80dc9" (UID: "4f69d681-27e5-4fb9-a199-3c6847c80dc9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.289908 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a893348-52b9-474c-8031-e67020c445eb-config-data\") pod \"nova-scheduler-0\" (UID: \"7a893348-52b9-474c-8031-e67020c445eb\") " pod="openstack/nova-scheduler-0" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.289993 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kl7jf\" (UniqueName: \"kubernetes.io/projected/7a893348-52b9-474c-8031-e67020c445eb-kube-api-access-kl7jf\") pod \"nova-scheduler-0\" (UID: \"7a893348-52b9-474c-8031-e67020c445eb\") " pod="openstack/nova-scheduler-0" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.290033 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a893348-52b9-474c-8031-e67020c445eb-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7a893348-52b9-474c-8031-e67020c445eb\") " pod="openstack/nova-scheduler-0" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.290214 4955 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f69d681-27e5-4fb9-a199-3c6847c80dc9-logs\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.293482 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f69d681-27e5-4fb9-a199-3c6847c80dc9-kube-api-access-fxmm7" (OuterVolumeSpecName: "kube-api-access-fxmm7") pod "4f69d681-27e5-4fb9-a199-3c6847c80dc9" (UID: "4f69d681-27e5-4fb9-a199-3c6847c80dc9"). InnerVolumeSpecName "kube-api-access-fxmm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.315062 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f69d681-27e5-4fb9-a199-3c6847c80dc9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4f69d681-27e5-4fb9-a199-3c6847c80dc9" (UID: "4f69d681-27e5-4fb9-a199-3c6847c80dc9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.316373 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f69d681-27e5-4fb9-a199-3c6847c80dc9-config-data" (OuterVolumeSpecName: "config-data") pod "4f69d681-27e5-4fb9-a199-3c6847c80dc9" (UID: "4f69d681-27e5-4fb9-a199-3c6847c80dc9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.391438 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a893348-52b9-474c-8031-e67020c445eb-config-data\") pod \"nova-scheduler-0\" (UID: \"7a893348-52b9-474c-8031-e67020c445eb\") " pod="openstack/nova-scheduler-0" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.391757 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kl7jf\" (UniqueName: \"kubernetes.io/projected/7a893348-52b9-474c-8031-e67020c445eb-kube-api-access-kl7jf\") pod \"nova-scheduler-0\" (UID: \"7a893348-52b9-474c-8031-e67020c445eb\") " pod="openstack/nova-scheduler-0" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.391998 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a893348-52b9-474c-8031-e67020c445eb-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7a893348-52b9-474c-8031-e67020c445eb\") " pod="openstack/nova-scheduler-0" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.392437 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f69d681-27e5-4fb9-a199-3c6847c80dc9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.394038 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f69d681-27e5-4fb9-a199-3c6847c80dc9-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.394206 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxmm7\" (UniqueName: \"kubernetes.io/projected/4f69d681-27e5-4fb9-a199-3c6847c80dc9-kube-api-access-fxmm7\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.396530 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a893348-52b9-474c-8031-e67020c445eb-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7a893348-52b9-474c-8031-e67020c445eb\") " pod="openstack/nova-scheduler-0" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.397094 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a893348-52b9-474c-8031-e67020c445eb-config-data\") pod \"nova-scheduler-0\" (UID: \"7a893348-52b9-474c-8031-e67020c445eb\") " pod="openstack/nova-scheduler-0" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.412529 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kl7jf\" (UniqueName: \"kubernetes.io/projected/7a893348-52b9-474c-8031-e67020c445eb-kube-api-access-kl7jf\") pod \"nova-scheduler-0\" (UID: \"7a893348-52b9-474c-8031-e67020c445eb\") " pod="openstack/nova-scheduler-0" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.508011 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.572720 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.607916 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.615195 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.617683 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.626172 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.628301 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.717727 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjwmk\" (UniqueName: \"kubernetes.io/projected/adfa8017-1c4c-4b31-90a8-c7e3ac999dd0-kube-api-access-vjwmk\") pod \"nova-api-0\" (UID: \"adfa8017-1c4c-4b31-90a8-c7e3ac999dd0\") " pod="openstack/nova-api-0" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.717839 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adfa8017-1c4c-4b31-90a8-c7e3ac999dd0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"adfa8017-1c4c-4b31-90a8-c7e3ac999dd0\") " pod="openstack/nova-api-0" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.717867 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adfa8017-1c4c-4b31-90a8-c7e3ac999dd0-logs\") pod \"nova-api-0\" (UID: \"adfa8017-1c4c-4b31-90a8-c7e3ac999dd0\") " pod="openstack/nova-api-0" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.717891 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adfa8017-1c4c-4b31-90a8-c7e3ac999dd0-config-data\") pod \"nova-api-0\" (UID: \"adfa8017-1c4c-4b31-90a8-c7e3ac999dd0\") " pod="openstack/nova-api-0" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.819298 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjwmk\" (UniqueName: \"kubernetes.io/projected/adfa8017-1c4c-4b31-90a8-c7e3ac999dd0-kube-api-access-vjwmk\") pod \"nova-api-0\" (UID: \"adfa8017-1c4c-4b31-90a8-c7e3ac999dd0\") " pod="openstack/nova-api-0" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.819388 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adfa8017-1c4c-4b31-90a8-c7e3ac999dd0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"adfa8017-1c4c-4b31-90a8-c7e3ac999dd0\") " pod="openstack/nova-api-0" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.819411 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adfa8017-1c4c-4b31-90a8-c7e3ac999dd0-logs\") pod \"nova-api-0\" (UID: \"adfa8017-1c4c-4b31-90a8-c7e3ac999dd0\") " pod="openstack/nova-api-0" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.819434 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adfa8017-1c4c-4b31-90a8-c7e3ac999dd0-config-data\") pod \"nova-api-0\" (UID: \"adfa8017-1c4c-4b31-90a8-c7e3ac999dd0\") " pod="openstack/nova-api-0" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.821082 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adfa8017-1c4c-4b31-90a8-c7e3ac999dd0-logs\") pod \"nova-api-0\" (UID: \"adfa8017-1c4c-4b31-90a8-c7e3ac999dd0\") " pod="openstack/nova-api-0" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.825135 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adfa8017-1c4c-4b31-90a8-c7e3ac999dd0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"adfa8017-1c4c-4b31-90a8-c7e3ac999dd0\") " pod="openstack/nova-api-0" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.831738 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adfa8017-1c4c-4b31-90a8-c7e3ac999dd0-config-data\") pod \"nova-api-0\" (UID: \"adfa8017-1c4c-4b31-90a8-c7e3ac999dd0\") " pod="openstack/nova-api-0" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.839836 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjwmk\" (UniqueName: \"kubernetes.io/projected/adfa8017-1c4c-4b31-90a8-c7e3ac999dd0-kube-api-access-vjwmk\") pod \"nova-api-0\" (UID: \"adfa8017-1c4c-4b31-90a8-c7e3ac999dd0\") " pod="openstack/nova-api-0" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.941393 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 17 13:25:21 crc kubenswrapper[4955]: I0217 13:25:21.944681 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 17 13:25:21 crc kubenswrapper[4955]: W0217 13:25:21.949569 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7a893348_52b9_474c_8031_e67020c445eb.slice/crio-fe1283912db36870d1dfdadb33f8eb25481b222b96a63fd493c39b168d051535 WatchSource:0}: Error finding container fe1283912db36870d1dfdadb33f8eb25481b222b96a63fd493c39b168d051535: Status 404 returned error can't find the container with id fe1283912db36870d1dfdadb33f8eb25481b222b96a63fd493c39b168d051535 Feb 17 13:25:22 crc kubenswrapper[4955]: I0217 13:25:22.243345 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5" path="/var/lib/kubelet/pods/01100c9a-bf1b-4d2a-bc47-c6f0090ebaa5/volumes" Feb 17 13:25:22 crc kubenswrapper[4955]: I0217 13:25:22.244326 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f69d681-27e5-4fb9-a199-3c6847c80dc9" path="/var/lib/kubelet/pods/4f69d681-27e5-4fb9-a199-3c6847c80dc9/volumes" Feb 17 13:25:22 crc kubenswrapper[4955]: I0217 13:25:22.244914 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7a893348-52b9-474c-8031-e67020c445eb","Type":"ContainerStarted","Data":"5e0370a1ee53a812b14d0cf1a49db5bf79f6f7d91dbbfb8601a5bc8ec1b0f646"} Feb 17 13:25:22 crc kubenswrapper[4955]: I0217 13:25:22.244937 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7a893348-52b9-474c-8031-e67020c445eb","Type":"ContainerStarted","Data":"fe1283912db36870d1dfdadb33f8eb25481b222b96a63fd493c39b168d051535"} Feb 17 13:25:22 crc kubenswrapper[4955]: I0217 13:25:22.264070 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.26405442 podStartE2EDuration="1.26405442s" podCreationTimestamp="2026-02-17 13:25:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:25:22.260986252 +0000 UTC m=+1260.783766716" watchObservedRunningTime="2026-02-17 13:25:22.26405442 +0000 UTC m=+1260.786783963" Feb 17 13:25:22 crc kubenswrapper[4955]: I0217 13:25:22.384316 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 17 13:25:23 crc kubenswrapper[4955]: I0217 13:25:23.267740 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"adfa8017-1c4c-4b31-90a8-c7e3ac999dd0","Type":"ContainerStarted","Data":"767b94202f66a98ef0dea13f1984f9a22753af63d8eb02d11ba3bb1a9b0cbc31"} Feb 17 13:25:23 crc kubenswrapper[4955]: I0217 13:25:23.268090 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"adfa8017-1c4c-4b31-90a8-c7e3ac999dd0","Type":"ContainerStarted","Data":"4627d2789659ddf3ab32efe87cc56e52503a33e443792449cb1ee6abb670881c"} Feb 17 13:25:23 crc kubenswrapper[4955]: I0217 13:25:23.268110 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"adfa8017-1c4c-4b31-90a8-c7e3ac999dd0","Type":"ContainerStarted","Data":"418ba7456744415a6978fdbfc9b893a15d923d2c3a177bfc0e2a5deea64f36ea"} Feb 17 13:25:23 crc kubenswrapper[4955]: I0217 13:25:23.288624 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.288609412 podStartE2EDuration="2.288609412s" podCreationTimestamp="2026-02-17 13:25:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:25:23.285418591 +0000 UTC m=+1261.808148144" watchObservedRunningTime="2026-02-17 13:25:23.288609412 +0000 UTC m=+1261.811338955" Feb 17 13:25:26 crc kubenswrapper[4955]: I0217 13:25:26.508169 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 17 13:25:26 crc kubenswrapper[4955]: I0217 13:25:26.608097 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Feb 17 13:25:31 crc kubenswrapper[4955]: I0217 13:25:31.508378 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 17 13:25:31 crc kubenswrapper[4955]: I0217 13:25:31.549736 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 17 13:25:31 crc kubenswrapper[4955]: I0217 13:25:31.942460 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 17 13:25:31 crc kubenswrapper[4955]: I0217 13:25:31.942510 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 17 13:25:32 crc kubenswrapper[4955]: I0217 13:25:32.388915 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 17 13:25:33 crc kubenswrapper[4955]: I0217 13:25:33.025027 4955 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="adfa8017-1c4c-4b31-90a8-c7e3ac999dd0" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.201:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 17 13:25:33 crc kubenswrapper[4955]: I0217 13:25:33.025134 4955 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="adfa8017-1c4c-4b31-90a8-c7e3ac999dd0" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.201:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 17 13:25:38 crc kubenswrapper[4955]: E0217 13:25:38.415021 4955 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod242f0cab_eded_4013_8ed0_f9ddd4de5fd0.slice/crio-142e9ee4cc1286ea2bc083be3162bcbf9edc08dc9caa0a0b95ec9437e6c3b941.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod242f0cab_eded_4013_8ed0_f9ddd4de5fd0.slice/crio-conmon-142e9ee4cc1286ea2bc083be3162bcbf9edc08dc9caa0a0b95ec9437e6c3b941.scope\": RecentStats: unable to find data in memory cache]" Feb 17 13:25:38 crc kubenswrapper[4955]: I0217 13:25:38.428364 4955 generic.go:334] "Generic (PLEG): container finished" podID="242f0cab-eded-4013-8ed0-f9ddd4de5fd0" containerID="142e9ee4cc1286ea2bc083be3162bcbf9edc08dc9caa0a0b95ec9437e6c3b941" exitCode=137 Feb 17 13:25:38 crc kubenswrapper[4955]: I0217 13:25:38.428454 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"242f0cab-eded-4013-8ed0-f9ddd4de5fd0","Type":"ContainerDied","Data":"142e9ee4cc1286ea2bc083be3162bcbf9edc08dc9caa0a0b95ec9437e6c3b941"} Feb 17 13:25:38 crc kubenswrapper[4955]: I0217 13:25:38.431322 4955 generic.go:334] "Generic (PLEG): container finished" podID="c7a1ed55-8add-4f6d-8210-76ccf42a58ab" containerID="4a08ccf38d2598943d136d7704d1b698199baa699a28452a09e8febea77a5314" exitCode=137 Feb 17 13:25:38 crc kubenswrapper[4955]: I0217 13:25:38.431375 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c7a1ed55-8add-4f6d-8210-76ccf42a58ab","Type":"ContainerDied","Data":"4a08ccf38d2598943d136d7704d1b698199baa699a28452a09e8febea77a5314"} Feb 17 13:25:38 crc kubenswrapper[4955]: I0217 13:25:38.431406 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c7a1ed55-8add-4f6d-8210-76ccf42a58ab","Type":"ContainerDied","Data":"281d1febd28a5c3d81367f3c198ee3b0972633283889f94f1d27281caf057e02"} Feb 17 13:25:38 crc kubenswrapper[4955]: I0217 13:25:38.431417 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="281d1febd28a5c3d81367f3c198ee3b0972633283889f94f1d27281caf057e02" Feb 17 13:25:38 crc kubenswrapper[4955]: I0217 13:25:38.438276 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 17 13:25:38 crc kubenswrapper[4955]: I0217 13:25:38.506394 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 17 13:25:38 crc kubenswrapper[4955]: I0217 13:25:38.525157 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 17 13:25:38 crc kubenswrapper[4955]: I0217 13:25:38.537767 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cp66t\" (UniqueName: \"kubernetes.io/projected/c7a1ed55-8add-4f6d-8210-76ccf42a58ab-kube-api-access-cp66t\") pod \"c7a1ed55-8add-4f6d-8210-76ccf42a58ab\" (UID: \"c7a1ed55-8add-4f6d-8210-76ccf42a58ab\") " Feb 17 13:25:38 crc kubenswrapper[4955]: I0217 13:25:38.538121 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/242f0cab-eded-4013-8ed0-f9ddd4de5fd0-config-data\") pod \"242f0cab-eded-4013-8ed0-f9ddd4de5fd0\" (UID: \"242f0cab-eded-4013-8ed0-f9ddd4de5fd0\") " Feb 17 13:25:38 crc kubenswrapper[4955]: I0217 13:25:38.538209 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/242f0cab-eded-4013-8ed0-f9ddd4de5fd0-combined-ca-bundle\") pod \"242f0cab-eded-4013-8ed0-f9ddd4de5fd0\" (UID: \"242f0cab-eded-4013-8ed0-f9ddd4de5fd0\") " Feb 17 13:25:38 crc kubenswrapper[4955]: I0217 13:25:38.538270 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7a1ed55-8add-4f6d-8210-76ccf42a58ab-combined-ca-bundle\") pod \"c7a1ed55-8add-4f6d-8210-76ccf42a58ab\" (UID: \"c7a1ed55-8add-4f6d-8210-76ccf42a58ab\") " Feb 17 13:25:38 crc kubenswrapper[4955]: I0217 13:25:38.538411 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c7a1ed55-8add-4f6d-8210-76ccf42a58ab-logs\") pod \"c7a1ed55-8add-4f6d-8210-76ccf42a58ab\" (UID: \"c7a1ed55-8add-4f6d-8210-76ccf42a58ab\") " Feb 17 13:25:38 crc kubenswrapper[4955]: I0217 13:25:38.538455 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrtph\" (UniqueName: \"kubernetes.io/projected/242f0cab-eded-4013-8ed0-f9ddd4de5fd0-kube-api-access-nrtph\") pod \"242f0cab-eded-4013-8ed0-f9ddd4de5fd0\" (UID: \"242f0cab-eded-4013-8ed0-f9ddd4de5fd0\") " Feb 17 13:25:38 crc kubenswrapper[4955]: I0217 13:25:38.538508 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7a1ed55-8add-4f6d-8210-76ccf42a58ab-config-data\") pod \"c7a1ed55-8add-4f6d-8210-76ccf42a58ab\" (UID: \"c7a1ed55-8add-4f6d-8210-76ccf42a58ab\") " Feb 17 13:25:38 crc kubenswrapper[4955]: I0217 13:25:38.538843 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7a1ed55-8add-4f6d-8210-76ccf42a58ab-logs" (OuterVolumeSpecName: "logs") pod "c7a1ed55-8add-4f6d-8210-76ccf42a58ab" (UID: "c7a1ed55-8add-4f6d-8210-76ccf42a58ab"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:25:38 crc kubenswrapper[4955]: I0217 13:25:38.539269 4955 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c7a1ed55-8add-4f6d-8210-76ccf42a58ab-logs\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:38 crc kubenswrapper[4955]: I0217 13:25:38.547439 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7a1ed55-8add-4f6d-8210-76ccf42a58ab-kube-api-access-cp66t" (OuterVolumeSpecName: "kube-api-access-cp66t") pod "c7a1ed55-8add-4f6d-8210-76ccf42a58ab" (UID: "c7a1ed55-8add-4f6d-8210-76ccf42a58ab"). InnerVolumeSpecName "kube-api-access-cp66t". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:25:38 crc kubenswrapper[4955]: I0217 13:25:38.548277 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/242f0cab-eded-4013-8ed0-f9ddd4de5fd0-kube-api-access-nrtph" (OuterVolumeSpecName: "kube-api-access-nrtph") pod "242f0cab-eded-4013-8ed0-f9ddd4de5fd0" (UID: "242f0cab-eded-4013-8ed0-f9ddd4de5fd0"). InnerVolumeSpecName "kube-api-access-nrtph". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:25:38 crc kubenswrapper[4955]: I0217 13:25:38.568673 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/242f0cab-eded-4013-8ed0-f9ddd4de5fd0-config-data" (OuterVolumeSpecName: "config-data") pod "242f0cab-eded-4013-8ed0-f9ddd4de5fd0" (UID: "242f0cab-eded-4013-8ed0-f9ddd4de5fd0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:25:38 crc kubenswrapper[4955]: I0217 13:25:38.581346 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/242f0cab-eded-4013-8ed0-f9ddd4de5fd0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "242f0cab-eded-4013-8ed0-f9ddd4de5fd0" (UID: "242f0cab-eded-4013-8ed0-f9ddd4de5fd0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:25:38 crc kubenswrapper[4955]: I0217 13:25:38.590726 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7a1ed55-8add-4f6d-8210-76ccf42a58ab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c7a1ed55-8add-4f6d-8210-76ccf42a58ab" (UID: "c7a1ed55-8add-4f6d-8210-76ccf42a58ab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:25:38 crc kubenswrapper[4955]: I0217 13:25:38.595073 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7a1ed55-8add-4f6d-8210-76ccf42a58ab-config-data" (OuterVolumeSpecName: "config-data") pod "c7a1ed55-8add-4f6d-8210-76ccf42a58ab" (UID: "c7a1ed55-8add-4f6d-8210-76ccf42a58ab"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:25:38 crc kubenswrapper[4955]: I0217 13:25:38.640967 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cp66t\" (UniqueName: \"kubernetes.io/projected/c7a1ed55-8add-4f6d-8210-76ccf42a58ab-kube-api-access-cp66t\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:38 crc kubenswrapper[4955]: I0217 13:25:38.641001 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/242f0cab-eded-4013-8ed0-f9ddd4de5fd0-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:38 crc kubenswrapper[4955]: I0217 13:25:38.641011 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/242f0cab-eded-4013-8ed0-f9ddd4de5fd0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:38 crc kubenswrapper[4955]: I0217 13:25:38.641021 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7a1ed55-8add-4f6d-8210-76ccf42a58ab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:38 crc kubenswrapper[4955]: I0217 13:25:38.641030 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrtph\" (UniqueName: \"kubernetes.io/projected/242f0cab-eded-4013-8ed0-f9ddd4de5fd0-kube-api-access-nrtph\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:38 crc kubenswrapper[4955]: I0217 13:25:38.641038 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7a1ed55-8add-4f6d-8210-76ccf42a58ab-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.441765 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.441756 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"242f0cab-eded-4013-8ed0-f9ddd4de5fd0","Type":"ContainerDied","Data":"73b19a13056adeea2860efe3a1fa16447b7dcb9ffd294f69964d6b214a0db361"} Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.442257 4955 scope.go:117] "RemoveContainer" containerID="142e9ee4cc1286ea2bc083be3162bcbf9edc08dc9caa0a0b95ec9437e6c3b941" Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.441844 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.580105 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.619870 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.633855 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.674114 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 17 13:25:39 crc kubenswrapper[4955]: E0217 13:25:39.674512 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="242f0cab-eded-4013-8ed0-f9ddd4de5fd0" containerName="nova-cell1-novncproxy-novncproxy" Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.674527 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="242f0cab-eded-4013-8ed0-f9ddd4de5fd0" containerName="nova-cell1-novncproxy-novncproxy" Feb 17 13:25:39 crc kubenswrapper[4955]: E0217 13:25:39.674548 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7a1ed55-8add-4f6d-8210-76ccf42a58ab" containerName="nova-metadata-metadata" Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.674554 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7a1ed55-8add-4f6d-8210-76ccf42a58ab" containerName="nova-metadata-metadata" Feb 17 13:25:39 crc kubenswrapper[4955]: E0217 13:25:39.674580 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7a1ed55-8add-4f6d-8210-76ccf42a58ab" containerName="nova-metadata-log" Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.674586 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7a1ed55-8add-4f6d-8210-76ccf42a58ab" containerName="nova-metadata-log" Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.674796 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7a1ed55-8add-4f6d-8210-76ccf42a58ab" containerName="nova-metadata-log" Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.674821 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="242f0cab-eded-4013-8ed0-f9ddd4de5fd0" containerName="nova-cell1-novncproxy-novncproxy" Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.674840 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7a1ed55-8add-4f6d-8210-76ccf42a58ab" containerName="nova-metadata-metadata" Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.675954 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.681187 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.681244 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.684869 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.692850 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.707295 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.709430 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.721312 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.721510 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.721659 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.721800 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.861926 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3fe1264c-d2ce-455a-99b9-88270f4dfb04-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3fe1264c-d2ce-455a-99b9-88270f4dfb04\") " pod="openstack/nova-metadata-0" Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.862004 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fe1264c-d2ce-455a-99b9-88270f4dfb04-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3fe1264c-d2ce-455a-99b9-88270f4dfb04\") " pod="openstack/nova-metadata-0" Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.862038 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3614da91-13c7-4d09-b508-933322e35f08-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3614da91-13c7-4d09-b508-933322e35f08\") " pod="openstack/nova-cell1-novncproxy-0" Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.862090 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fe1264c-d2ce-455a-99b9-88270f4dfb04-config-data\") pod \"nova-metadata-0\" (UID: \"3fe1264c-d2ce-455a-99b9-88270f4dfb04\") " pod="openstack/nova-metadata-0" Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.862170 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxckz\" (UniqueName: \"kubernetes.io/projected/3fe1264c-d2ce-455a-99b9-88270f4dfb04-kube-api-access-vxckz\") pod \"nova-metadata-0\" (UID: \"3fe1264c-d2ce-455a-99b9-88270f4dfb04\") " pod="openstack/nova-metadata-0" Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.862199 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5k727\" (UniqueName: \"kubernetes.io/projected/3614da91-13c7-4d09-b508-933322e35f08-kube-api-access-5k727\") pod \"nova-cell1-novncproxy-0\" (UID: \"3614da91-13c7-4d09-b508-933322e35f08\") " pod="openstack/nova-cell1-novncproxy-0" Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.862237 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/3614da91-13c7-4d09-b508-933322e35f08-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"3614da91-13c7-4d09-b508-933322e35f08\") " pod="openstack/nova-cell1-novncproxy-0" Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.862266 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3614da91-13c7-4d09-b508-933322e35f08-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3614da91-13c7-4d09-b508-933322e35f08\") " pod="openstack/nova-cell1-novncproxy-0" Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.862366 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/3614da91-13c7-4d09-b508-933322e35f08-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"3614da91-13c7-4d09-b508-933322e35f08\") " pod="openstack/nova-cell1-novncproxy-0" Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.862390 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3fe1264c-d2ce-455a-99b9-88270f4dfb04-logs\") pod \"nova-metadata-0\" (UID: \"3fe1264c-d2ce-455a-99b9-88270f4dfb04\") " pod="openstack/nova-metadata-0" Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.964251 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fe1264c-d2ce-455a-99b9-88270f4dfb04-config-data\") pod \"nova-metadata-0\" (UID: \"3fe1264c-d2ce-455a-99b9-88270f4dfb04\") " pod="openstack/nova-metadata-0" Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.964316 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxckz\" (UniqueName: \"kubernetes.io/projected/3fe1264c-d2ce-455a-99b9-88270f4dfb04-kube-api-access-vxckz\") pod \"nova-metadata-0\" (UID: \"3fe1264c-d2ce-455a-99b9-88270f4dfb04\") " pod="openstack/nova-metadata-0" Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.964335 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5k727\" (UniqueName: \"kubernetes.io/projected/3614da91-13c7-4d09-b508-933322e35f08-kube-api-access-5k727\") pod \"nova-cell1-novncproxy-0\" (UID: \"3614da91-13c7-4d09-b508-933322e35f08\") " pod="openstack/nova-cell1-novncproxy-0" Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.964360 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/3614da91-13c7-4d09-b508-933322e35f08-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"3614da91-13c7-4d09-b508-933322e35f08\") " pod="openstack/nova-cell1-novncproxy-0" Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.964384 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3614da91-13c7-4d09-b508-933322e35f08-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3614da91-13c7-4d09-b508-933322e35f08\") " pod="openstack/nova-cell1-novncproxy-0" Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.964465 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/3614da91-13c7-4d09-b508-933322e35f08-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"3614da91-13c7-4d09-b508-933322e35f08\") " pod="openstack/nova-cell1-novncproxy-0" Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.964484 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3fe1264c-d2ce-455a-99b9-88270f4dfb04-logs\") pod \"nova-metadata-0\" (UID: \"3fe1264c-d2ce-455a-99b9-88270f4dfb04\") " pod="openstack/nova-metadata-0" Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.964516 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3fe1264c-d2ce-455a-99b9-88270f4dfb04-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3fe1264c-d2ce-455a-99b9-88270f4dfb04\") " pod="openstack/nova-metadata-0" Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.964539 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fe1264c-d2ce-455a-99b9-88270f4dfb04-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3fe1264c-d2ce-455a-99b9-88270f4dfb04\") " pod="openstack/nova-metadata-0" Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.964556 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3614da91-13c7-4d09-b508-933322e35f08-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3614da91-13c7-4d09-b508-933322e35f08\") " pod="openstack/nova-cell1-novncproxy-0" Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.965939 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3fe1264c-d2ce-455a-99b9-88270f4dfb04-logs\") pod \"nova-metadata-0\" (UID: \"3fe1264c-d2ce-455a-99b9-88270f4dfb04\") " pod="openstack/nova-metadata-0" Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.970449 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/3614da91-13c7-4d09-b508-933322e35f08-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"3614da91-13c7-4d09-b508-933322e35f08\") " pod="openstack/nova-cell1-novncproxy-0" Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.973522 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fe1264c-d2ce-455a-99b9-88270f4dfb04-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3fe1264c-d2ce-455a-99b9-88270f4dfb04\") " pod="openstack/nova-metadata-0" Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.985404 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxckz\" (UniqueName: \"kubernetes.io/projected/3fe1264c-d2ce-455a-99b9-88270f4dfb04-kube-api-access-vxckz\") pod \"nova-metadata-0\" (UID: \"3fe1264c-d2ce-455a-99b9-88270f4dfb04\") " pod="openstack/nova-metadata-0" Feb 17 13:25:39 crc kubenswrapper[4955]: I0217 13:25:39.994470 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5k727\" (UniqueName: \"kubernetes.io/projected/3614da91-13c7-4d09-b508-933322e35f08-kube-api-access-5k727\") pod \"nova-cell1-novncproxy-0\" (UID: \"3614da91-13c7-4d09-b508-933322e35f08\") " pod="openstack/nova-cell1-novncproxy-0" Feb 17 13:25:40 crc kubenswrapper[4955]: I0217 13:25:40.006698 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3fe1264c-d2ce-455a-99b9-88270f4dfb04-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3fe1264c-d2ce-455a-99b9-88270f4dfb04\") " pod="openstack/nova-metadata-0" Feb 17 13:25:40 crc kubenswrapper[4955]: I0217 13:25:40.006698 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/3614da91-13c7-4d09-b508-933322e35f08-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"3614da91-13c7-4d09-b508-933322e35f08\") " pod="openstack/nova-cell1-novncproxy-0" Feb 17 13:25:40 crc kubenswrapper[4955]: I0217 13:25:40.006926 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fe1264c-d2ce-455a-99b9-88270f4dfb04-config-data\") pod \"nova-metadata-0\" (UID: \"3fe1264c-d2ce-455a-99b9-88270f4dfb04\") " pod="openstack/nova-metadata-0" Feb 17 13:25:40 crc kubenswrapper[4955]: I0217 13:25:40.008618 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3614da91-13c7-4d09-b508-933322e35f08-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3614da91-13c7-4d09-b508-933322e35f08\") " pod="openstack/nova-cell1-novncproxy-0" Feb 17 13:25:40 crc kubenswrapper[4955]: I0217 13:25:40.009557 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3614da91-13c7-4d09-b508-933322e35f08-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3614da91-13c7-4d09-b508-933322e35f08\") " pod="openstack/nova-cell1-novncproxy-0" Feb 17 13:25:40 crc kubenswrapper[4955]: I0217 13:25:40.013516 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 17 13:25:40 crc kubenswrapper[4955]: I0217 13:25:40.045574 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 17 13:25:40 crc kubenswrapper[4955]: I0217 13:25:40.234594 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="242f0cab-eded-4013-8ed0-f9ddd4de5fd0" path="/var/lib/kubelet/pods/242f0cab-eded-4013-8ed0-f9ddd4de5fd0/volumes" Feb 17 13:25:40 crc kubenswrapper[4955]: I0217 13:25:40.235183 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7a1ed55-8add-4f6d-8210-76ccf42a58ab" path="/var/lib/kubelet/pods/c7a1ed55-8add-4f6d-8210-76ccf42a58ab/volumes" Feb 17 13:25:40 crc kubenswrapper[4955]: I0217 13:25:40.485730 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 17 13:25:40 crc kubenswrapper[4955]: I0217 13:25:40.586449 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 17 13:25:40 crc kubenswrapper[4955]: W0217 13:25:40.615118 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3614da91_13c7_4d09_b508_933322e35f08.slice/crio-2fd08a05064d5c8c91760448e84137ae58fe46a52237103bc66388d90e357dbd WatchSource:0}: Error finding container 2fd08a05064d5c8c91760448e84137ae58fe46a52237103bc66388d90e357dbd: Status 404 returned error can't find the container with id 2fd08a05064d5c8c91760448e84137ae58fe46a52237103bc66388d90e357dbd Feb 17 13:25:41 crc kubenswrapper[4955]: I0217 13:25:41.463978 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3fe1264c-d2ce-455a-99b9-88270f4dfb04","Type":"ContainerStarted","Data":"fa8c554eed043625d8a3e30ac076a7f926ace945ba0d86ad179212d29d9bae2e"} Feb 17 13:25:41 crc kubenswrapper[4955]: I0217 13:25:41.464743 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3fe1264c-d2ce-455a-99b9-88270f4dfb04","Type":"ContainerStarted","Data":"55bcc64e4aacdb624eb82ffeda3b66fe603131d8da1ae9d31139305217a5f598"} Feb 17 13:25:41 crc kubenswrapper[4955]: I0217 13:25:41.464772 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3fe1264c-d2ce-455a-99b9-88270f4dfb04","Type":"ContainerStarted","Data":"e29f177fa16272579c55e170117e853d9f3bfbfd1a79199137846ee8ef65ec82"} Feb 17 13:25:41 crc kubenswrapper[4955]: I0217 13:25:41.466451 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3614da91-13c7-4d09-b508-933322e35f08","Type":"ContainerStarted","Data":"2ed3c3a5b497ad6cd2921169007d15dc6651bbaecaaf316c61da6713331e1e14"} Feb 17 13:25:41 crc kubenswrapper[4955]: I0217 13:25:41.466501 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3614da91-13c7-4d09-b508-933322e35f08","Type":"ContainerStarted","Data":"2fd08a05064d5c8c91760448e84137ae58fe46a52237103bc66388d90e357dbd"} Feb 17 13:25:41 crc kubenswrapper[4955]: I0217 13:25:41.495643 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.4956206180000002 podStartE2EDuration="2.495620618s" podCreationTimestamp="2026-02-17 13:25:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:25:41.483874271 +0000 UTC m=+1280.006603834" watchObservedRunningTime="2026-02-17 13:25:41.495620618 +0000 UTC m=+1280.018350171" Feb 17 13:25:41 crc kubenswrapper[4955]: I0217 13:25:41.946081 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 17 13:25:41 crc kubenswrapper[4955]: I0217 13:25:41.946701 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 17 13:25:41 crc kubenswrapper[4955]: I0217 13:25:41.948726 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 17 13:25:41 crc kubenswrapper[4955]: I0217 13:25:41.950380 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 17 13:25:42 crc kubenswrapper[4955]: I0217 13:25:42.005830 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.005807419 podStartE2EDuration="3.005807419s" podCreationTimestamp="2026-02-17 13:25:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:25:41.505777699 +0000 UTC m=+1280.028507262" watchObservedRunningTime="2026-02-17 13:25:42.005807419 +0000 UTC m=+1280.528536962" Feb 17 13:25:42 crc kubenswrapper[4955]: I0217 13:25:42.475402 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 17 13:25:42 crc kubenswrapper[4955]: I0217 13:25:42.479663 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 17 13:25:42 crc kubenswrapper[4955]: I0217 13:25:42.678141 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-9fmsq"] Feb 17 13:25:42 crc kubenswrapper[4955]: I0217 13:25:42.680073 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-9fmsq" Feb 17 13:25:42 crc kubenswrapper[4955]: I0217 13:25:42.689817 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-9fmsq"] Feb 17 13:25:42 crc kubenswrapper[4955]: I0217 13:25:42.824127 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ea144e9-32f5-4f5d-85fd-a6ec6e73add3-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-9fmsq\" (UID: \"3ea144e9-32f5-4f5d-85fd-a6ec6e73add3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-9fmsq" Feb 17 13:25:42 crc kubenswrapper[4955]: I0217 13:25:42.824231 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3ea144e9-32f5-4f5d-85fd-a6ec6e73add3-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-9fmsq\" (UID: \"3ea144e9-32f5-4f5d-85fd-a6ec6e73add3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-9fmsq" Feb 17 13:25:42 crc kubenswrapper[4955]: I0217 13:25:42.824263 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ea144e9-32f5-4f5d-85fd-a6ec6e73add3-config\") pod \"dnsmasq-dns-cd5cbd7b9-9fmsq\" (UID: \"3ea144e9-32f5-4f5d-85fd-a6ec6e73add3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-9fmsq" Feb 17 13:25:42 crc kubenswrapper[4955]: I0217 13:25:42.824338 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ea144e9-32f5-4f5d-85fd-a6ec6e73add3-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-9fmsq\" (UID: \"3ea144e9-32f5-4f5d-85fd-a6ec6e73add3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-9fmsq" Feb 17 13:25:42 crc kubenswrapper[4955]: I0217 13:25:42.824395 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdq8t\" (UniqueName: \"kubernetes.io/projected/3ea144e9-32f5-4f5d-85fd-a6ec6e73add3-kube-api-access-zdq8t\") pod \"dnsmasq-dns-cd5cbd7b9-9fmsq\" (UID: \"3ea144e9-32f5-4f5d-85fd-a6ec6e73add3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-9fmsq" Feb 17 13:25:42 crc kubenswrapper[4955]: I0217 13:25:42.824434 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ea144e9-32f5-4f5d-85fd-a6ec6e73add3-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-9fmsq\" (UID: \"3ea144e9-32f5-4f5d-85fd-a6ec6e73add3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-9fmsq" Feb 17 13:25:42 crc kubenswrapper[4955]: I0217 13:25:42.925955 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdq8t\" (UniqueName: \"kubernetes.io/projected/3ea144e9-32f5-4f5d-85fd-a6ec6e73add3-kube-api-access-zdq8t\") pod \"dnsmasq-dns-cd5cbd7b9-9fmsq\" (UID: \"3ea144e9-32f5-4f5d-85fd-a6ec6e73add3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-9fmsq" Feb 17 13:25:42 crc kubenswrapper[4955]: I0217 13:25:42.926009 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ea144e9-32f5-4f5d-85fd-a6ec6e73add3-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-9fmsq\" (UID: \"3ea144e9-32f5-4f5d-85fd-a6ec6e73add3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-9fmsq" Feb 17 13:25:42 crc kubenswrapper[4955]: I0217 13:25:42.926076 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ea144e9-32f5-4f5d-85fd-a6ec6e73add3-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-9fmsq\" (UID: \"3ea144e9-32f5-4f5d-85fd-a6ec6e73add3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-9fmsq" Feb 17 13:25:42 crc kubenswrapper[4955]: I0217 13:25:42.926139 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3ea144e9-32f5-4f5d-85fd-a6ec6e73add3-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-9fmsq\" (UID: \"3ea144e9-32f5-4f5d-85fd-a6ec6e73add3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-9fmsq" Feb 17 13:25:42 crc kubenswrapper[4955]: I0217 13:25:42.926203 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ea144e9-32f5-4f5d-85fd-a6ec6e73add3-config\") pod \"dnsmasq-dns-cd5cbd7b9-9fmsq\" (UID: \"3ea144e9-32f5-4f5d-85fd-a6ec6e73add3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-9fmsq" Feb 17 13:25:42 crc kubenswrapper[4955]: I0217 13:25:42.927065 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3ea144e9-32f5-4f5d-85fd-a6ec6e73add3-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-9fmsq\" (UID: \"3ea144e9-32f5-4f5d-85fd-a6ec6e73add3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-9fmsq" Feb 17 13:25:42 crc kubenswrapper[4955]: I0217 13:25:42.927151 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ea144e9-32f5-4f5d-85fd-a6ec6e73add3-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-9fmsq\" (UID: \"3ea144e9-32f5-4f5d-85fd-a6ec6e73add3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-9fmsq" Feb 17 13:25:42 crc kubenswrapper[4955]: I0217 13:25:42.927160 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ea144e9-32f5-4f5d-85fd-a6ec6e73add3-config\") pod \"dnsmasq-dns-cd5cbd7b9-9fmsq\" (UID: \"3ea144e9-32f5-4f5d-85fd-a6ec6e73add3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-9fmsq" Feb 17 13:25:42 crc kubenswrapper[4955]: I0217 13:25:42.928235 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ea144e9-32f5-4f5d-85fd-a6ec6e73add3-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-9fmsq\" (UID: \"3ea144e9-32f5-4f5d-85fd-a6ec6e73add3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-9fmsq" Feb 17 13:25:42 crc kubenswrapper[4955]: I0217 13:25:42.928354 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ea144e9-32f5-4f5d-85fd-a6ec6e73add3-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-9fmsq\" (UID: \"3ea144e9-32f5-4f5d-85fd-a6ec6e73add3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-9fmsq" Feb 17 13:25:42 crc kubenswrapper[4955]: I0217 13:25:42.928604 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ea144e9-32f5-4f5d-85fd-a6ec6e73add3-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-9fmsq\" (UID: \"3ea144e9-32f5-4f5d-85fd-a6ec6e73add3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-9fmsq" Feb 17 13:25:42 crc kubenswrapper[4955]: I0217 13:25:42.951140 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdq8t\" (UniqueName: \"kubernetes.io/projected/3ea144e9-32f5-4f5d-85fd-a6ec6e73add3-kube-api-access-zdq8t\") pod \"dnsmasq-dns-cd5cbd7b9-9fmsq\" (UID: \"3ea144e9-32f5-4f5d-85fd-a6ec6e73add3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-9fmsq" Feb 17 13:25:43 crc kubenswrapper[4955]: I0217 13:25:43.001606 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-9fmsq" Feb 17 13:25:43 crc kubenswrapper[4955]: I0217 13:25:43.577151 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-9fmsq"] Feb 17 13:25:43 crc kubenswrapper[4955]: W0217 13:25:43.578027 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ea144e9_32f5_4f5d_85fd_a6ec6e73add3.slice/crio-b877b588b2d967a7f359cfb8338966090fa0ab03a837b925937fe7e9e3b3562d WatchSource:0}: Error finding container b877b588b2d967a7f359cfb8338966090fa0ab03a837b925937fe7e9e3b3562d: Status 404 returned error can't find the container with id b877b588b2d967a7f359cfb8338966090fa0ab03a837b925937fe7e9e3b3562d Feb 17 13:25:44 crc kubenswrapper[4955]: I0217 13:25:44.493322 4955 generic.go:334] "Generic (PLEG): container finished" podID="3ea144e9-32f5-4f5d-85fd-a6ec6e73add3" containerID="22214ad7f31c951e9c4565fdbc15df40fec663e16d96a70ca906278b9bd1a11a" exitCode=0 Feb 17 13:25:44 crc kubenswrapper[4955]: I0217 13:25:44.493377 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-9fmsq" event={"ID":"3ea144e9-32f5-4f5d-85fd-a6ec6e73add3","Type":"ContainerDied","Data":"22214ad7f31c951e9c4565fdbc15df40fec663e16d96a70ca906278b9bd1a11a"} Feb 17 13:25:44 crc kubenswrapper[4955]: I0217 13:25:44.493455 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-9fmsq" event={"ID":"3ea144e9-32f5-4f5d-85fd-a6ec6e73add3","Type":"ContainerStarted","Data":"b877b588b2d967a7f359cfb8338966090fa0ab03a837b925937fe7e9e3b3562d"} Feb 17 13:25:44 crc kubenswrapper[4955]: I0217 13:25:44.954086 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:25:44 crc kubenswrapper[4955]: I0217 13:25:44.955150 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e1c521d2-55b0-4386-9375-7c7b64d1610f" containerName="sg-core" containerID="cri-o://1ecfa2946eb9cafc9cad6f71311771dcc302de5797b49660b697cd2042a0fff9" gracePeriod=30 Feb 17 13:25:44 crc kubenswrapper[4955]: I0217 13:25:44.955188 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e1c521d2-55b0-4386-9375-7c7b64d1610f" containerName="proxy-httpd" containerID="cri-o://61150ce92008ba120f051cc2960e790a0f85e45b49f99bc87e4d000b7f6f0af5" gracePeriod=30 Feb 17 13:25:44 crc kubenswrapper[4955]: I0217 13:25:44.955413 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e1c521d2-55b0-4386-9375-7c7b64d1610f" containerName="ceilometer-notification-agent" containerID="cri-o://744b17f1d9b104c8caa4129128acc8dc79daba9c743ce68b87f254b512b4d903" gracePeriod=30 Feb 17 13:25:44 crc kubenswrapper[4955]: I0217 13:25:44.956132 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e1c521d2-55b0-4386-9375-7c7b64d1610f" containerName="ceilometer-central-agent" containerID="cri-o://089fe3dff83ab164a4663da073016319402cb07ed7779e22c97e9d51bd600efa" gracePeriod=30 Feb 17 13:25:45 crc kubenswrapper[4955]: I0217 13:25:45.013844 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 17 13:25:45 crc kubenswrapper[4955]: I0217 13:25:45.013912 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 17 13:25:45 crc kubenswrapper[4955]: I0217 13:25:45.046982 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Feb 17 13:25:45 crc kubenswrapper[4955]: I0217 13:25:45.311140 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 17 13:25:45 crc kubenswrapper[4955]: I0217 13:25:45.509864 4955 generic.go:334] "Generic (PLEG): container finished" podID="e1c521d2-55b0-4386-9375-7c7b64d1610f" containerID="61150ce92008ba120f051cc2960e790a0f85e45b49f99bc87e4d000b7f6f0af5" exitCode=0 Feb 17 13:25:45 crc kubenswrapper[4955]: I0217 13:25:45.509895 4955 generic.go:334] "Generic (PLEG): container finished" podID="e1c521d2-55b0-4386-9375-7c7b64d1610f" containerID="1ecfa2946eb9cafc9cad6f71311771dcc302de5797b49660b697cd2042a0fff9" exitCode=2 Feb 17 13:25:45 crc kubenswrapper[4955]: I0217 13:25:45.509903 4955 generic.go:334] "Generic (PLEG): container finished" podID="e1c521d2-55b0-4386-9375-7c7b64d1610f" containerID="089fe3dff83ab164a4663da073016319402cb07ed7779e22c97e9d51bd600efa" exitCode=0 Feb 17 13:25:45 crc kubenswrapper[4955]: I0217 13:25:45.509951 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1c521d2-55b0-4386-9375-7c7b64d1610f","Type":"ContainerDied","Data":"61150ce92008ba120f051cc2960e790a0f85e45b49f99bc87e4d000b7f6f0af5"} Feb 17 13:25:45 crc kubenswrapper[4955]: I0217 13:25:45.509987 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1c521d2-55b0-4386-9375-7c7b64d1610f","Type":"ContainerDied","Data":"1ecfa2946eb9cafc9cad6f71311771dcc302de5797b49660b697cd2042a0fff9"} Feb 17 13:25:45 crc kubenswrapper[4955]: I0217 13:25:45.509997 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1c521d2-55b0-4386-9375-7c7b64d1610f","Type":"ContainerDied","Data":"089fe3dff83ab164a4663da073016319402cb07ed7779e22c97e9d51bd600efa"} Feb 17 13:25:45 crc kubenswrapper[4955]: I0217 13:25:45.512129 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="adfa8017-1c4c-4b31-90a8-c7e3ac999dd0" containerName="nova-api-log" containerID="cri-o://4627d2789659ddf3ab32efe87cc56e52503a33e443792449cb1ee6abb670881c" gracePeriod=30 Feb 17 13:25:45 crc kubenswrapper[4955]: I0217 13:25:45.512301 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-9fmsq" event={"ID":"3ea144e9-32f5-4f5d-85fd-a6ec6e73add3","Type":"ContainerStarted","Data":"8b42b497e8dd742535e0e73c9eb2c77ffc626ae2e0cc6c6f483d434219efb640"} Feb 17 13:25:45 crc kubenswrapper[4955]: I0217 13:25:45.512506 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cd5cbd7b9-9fmsq" Feb 17 13:25:45 crc kubenswrapper[4955]: I0217 13:25:45.512558 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="adfa8017-1c4c-4b31-90a8-c7e3ac999dd0" containerName="nova-api-api" containerID="cri-o://767b94202f66a98ef0dea13f1984f9a22753af63d8eb02d11ba3bb1a9b0cbc31" gracePeriod=30 Feb 17 13:25:45 crc kubenswrapper[4955]: I0217 13:25:45.538214 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cd5cbd7b9-9fmsq" podStartSLOduration=3.5381952119999998 podStartE2EDuration="3.538195212s" podCreationTimestamp="2026-02-17 13:25:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:25:45.535969638 +0000 UTC m=+1284.058699181" watchObservedRunningTime="2026-02-17 13:25:45.538195212 +0000 UTC m=+1284.060924755" Feb 17 13:25:46 crc kubenswrapper[4955]: I0217 13:25:46.525347 4955 generic.go:334] "Generic (PLEG): container finished" podID="adfa8017-1c4c-4b31-90a8-c7e3ac999dd0" containerID="4627d2789659ddf3ab32efe87cc56e52503a33e443792449cb1ee6abb670881c" exitCode=143 Feb 17 13:25:46 crc kubenswrapper[4955]: I0217 13:25:46.525450 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"adfa8017-1c4c-4b31-90a8-c7e3ac999dd0","Type":"ContainerDied","Data":"4627d2789659ddf3ab32efe87cc56e52503a33e443792449cb1ee6abb670881c"} Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.442896 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.539075 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1c521d2-55b0-4386-9375-7c7b64d1610f-scripts\") pod \"e1c521d2-55b0-4386-9375-7c7b64d1610f\" (UID: \"e1c521d2-55b0-4386-9375-7c7b64d1610f\") " Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.539165 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1c521d2-55b0-4386-9375-7c7b64d1610f-run-httpd\") pod \"e1c521d2-55b0-4386-9375-7c7b64d1610f\" (UID: \"e1c521d2-55b0-4386-9375-7c7b64d1610f\") " Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.539190 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1c521d2-55b0-4386-9375-7c7b64d1610f-config-data\") pod \"e1c521d2-55b0-4386-9375-7c7b64d1610f\" (UID: \"e1c521d2-55b0-4386-9375-7c7b64d1610f\") " Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.539250 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kccmt\" (UniqueName: \"kubernetes.io/projected/e1c521d2-55b0-4386-9375-7c7b64d1610f-kube-api-access-kccmt\") pod \"e1c521d2-55b0-4386-9375-7c7b64d1610f\" (UID: \"e1c521d2-55b0-4386-9375-7c7b64d1610f\") " Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.539272 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1c521d2-55b0-4386-9375-7c7b64d1610f-combined-ca-bundle\") pod \"e1c521d2-55b0-4386-9375-7c7b64d1610f\" (UID: \"e1c521d2-55b0-4386-9375-7c7b64d1610f\") " Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.539325 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1c521d2-55b0-4386-9375-7c7b64d1610f-ceilometer-tls-certs\") pod \"e1c521d2-55b0-4386-9375-7c7b64d1610f\" (UID: \"e1c521d2-55b0-4386-9375-7c7b64d1610f\") " Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.539360 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1c521d2-55b0-4386-9375-7c7b64d1610f-log-httpd\") pod \"e1c521d2-55b0-4386-9375-7c7b64d1610f\" (UID: \"e1c521d2-55b0-4386-9375-7c7b64d1610f\") " Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.540156 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e1c521d2-55b0-4386-9375-7c7b64d1610f-sg-core-conf-yaml\") pod \"e1c521d2-55b0-4386-9375-7c7b64d1610f\" (UID: \"e1c521d2-55b0-4386-9375-7c7b64d1610f\") " Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.540356 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1c521d2-55b0-4386-9375-7c7b64d1610f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e1c521d2-55b0-4386-9375-7c7b64d1610f" (UID: "e1c521d2-55b0-4386-9375-7c7b64d1610f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.540825 4955 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1c521d2-55b0-4386-9375-7c7b64d1610f-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.542495 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1c521d2-55b0-4386-9375-7c7b64d1610f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e1c521d2-55b0-4386-9375-7c7b64d1610f" (UID: "e1c521d2-55b0-4386-9375-7c7b64d1610f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.547376 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1c521d2-55b0-4386-9375-7c7b64d1610f-kube-api-access-kccmt" (OuterVolumeSpecName: "kube-api-access-kccmt") pod "e1c521d2-55b0-4386-9375-7c7b64d1610f" (UID: "e1c521d2-55b0-4386-9375-7c7b64d1610f"). InnerVolumeSpecName "kube-api-access-kccmt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.550312 4955 generic.go:334] "Generic (PLEG): container finished" podID="e1c521d2-55b0-4386-9375-7c7b64d1610f" containerID="744b17f1d9b104c8caa4129128acc8dc79daba9c743ce68b87f254b512b4d903" exitCode=0 Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.550354 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1c521d2-55b0-4386-9375-7c7b64d1610f","Type":"ContainerDied","Data":"744b17f1d9b104c8caa4129128acc8dc79daba9c743ce68b87f254b512b4d903"} Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.550403 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1c521d2-55b0-4386-9375-7c7b64d1610f","Type":"ContainerDied","Data":"a129372e16752f4b32b48de60cd95e470f523f4ab09f370e483ead6fd71565b9"} Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.550422 4955 scope.go:117] "RemoveContainer" containerID="61150ce92008ba120f051cc2960e790a0f85e45b49f99bc87e4d000b7f6f0af5" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.550692 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.552088 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1c521d2-55b0-4386-9375-7c7b64d1610f-scripts" (OuterVolumeSpecName: "scripts") pod "e1c521d2-55b0-4386-9375-7c7b64d1610f" (UID: "e1c521d2-55b0-4386-9375-7c7b64d1610f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.608578 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1c521d2-55b0-4386-9375-7c7b64d1610f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e1c521d2-55b0-4386-9375-7c7b64d1610f" (UID: "e1c521d2-55b0-4386-9375-7c7b64d1610f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.641687 4955 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1c521d2-55b0-4386-9375-7c7b64d1610f-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.641725 4955 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e1c521d2-55b0-4386-9375-7c7b64d1610f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.641740 4955 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1c521d2-55b0-4386-9375-7c7b64d1610f-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.641752 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kccmt\" (UniqueName: \"kubernetes.io/projected/e1c521d2-55b0-4386-9375-7c7b64d1610f-kube-api-access-kccmt\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.651997 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1c521d2-55b0-4386-9375-7c7b64d1610f-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "e1c521d2-55b0-4386-9375-7c7b64d1610f" (UID: "e1c521d2-55b0-4386-9375-7c7b64d1610f"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.672959 4955 scope.go:117] "RemoveContainer" containerID="1ecfa2946eb9cafc9cad6f71311771dcc302de5797b49660b697cd2042a0fff9" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.692129 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1c521d2-55b0-4386-9375-7c7b64d1610f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e1c521d2-55b0-4386-9375-7c7b64d1610f" (UID: "e1c521d2-55b0-4386-9375-7c7b64d1610f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.693572 4955 scope.go:117] "RemoveContainer" containerID="744b17f1d9b104c8caa4129128acc8dc79daba9c743ce68b87f254b512b4d903" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.695531 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1c521d2-55b0-4386-9375-7c7b64d1610f-config-data" (OuterVolumeSpecName: "config-data") pod "e1c521d2-55b0-4386-9375-7c7b64d1610f" (UID: "e1c521d2-55b0-4386-9375-7c7b64d1610f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.718018 4955 scope.go:117] "RemoveContainer" containerID="089fe3dff83ab164a4663da073016319402cb07ed7779e22c97e9d51bd600efa" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.743491 4955 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1c521d2-55b0-4386-9375-7c7b64d1610f-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.743522 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1c521d2-55b0-4386-9375-7c7b64d1610f-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.743534 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1c521d2-55b0-4386-9375-7c7b64d1610f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.791410 4955 scope.go:117] "RemoveContainer" containerID="61150ce92008ba120f051cc2960e790a0f85e45b49f99bc87e4d000b7f6f0af5" Feb 17 13:25:48 crc kubenswrapper[4955]: E0217 13:25:48.791957 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61150ce92008ba120f051cc2960e790a0f85e45b49f99bc87e4d000b7f6f0af5\": container with ID starting with 61150ce92008ba120f051cc2960e790a0f85e45b49f99bc87e4d000b7f6f0af5 not found: ID does not exist" containerID="61150ce92008ba120f051cc2960e790a0f85e45b49f99bc87e4d000b7f6f0af5" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.792205 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61150ce92008ba120f051cc2960e790a0f85e45b49f99bc87e4d000b7f6f0af5"} err="failed to get container status \"61150ce92008ba120f051cc2960e790a0f85e45b49f99bc87e4d000b7f6f0af5\": rpc error: code = NotFound desc = could not find container \"61150ce92008ba120f051cc2960e790a0f85e45b49f99bc87e4d000b7f6f0af5\": container with ID starting with 61150ce92008ba120f051cc2960e790a0f85e45b49f99bc87e4d000b7f6f0af5 not found: ID does not exist" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.792411 4955 scope.go:117] "RemoveContainer" containerID="1ecfa2946eb9cafc9cad6f71311771dcc302de5797b49660b697cd2042a0fff9" Feb 17 13:25:48 crc kubenswrapper[4955]: E0217 13:25:48.792911 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ecfa2946eb9cafc9cad6f71311771dcc302de5797b49660b697cd2042a0fff9\": container with ID starting with 1ecfa2946eb9cafc9cad6f71311771dcc302de5797b49660b697cd2042a0fff9 not found: ID does not exist" containerID="1ecfa2946eb9cafc9cad6f71311771dcc302de5797b49660b697cd2042a0fff9" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.792954 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ecfa2946eb9cafc9cad6f71311771dcc302de5797b49660b697cd2042a0fff9"} err="failed to get container status \"1ecfa2946eb9cafc9cad6f71311771dcc302de5797b49660b697cd2042a0fff9\": rpc error: code = NotFound desc = could not find container \"1ecfa2946eb9cafc9cad6f71311771dcc302de5797b49660b697cd2042a0fff9\": container with ID starting with 1ecfa2946eb9cafc9cad6f71311771dcc302de5797b49660b697cd2042a0fff9 not found: ID does not exist" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.792989 4955 scope.go:117] "RemoveContainer" containerID="744b17f1d9b104c8caa4129128acc8dc79daba9c743ce68b87f254b512b4d903" Feb 17 13:25:48 crc kubenswrapper[4955]: E0217 13:25:48.793380 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"744b17f1d9b104c8caa4129128acc8dc79daba9c743ce68b87f254b512b4d903\": container with ID starting with 744b17f1d9b104c8caa4129128acc8dc79daba9c743ce68b87f254b512b4d903 not found: ID does not exist" containerID="744b17f1d9b104c8caa4129128acc8dc79daba9c743ce68b87f254b512b4d903" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.793531 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"744b17f1d9b104c8caa4129128acc8dc79daba9c743ce68b87f254b512b4d903"} err="failed to get container status \"744b17f1d9b104c8caa4129128acc8dc79daba9c743ce68b87f254b512b4d903\": rpc error: code = NotFound desc = could not find container \"744b17f1d9b104c8caa4129128acc8dc79daba9c743ce68b87f254b512b4d903\": container with ID starting with 744b17f1d9b104c8caa4129128acc8dc79daba9c743ce68b87f254b512b4d903 not found: ID does not exist" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.793628 4955 scope.go:117] "RemoveContainer" containerID="089fe3dff83ab164a4663da073016319402cb07ed7779e22c97e9d51bd600efa" Feb 17 13:25:48 crc kubenswrapper[4955]: E0217 13:25:48.793996 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"089fe3dff83ab164a4663da073016319402cb07ed7779e22c97e9d51bd600efa\": container with ID starting with 089fe3dff83ab164a4663da073016319402cb07ed7779e22c97e9d51bd600efa not found: ID does not exist" containerID="089fe3dff83ab164a4663da073016319402cb07ed7779e22c97e9d51bd600efa" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.794039 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"089fe3dff83ab164a4663da073016319402cb07ed7779e22c97e9d51bd600efa"} err="failed to get container status \"089fe3dff83ab164a4663da073016319402cb07ed7779e22c97e9d51bd600efa\": rpc error: code = NotFound desc = could not find container \"089fe3dff83ab164a4663da073016319402cb07ed7779e22c97e9d51bd600efa\": container with ID starting with 089fe3dff83ab164a4663da073016319402cb07ed7779e22c97e9d51bd600efa not found: ID does not exist" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.893370 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.904609 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.929266 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:25:48 crc kubenswrapper[4955]: E0217 13:25:48.930508 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1c521d2-55b0-4386-9375-7c7b64d1610f" containerName="sg-core" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.930625 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1c521d2-55b0-4386-9375-7c7b64d1610f" containerName="sg-core" Feb 17 13:25:48 crc kubenswrapper[4955]: E0217 13:25:48.930692 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1c521d2-55b0-4386-9375-7c7b64d1610f" containerName="proxy-httpd" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.930742 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1c521d2-55b0-4386-9375-7c7b64d1610f" containerName="proxy-httpd" Feb 17 13:25:48 crc kubenswrapper[4955]: E0217 13:25:48.930834 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1c521d2-55b0-4386-9375-7c7b64d1610f" containerName="ceilometer-notification-agent" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.930935 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1c521d2-55b0-4386-9375-7c7b64d1610f" containerName="ceilometer-notification-agent" Feb 17 13:25:48 crc kubenswrapper[4955]: E0217 13:25:48.931003 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1c521d2-55b0-4386-9375-7c7b64d1610f" containerName="ceilometer-central-agent" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.931062 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1c521d2-55b0-4386-9375-7c7b64d1610f" containerName="ceilometer-central-agent" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.931323 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1c521d2-55b0-4386-9375-7c7b64d1610f" containerName="sg-core" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.931400 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1c521d2-55b0-4386-9375-7c7b64d1610f" containerName="ceilometer-central-agent" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.931465 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1c521d2-55b0-4386-9375-7c7b64d1610f" containerName="proxy-httpd" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.931574 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1c521d2-55b0-4386-9375-7c7b64d1610f" containerName="ceilometer-notification-agent" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.933163 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.947050 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.947057 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.947626 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.960103 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c46f44da-c85d-4a6c-a5db-28e900c054da-run-httpd\") pod \"ceilometer-0\" (UID: \"c46f44da-c85d-4a6c-a5db-28e900c054da\") " pod="openstack/ceilometer-0" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.960162 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c46f44da-c85d-4a6c-a5db-28e900c054da-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c46f44da-c85d-4a6c-a5db-28e900c054da\") " pod="openstack/ceilometer-0" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.960206 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c46f44da-c85d-4a6c-a5db-28e900c054da-log-httpd\") pod \"ceilometer-0\" (UID: \"c46f44da-c85d-4a6c-a5db-28e900c054da\") " pod="openstack/ceilometer-0" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.960255 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c46f44da-c85d-4a6c-a5db-28e900c054da-config-data\") pod \"ceilometer-0\" (UID: \"c46f44da-c85d-4a6c-a5db-28e900c054da\") " pod="openstack/ceilometer-0" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.960294 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c46f44da-c85d-4a6c-a5db-28e900c054da-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c46f44da-c85d-4a6c-a5db-28e900c054da\") " pod="openstack/ceilometer-0" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.960381 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6fvh\" (UniqueName: \"kubernetes.io/projected/c46f44da-c85d-4a6c-a5db-28e900c054da-kube-api-access-g6fvh\") pod \"ceilometer-0\" (UID: \"c46f44da-c85d-4a6c-a5db-28e900c054da\") " pod="openstack/ceilometer-0" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.960434 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c46f44da-c85d-4a6c-a5db-28e900c054da-scripts\") pod \"ceilometer-0\" (UID: \"c46f44da-c85d-4a6c-a5db-28e900c054da\") " pod="openstack/ceilometer-0" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.960459 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c46f44da-c85d-4a6c-a5db-28e900c054da-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c46f44da-c85d-4a6c-a5db-28e900c054da\") " pod="openstack/ceilometer-0" Feb 17 13:25:48 crc kubenswrapper[4955]: I0217 13:25:48.970140 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.062361 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c46f44da-c85d-4a6c-a5db-28e900c054da-scripts\") pod \"ceilometer-0\" (UID: \"c46f44da-c85d-4a6c-a5db-28e900c054da\") " pod="openstack/ceilometer-0" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.062692 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c46f44da-c85d-4a6c-a5db-28e900c054da-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c46f44da-c85d-4a6c-a5db-28e900c054da\") " pod="openstack/ceilometer-0" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.062763 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c46f44da-c85d-4a6c-a5db-28e900c054da-run-httpd\") pod \"ceilometer-0\" (UID: \"c46f44da-c85d-4a6c-a5db-28e900c054da\") " pod="openstack/ceilometer-0" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.062797 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c46f44da-c85d-4a6c-a5db-28e900c054da-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c46f44da-c85d-4a6c-a5db-28e900c054da\") " pod="openstack/ceilometer-0" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.062831 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c46f44da-c85d-4a6c-a5db-28e900c054da-log-httpd\") pod \"ceilometer-0\" (UID: \"c46f44da-c85d-4a6c-a5db-28e900c054da\") " pod="openstack/ceilometer-0" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.062878 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c46f44da-c85d-4a6c-a5db-28e900c054da-config-data\") pod \"ceilometer-0\" (UID: \"c46f44da-c85d-4a6c-a5db-28e900c054da\") " pod="openstack/ceilometer-0" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.062911 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c46f44da-c85d-4a6c-a5db-28e900c054da-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c46f44da-c85d-4a6c-a5db-28e900c054da\") " pod="openstack/ceilometer-0" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.062929 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6fvh\" (UniqueName: \"kubernetes.io/projected/c46f44da-c85d-4a6c-a5db-28e900c054da-kube-api-access-g6fvh\") pod \"ceilometer-0\" (UID: \"c46f44da-c85d-4a6c-a5db-28e900c054da\") " pod="openstack/ceilometer-0" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.063590 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c46f44da-c85d-4a6c-a5db-28e900c054da-log-httpd\") pod \"ceilometer-0\" (UID: \"c46f44da-c85d-4a6c-a5db-28e900c054da\") " pod="openstack/ceilometer-0" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.063696 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c46f44da-c85d-4a6c-a5db-28e900c054da-run-httpd\") pod \"ceilometer-0\" (UID: \"c46f44da-c85d-4a6c-a5db-28e900c054da\") " pod="openstack/ceilometer-0" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.068338 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c46f44da-c85d-4a6c-a5db-28e900c054da-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c46f44da-c85d-4a6c-a5db-28e900c054da\") " pod="openstack/ceilometer-0" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.068660 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c46f44da-c85d-4a6c-a5db-28e900c054da-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c46f44da-c85d-4a6c-a5db-28e900c054da\") " pod="openstack/ceilometer-0" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.068863 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c46f44da-c85d-4a6c-a5db-28e900c054da-config-data\") pod \"ceilometer-0\" (UID: \"c46f44da-c85d-4a6c-a5db-28e900c054da\") " pod="openstack/ceilometer-0" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.071308 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c46f44da-c85d-4a6c-a5db-28e900c054da-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c46f44da-c85d-4a6c-a5db-28e900c054da\") " pod="openstack/ceilometer-0" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.071341 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c46f44da-c85d-4a6c-a5db-28e900c054da-scripts\") pod \"ceilometer-0\" (UID: \"c46f44da-c85d-4a6c-a5db-28e900c054da\") " pod="openstack/ceilometer-0" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.083837 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6fvh\" (UniqueName: \"kubernetes.io/projected/c46f44da-c85d-4a6c-a5db-28e900c054da-kube-api-access-g6fvh\") pod \"ceilometer-0\" (UID: \"c46f44da-c85d-4a6c-a5db-28e900c054da\") " pod="openstack/ceilometer-0" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.201497 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.268525 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adfa8017-1c4c-4b31-90a8-c7e3ac999dd0-logs\") pod \"adfa8017-1c4c-4b31-90a8-c7e3ac999dd0\" (UID: \"adfa8017-1c4c-4b31-90a8-c7e3ac999dd0\") " Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.268665 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vjwmk\" (UniqueName: \"kubernetes.io/projected/adfa8017-1c4c-4b31-90a8-c7e3ac999dd0-kube-api-access-vjwmk\") pod \"adfa8017-1c4c-4b31-90a8-c7e3ac999dd0\" (UID: \"adfa8017-1c4c-4b31-90a8-c7e3ac999dd0\") " Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.268767 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adfa8017-1c4c-4b31-90a8-c7e3ac999dd0-config-data\") pod \"adfa8017-1c4c-4b31-90a8-c7e3ac999dd0\" (UID: \"adfa8017-1c4c-4b31-90a8-c7e3ac999dd0\") " Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.268839 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adfa8017-1c4c-4b31-90a8-c7e3ac999dd0-combined-ca-bundle\") pod \"adfa8017-1c4c-4b31-90a8-c7e3ac999dd0\" (UID: \"adfa8017-1c4c-4b31-90a8-c7e3ac999dd0\") " Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.269134 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adfa8017-1c4c-4b31-90a8-c7e3ac999dd0-logs" (OuterVolumeSpecName: "logs") pod "adfa8017-1c4c-4b31-90a8-c7e3ac999dd0" (UID: "adfa8017-1c4c-4b31-90a8-c7e3ac999dd0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.269677 4955 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adfa8017-1c4c-4b31-90a8-c7e3ac999dd0-logs\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.276710 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.307797 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adfa8017-1c4c-4b31-90a8-c7e3ac999dd0-kube-api-access-vjwmk" (OuterVolumeSpecName: "kube-api-access-vjwmk") pod "adfa8017-1c4c-4b31-90a8-c7e3ac999dd0" (UID: "adfa8017-1c4c-4b31-90a8-c7e3ac999dd0"). InnerVolumeSpecName "kube-api-access-vjwmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.370957 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vjwmk\" (UniqueName: \"kubernetes.io/projected/adfa8017-1c4c-4b31-90a8-c7e3ac999dd0-kube-api-access-vjwmk\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.374973 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adfa8017-1c4c-4b31-90a8-c7e3ac999dd0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "adfa8017-1c4c-4b31-90a8-c7e3ac999dd0" (UID: "adfa8017-1c4c-4b31-90a8-c7e3ac999dd0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.407660 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adfa8017-1c4c-4b31-90a8-c7e3ac999dd0-config-data" (OuterVolumeSpecName: "config-data") pod "adfa8017-1c4c-4b31-90a8-c7e3ac999dd0" (UID: "adfa8017-1c4c-4b31-90a8-c7e3ac999dd0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.475108 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adfa8017-1c4c-4b31-90a8-c7e3ac999dd0-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.475343 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adfa8017-1c4c-4b31-90a8-c7e3ac999dd0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.567452 4955 generic.go:334] "Generic (PLEG): container finished" podID="adfa8017-1c4c-4b31-90a8-c7e3ac999dd0" containerID="767b94202f66a98ef0dea13f1984f9a22753af63d8eb02d11ba3bb1a9b0cbc31" exitCode=0 Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.567502 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"adfa8017-1c4c-4b31-90a8-c7e3ac999dd0","Type":"ContainerDied","Data":"767b94202f66a98ef0dea13f1984f9a22753af63d8eb02d11ba3bb1a9b0cbc31"} Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.567603 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"adfa8017-1c4c-4b31-90a8-c7e3ac999dd0","Type":"ContainerDied","Data":"418ba7456744415a6978fdbfc9b893a15d923d2c3a177bfc0e2a5deea64f36ea"} Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.567633 4955 scope.go:117] "RemoveContainer" containerID="767b94202f66a98ef0dea13f1984f9a22753af63d8eb02d11ba3bb1a9b0cbc31" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.567551 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.639289 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.647340 4955 scope.go:117] "RemoveContainer" containerID="4627d2789659ddf3ab32efe87cc56e52503a33e443792449cb1ee6abb670881c" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.657226 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.676691 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 17 13:25:49 crc kubenswrapper[4955]: E0217 13:25:49.677257 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adfa8017-1c4c-4b31-90a8-c7e3ac999dd0" containerName="nova-api-log" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.677271 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="adfa8017-1c4c-4b31-90a8-c7e3ac999dd0" containerName="nova-api-log" Feb 17 13:25:49 crc kubenswrapper[4955]: E0217 13:25:49.677297 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adfa8017-1c4c-4b31-90a8-c7e3ac999dd0" containerName="nova-api-api" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.677305 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="adfa8017-1c4c-4b31-90a8-c7e3ac999dd0" containerName="nova-api-api" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.677470 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="adfa8017-1c4c-4b31-90a8-c7e3ac999dd0" containerName="nova-api-log" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.677494 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="adfa8017-1c4c-4b31-90a8-c7e3ac999dd0" containerName="nova-api-api" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.678488 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.687146 4955 scope.go:117] "RemoveContainer" containerID="767b94202f66a98ef0dea13f1984f9a22753af63d8eb02d11ba3bb1a9b0cbc31" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.687403 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.687523 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.687691 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Feb 17 13:25:49 crc kubenswrapper[4955]: E0217 13:25:49.688764 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"767b94202f66a98ef0dea13f1984f9a22753af63d8eb02d11ba3bb1a9b0cbc31\": container with ID starting with 767b94202f66a98ef0dea13f1984f9a22753af63d8eb02d11ba3bb1a9b0cbc31 not found: ID does not exist" containerID="767b94202f66a98ef0dea13f1984f9a22753af63d8eb02d11ba3bb1a9b0cbc31" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.688873 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"767b94202f66a98ef0dea13f1984f9a22753af63d8eb02d11ba3bb1a9b0cbc31"} err="failed to get container status \"767b94202f66a98ef0dea13f1984f9a22753af63d8eb02d11ba3bb1a9b0cbc31\": rpc error: code = NotFound desc = could not find container \"767b94202f66a98ef0dea13f1984f9a22753af63d8eb02d11ba3bb1a9b0cbc31\": container with ID starting with 767b94202f66a98ef0dea13f1984f9a22753af63d8eb02d11ba3bb1a9b0cbc31 not found: ID does not exist" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.688904 4955 scope.go:117] "RemoveContainer" containerID="4627d2789659ddf3ab32efe87cc56e52503a33e443792449cb1ee6abb670881c" Feb 17 13:25:49 crc kubenswrapper[4955]: E0217 13:25:49.689222 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4627d2789659ddf3ab32efe87cc56e52503a33e443792449cb1ee6abb670881c\": container with ID starting with 4627d2789659ddf3ab32efe87cc56e52503a33e443792449cb1ee6abb670881c not found: ID does not exist" containerID="4627d2789659ddf3ab32efe87cc56e52503a33e443792449cb1ee6abb670881c" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.689256 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4627d2789659ddf3ab32efe87cc56e52503a33e443792449cb1ee6abb670881c"} err="failed to get container status \"4627d2789659ddf3ab32efe87cc56e52503a33e443792449cb1ee6abb670881c\": rpc error: code = NotFound desc = could not find container \"4627d2789659ddf3ab32efe87cc56e52503a33e443792449cb1ee6abb670881c\": container with ID starting with 4627d2789659ddf3ab32efe87cc56e52503a33e443792449cb1ee6abb670881c not found: ID does not exist" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.689768 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.781772 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cb6b03f-a3f9-4a85-9db3-c504f959634e-public-tls-certs\") pod \"nova-api-0\" (UID: \"6cb6b03f-a3f9-4a85-9db3-c504f959634e\") " pod="openstack/nova-api-0" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.781940 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2pfl\" (UniqueName: \"kubernetes.io/projected/6cb6b03f-a3f9-4a85-9db3-c504f959634e-kube-api-access-q2pfl\") pod \"nova-api-0\" (UID: \"6cb6b03f-a3f9-4a85-9db3-c504f959634e\") " pod="openstack/nova-api-0" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.782040 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6cb6b03f-a3f9-4a85-9db3-c504f959634e-logs\") pod \"nova-api-0\" (UID: \"6cb6b03f-a3f9-4a85-9db3-c504f959634e\") " pod="openstack/nova-api-0" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.782102 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cb6b03f-a3f9-4a85-9db3-c504f959634e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6cb6b03f-a3f9-4a85-9db3-c504f959634e\") " pod="openstack/nova-api-0" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.782139 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cb6b03f-a3f9-4a85-9db3-c504f959634e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6cb6b03f-a3f9-4a85-9db3-c504f959634e\") " pod="openstack/nova-api-0" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.782450 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cb6b03f-a3f9-4a85-9db3-c504f959634e-config-data\") pod \"nova-api-0\" (UID: \"6cb6b03f-a3f9-4a85-9db3-c504f959634e\") " pod="openstack/nova-api-0" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.834854 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.837478 4955 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.884520 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cb6b03f-a3f9-4a85-9db3-c504f959634e-config-data\") pod \"nova-api-0\" (UID: \"6cb6b03f-a3f9-4a85-9db3-c504f959634e\") " pod="openstack/nova-api-0" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.884580 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cb6b03f-a3f9-4a85-9db3-c504f959634e-public-tls-certs\") pod \"nova-api-0\" (UID: \"6cb6b03f-a3f9-4a85-9db3-c504f959634e\") " pod="openstack/nova-api-0" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.884624 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2pfl\" (UniqueName: \"kubernetes.io/projected/6cb6b03f-a3f9-4a85-9db3-c504f959634e-kube-api-access-q2pfl\") pod \"nova-api-0\" (UID: \"6cb6b03f-a3f9-4a85-9db3-c504f959634e\") " pod="openstack/nova-api-0" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.884656 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6cb6b03f-a3f9-4a85-9db3-c504f959634e-logs\") pod \"nova-api-0\" (UID: \"6cb6b03f-a3f9-4a85-9db3-c504f959634e\") " pod="openstack/nova-api-0" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.884990 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cb6b03f-a3f9-4a85-9db3-c504f959634e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6cb6b03f-a3f9-4a85-9db3-c504f959634e\") " pod="openstack/nova-api-0" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.885024 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cb6b03f-a3f9-4a85-9db3-c504f959634e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6cb6b03f-a3f9-4a85-9db3-c504f959634e\") " pod="openstack/nova-api-0" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.885236 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6cb6b03f-a3f9-4a85-9db3-c504f959634e-logs\") pod \"nova-api-0\" (UID: \"6cb6b03f-a3f9-4a85-9db3-c504f959634e\") " pod="openstack/nova-api-0" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.889668 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cb6b03f-a3f9-4a85-9db3-c504f959634e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6cb6b03f-a3f9-4a85-9db3-c504f959634e\") " pod="openstack/nova-api-0" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.892131 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cb6b03f-a3f9-4a85-9db3-c504f959634e-config-data\") pod \"nova-api-0\" (UID: \"6cb6b03f-a3f9-4a85-9db3-c504f959634e\") " pod="openstack/nova-api-0" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.892856 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cb6b03f-a3f9-4a85-9db3-c504f959634e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6cb6b03f-a3f9-4a85-9db3-c504f959634e\") " pod="openstack/nova-api-0" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.898079 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cb6b03f-a3f9-4a85-9db3-c504f959634e-public-tls-certs\") pod \"nova-api-0\" (UID: \"6cb6b03f-a3f9-4a85-9db3-c504f959634e\") " pod="openstack/nova-api-0" Feb 17 13:25:49 crc kubenswrapper[4955]: I0217 13:25:49.902191 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2pfl\" (UniqueName: \"kubernetes.io/projected/6cb6b03f-a3f9-4a85-9db3-c504f959634e-kube-api-access-q2pfl\") pod \"nova-api-0\" (UID: \"6cb6b03f-a3f9-4a85-9db3-c504f959634e\") " pod="openstack/nova-api-0" Feb 17 13:25:50 crc kubenswrapper[4955]: I0217 13:25:50.006262 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 17 13:25:50 crc kubenswrapper[4955]: I0217 13:25:50.014117 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 17 13:25:50 crc kubenswrapper[4955]: I0217 13:25:50.014162 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 17 13:25:50 crc kubenswrapper[4955]: I0217 13:25:50.046964 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Feb 17 13:25:50 crc kubenswrapper[4955]: I0217 13:25:50.079905 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Feb 17 13:25:50 crc kubenswrapper[4955]: I0217 13:25:50.237615 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adfa8017-1c4c-4b31-90a8-c7e3ac999dd0" path="/var/lib/kubelet/pods/adfa8017-1c4c-4b31-90a8-c7e3ac999dd0/volumes" Feb 17 13:25:50 crc kubenswrapper[4955]: I0217 13:25:50.246556 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1c521d2-55b0-4386-9375-7c7b64d1610f" path="/var/lib/kubelet/pods/e1c521d2-55b0-4386-9375-7c7b64d1610f/volumes" Feb 17 13:25:50 crc kubenswrapper[4955]: I0217 13:25:50.505408 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 17 13:25:50 crc kubenswrapper[4955]: I0217 13:25:50.582902 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6cb6b03f-a3f9-4a85-9db3-c504f959634e","Type":"ContainerStarted","Data":"4e1768eb4dfc8221cd30f19e3115718e6122d70e484cb3f5a3aa7e8547a3265c"} Feb 17 13:25:50 crc kubenswrapper[4955]: I0217 13:25:50.589983 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c46f44da-c85d-4a6c-a5db-28e900c054da","Type":"ContainerStarted","Data":"beb45803ba349f876b1d2225b3d476d0fb293c061f192088f994cef10c74e899"} Feb 17 13:25:50 crc kubenswrapper[4955]: I0217 13:25:50.633378 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Feb 17 13:25:50 crc kubenswrapper[4955]: I0217 13:25:50.892611 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-4g28c"] Feb 17 13:25:50 crc kubenswrapper[4955]: I0217 13:25:50.894090 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-4g28c" Feb 17 13:25:50 crc kubenswrapper[4955]: I0217 13:25:50.896772 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Feb 17 13:25:50 crc kubenswrapper[4955]: I0217 13:25:50.896942 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Feb 17 13:25:50 crc kubenswrapper[4955]: I0217 13:25:50.902007 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-4g28c"] Feb 17 13:25:50 crc kubenswrapper[4955]: I0217 13:25:50.906102 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnrps\" (UniqueName: \"kubernetes.io/projected/f3a87515-64e3-4ed0-bc43-39ea2caa11c0-kube-api-access-cnrps\") pod \"nova-cell1-cell-mapping-4g28c\" (UID: \"f3a87515-64e3-4ed0-bc43-39ea2caa11c0\") " pod="openstack/nova-cell1-cell-mapping-4g28c" Feb 17 13:25:50 crc kubenswrapper[4955]: I0217 13:25:50.906162 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3a87515-64e3-4ed0-bc43-39ea2caa11c0-config-data\") pod \"nova-cell1-cell-mapping-4g28c\" (UID: \"f3a87515-64e3-4ed0-bc43-39ea2caa11c0\") " pod="openstack/nova-cell1-cell-mapping-4g28c" Feb 17 13:25:50 crc kubenswrapper[4955]: I0217 13:25:50.906202 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3a87515-64e3-4ed0-bc43-39ea2caa11c0-scripts\") pod \"nova-cell1-cell-mapping-4g28c\" (UID: \"f3a87515-64e3-4ed0-bc43-39ea2caa11c0\") " pod="openstack/nova-cell1-cell-mapping-4g28c" Feb 17 13:25:50 crc kubenswrapper[4955]: I0217 13:25:50.906248 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3a87515-64e3-4ed0-bc43-39ea2caa11c0-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-4g28c\" (UID: \"f3a87515-64e3-4ed0-bc43-39ea2caa11c0\") " pod="openstack/nova-cell1-cell-mapping-4g28c" Feb 17 13:25:51 crc kubenswrapper[4955]: I0217 13:25:51.008421 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3a87515-64e3-4ed0-bc43-39ea2caa11c0-scripts\") pod \"nova-cell1-cell-mapping-4g28c\" (UID: \"f3a87515-64e3-4ed0-bc43-39ea2caa11c0\") " pod="openstack/nova-cell1-cell-mapping-4g28c" Feb 17 13:25:51 crc kubenswrapper[4955]: I0217 13:25:51.008545 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3a87515-64e3-4ed0-bc43-39ea2caa11c0-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-4g28c\" (UID: \"f3a87515-64e3-4ed0-bc43-39ea2caa11c0\") " pod="openstack/nova-cell1-cell-mapping-4g28c" Feb 17 13:25:51 crc kubenswrapper[4955]: I0217 13:25:51.008648 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnrps\" (UniqueName: \"kubernetes.io/projected/f3a87515-64e3-4ed0-bc43-39ea2caa11c0-kube-api-access-cnrps\") pod \"nova-cell1-cell-mapping-4g28c\" (UID: \"f3a87515-64e3-4ed0-bc43-39ea2caa11c0\") " pod="openstack/nova-cell1-cell-mapping-4g28c" Feb 17 13:25:51 crc kubenswrapper[4955]: I0217 13:25:51.008711 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3a87515-64e3-4ed0-bc43-39ea2caa11c0-config-data\") pod \"nova-cell1-cell-mapping-4g28c\" (UID: \"f3a87515-64e3-4ed0-bc43-39ea2caa11c0\") " pod="openstack/nova-cell1-cell-mapping-4g28c" Feb 17 13:25:51 crc kubenswrapper[4955]: I0217 13:25:51.013439 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3a87515-64e3-4ed0-bc43-39ea2caa11c0-config-data\") pod \"nova-cell1-cell-mapping-4g28c\" (UID: \"f3a87515-64e3-4ed0-bc43-39ea2caa11c0\") " pod="openstack/nova-cell1-cell-mapping-4g28c" Feb 17 13:25:51 crc kubenswrapper[4955]: I0217 13:25:51.013464 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3a87515-64e3-4ed0-bc43-39ea2caa11c0-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-4g28c\" (UID: \"f3a87515-64e3-4ed0-bc43-39ea2caa11c0\") " pod="openstack/nova-cell1-cell-mapping-4g28c" Feb 17 13:25:51 crc kubenswrapper[4955]: I0217 13:25:51.014372 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3a87515-64e3-4ed0-bc43-39ea2caa11c0-scripts\") pod \"nova-cell1-cell-mapping-4g28c\" (UID: \"f3a87515-64e3-4ed0-bc43-39ea2caa11c0\") " pod="openstack/nova-cell1-cell-mapping-4g28c" Feb 17 13:25:51 crc kubenswrapper[4955]: I0217 13:25:51.034298 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnrps\" (UniqueName: \"kubernetes.io/projected/f3a87515-64e3-4ed0-bc43-39ea2caa11c0-kube-api-access-cnrps\") pod \"nova-cell1-cell-mapping-4g28c\" (UID: \"f3a87515-64e3-4ed0-bc43-39ea2caa11c0\") " pod="openstack/nova-cell1-cell-mapping-4g28c" Feb 17 13:25:51 crc kubenswrapper[4955]: I0217 13:25:51.066051 4955 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="3fe1264c-d2ce-455a-99b9-88270f4dfb04" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 17 13:25:51 crc kubenswrapper[4955]: I0217 13:25:51.066163 4955 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="3fe1264c-d2ce-455a-99b9-88270f4dfb04" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 17 13:25:51 crc kubenswrapper[4955]: I0217 13:25:51.210381 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-4g28c" Feb 17 13:25:51 crc kubenswrapper[4955]: I0217 13:25:51.614810 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6cb6b03f-a3f9-4a85-9db3-c504f959634e","Type":"ContainerStarted","Data":"9456cf3d2171a35a016c053f0b796677c21a625fcc03cc85a69903a46e9f1888"} Feb 17 13:25:51 crc kubenswrapper[4955]: I0217 13:25:51.615335 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6cb6b03f-a3f9-4a85-9db3-c504f959634e","Type":"ContainerStarted","Data":"dd97550f6e291125acb0ee9d6a3bb7bea3996a670b71e9f34904446a033c0d48"} Feb 17 13:25:51 crc kubenswrapper[4955]: I0217 13:25:51.618416 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c46f44da-c85d-4a6c-a5db-28e900c054da","Type":"ContainerStarted","Data":"111b21ad1db00f438bf335d152b075994dda1c79a730df7b9860a7b7b6a8ccdf"} Feb 17 13:25:51 crc kubenswrapper[4955]: I0217 13:25:51.618452 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c46f44da-c85d-4a6c-a5db-28e900c054da","Type":"ContainerStarted","Data":"ae4ea238aa4471975632720842d6850c6bca317059ffd564b8d04a83dbedaa2d"} Feb 17 13:25:51 crc kubenswrapper[4955]: I0217 13:25:51.630549 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.630528519 podStartE2EDuration="2.630528519s" podCreationTimestamp="2026-02-17 13:25:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:25:51.629442128 +0000 UTC m=+1290.152171691" watchObservedRunningTime="2026-02-17 13:25:51.630528519 +0000 UTC m=+1290.153258062" Feb 17 13:25:51 crc kubenswrapper[4955]: I0217 13:25:51.746708 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-4g28c"] Feb 17 13:25:51 crc kubenswrapper[4955]: W0217 13:25:51.749319 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf3a87515_64e3_4ed0_bc43_39ea2caa11c0.slice/crio-1bd5102c0d2874e12dd41bf4c88a13b50ea2d30bb954f62efd6121339cd3d9b5 WatchSource:0}: Error finding container 1bd5102c0d2874e12dd41bf4c88a13b50ea2d30bb954f62efd6121339cd3d9b5: Status 404 returned error can't find the container with id 1bd5102c0d2874e12dd41bf4c88a13b50ea2d30bb954f62efd6121339cd3d9b5 Feb 17 13:25:52 crc kubenswrapper[4955]: I0217 13:25:52.637401 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-4g28c" event={"ID":"f3a87515-64e3-4ed0-bc43-39ea2caa11c0","Type":"ContainerStarted","Data":"bdbc3d9379e7a7d7f836ef2ed615cfbcc7a7d7f6708259816dc806b94dd3b3cd"} Feb 17 13:25:52 crc kubenswrapper[4955]: I0217 13:25:52.638066 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-4g28c" event={"ID":"f3a87515-64e3-4ed0-bc43-39ea2caa11c0","Type":"ContainerStarted","Data":"1bd5102c0d2874e12dd41bf4c88a13b50ea2d30bb954f62efd6121339cd3d9b5"} Feb 17 13:25:52 crc kubenswrapper[4955]: I0217 13:25:52.661436 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-4g28c" podStartSLOduration=2.6614181329999997 podStartE2EDuration="2.661418133s" podCreationTimestamp="2026-02-17 13:25:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:25:52.65223427 +0000 UTC m=+1291.174963813" watchObservedRunningTime="2026-02-17 13:25:52.661418133 +0000 UTC m=+1291.184147666" Feb 17 13:25:53 crc kubenswrapper[4955]: I0217 13:25:53.002900 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cd5cbd7b9-9fmsq" Feb 17 13:25:53 crc kubenswrapper[4955]: I0217 13:25:53.062532 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-b6cbh"] Feb 17 13:25:53 crc kubenswrapper[4955]: I0217 13:25:53.062840 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-bccf8f775-b6cbh" podUID="987ec26b-c18e-47ec-be15-24059d88e961" containerName="dnsmasq-dns" containerID="cri-o://4513616fbc9ee9a8b58471bb35d22e68a4a066e423e406eada0ad59580bc24f2" gracePeriod=10 Feb 17 13:25:53 crc kubenswrapper[4955]: I0217 13:25:53.592955 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-b6cbh" Feb 17 13:25:53 crc kubenswrapper[4955]: I0217 13:25:53.653273 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c46f44da-c85d-4a6c-a5db-28e900c054da","Type":"ContainerStarted","Data":"23c0b54b8032acd4f2b1092a19e76ee3213d6d190c5ffbb249dc745aadfc9b78"} Feb 17 13:25:53 crc kubenswrapper[4955]: I0217 13:25:53.656375 4955 generic.go:334] "Generic (PLEG): container finished" podID="987ec26b-c18e-47ec-be15-24059d88e961" containerID="4513616fbc9ee9a8b58471bb35d22e68a4a066e423e406eada0ad59580bc24f2" exitCode=0 Feb 17 13:25:53 crc kubenswrapper[4955]: I0217 13:25:53.656459 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-b6cbh" Feb 17 13:25:53 crc kubenswrapper[4955]: I0217 13:25:53.656443 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-b6cbh" event={"ID":"987ec26b-c18e-47ec-be15-24059d88e961","Type":"ContainerDied","Data":"4513616fbc9ee9a8b58471bb35d22e68a4a066e423e406eada0ad59580bc24f2"} Feb 17 13:25:53 crc kubenswrapper[4955]: I0217 13:25:53.656522 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-b6cbh" event={"ID":"987ec26b-c18e-47ec-be15-24059d88e961","Type":"ContainerDied","Data":"0fe99668d6bc86a35848edfc3878bb6e957baf6ddd91c7e5be544e3d77f889bd"} Feb 17 13:25:53 crc kubenswrapper[4955]: I0217 13:25:53.656549 4955 scope.go:117] "RemoveContainer" containerID="4513616fbc9ee9a8b58471bb35d22e68a4a066e423e406eada0ad59580bc24f2" Feb 17 13:25:53 crc kubenswrapper[4955]: I0217 13:25:53.657945 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/987ec26b-c18e-47ec-be15-24059d88e961-config\") pod \"987ec26b-c18e-47ec-be15-24059d88e961\" (UID: \"987ec26b-c18e-47ec-be15-24059d88e961\") " Feb 17 13:25:53 crc kubenswrapper[4955]: I0217 13:25:53.657993 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/987ec26b-c18e-47ec-be15-24059d88e961-ovsdbserver-nb\") pod \"987ec26b-c18e-47ec-be15-24059d88e961\" (UID: \"987ec26b-c18e-47ec-be15-24059d88e961\") " Feb 17 13:25:53 crc kubenswrapper[4955]: I0217 13:25:53.658045 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnh22\" (UniqueName: \"kubernetes.io/projected/987ec26b-c18e-47ec-be15-24059d88e961-kube-api-access-qnh22\") pod \"987ec26b-c18e-47ec-be15-24059d88e961\" (UID: \"987ec26b-c18e-47ec-be15-24059d88e961\") " Feb 17 13:25:53 crc kubenswrapper[4955]: I0217 13:25:53.658194 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/987ec26b-c18e-47ec-be15-24059d88e961-dns-swift-storage-0\") pod \"987ec26b-c18e-47ec-be15-24059d88e961\" (UID: \"987ec26b-c18e-47ec-be15-24059d88e961\") " Feb 17 13:25:53 crc kubenswrapper[4955]: I0217 13:25:53.658278 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/987ec26b-c18e-47ec-be15-24059d88e961-ovsdbserver-sb\") pod \"987ec26b-c18e-47ec-be15-24059d88e961\" (UID: \"987ec26b-c18e-47ec-be15-24059d88e961\") " Feb 17 13:25:53 crc kubenswrapper[4955]: I0217 13:25:53.658403 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/987ec26b-c18e-47ec-be15-24059d88e961-dns-svc\") pod \"987ec26b-c18e-47ec-be15-24059d88e961\" (UID: \"987ec26b-c18e-47ec-be15-24059d88e961\") " Feb 17 13:25:53 crc kubenswrapper[4955]: I0217 13:25:53.688325 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/987ec26b-c18e-47ec-be15-24059d88e961-kube-api-access-qnh22" (OuterVolumeSpecName: "kube-api-access-qnh22") pod "987ec26b-c18e-47ec-be15-24059d88e961" (UID: "987ec26b-c18e-47ec-be15-24059d88e961"). InnerVolumeSpecName "kube-api-access-qnh22". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:25:53 crc kubenswrapper[4955]: I0217 13:25:53.721731 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/987ec26b-c18e-47ec-be15-24059d88e961-config" (OuterVolumeSpecName: "config") pod "987ec26b-c18e-47ec-be15-24059d88e961" (UID: "987ec26b-c18e-47ec-be15-24059d88e961"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:25:53 crc kubenswrapper[4955]: I0217 13:25:53.737525 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/987ec26b-c18e-47ec-be15-24059d88e961-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "987ec26b-c18e-47ec-be15-24059d88e961" (UID: "987ec26b-c18e-47ec-be15-24059d88e961"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:25:53 crc kubenswrapper[4955]: I0217 13:25:53.754137 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/987ec26b-c18e-47ec-be15-24059d88e961-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "987ec26b-c18e-47ec-be15-24059d88e961" (UID: "987ec26b-c18e-47ec-be15-24059d88e961"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:25:53 crc kubenswrapper[4955]: I0217 13:25:53.760647 4955 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/987ec26b-c18e-47ec-be15-24059d88e961-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:53 crc kubenswrapper[4955]: I0217 13:25:53.760682 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnh22\" (UniqueName: \"kubernetes.io/projected/987ec26b-c18e-47ec-be15-24059d88e961-kube-api-access-qnh22\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:53 crc kubenswrapper[4955]: I0217 13:25:53.760698 4955 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/987ec26b-c18e-47ec-be15-24059d88e961-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:53 crc kubenswrapper[4955]: I0217 13:25:53.760710 4955 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/987ec26b-c18e-47ec-be15-24059d88e961-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:53 crc kubenswrapper[4955]: I0217 13:25:53.781843 4955 scope.go:117] "RemoveContainer" containerID="d0e0c44eab047bf10331abf17b10ea10e7863a066569ee44f2da418f0d8edb45" Feb 17 13:25:53 crc kubenswrapper[4955]: I0217 13:25:53.787677 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/987ec26b-c18e-47ec-be15-24059d88e961-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "987ec26b-c18e-47ec-be15-24059d88e961" (UID: "987ec26b-c18e-47ec-be15-24059d88e961"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:25:53 crc kubenswrapper[4955]: I0217 13:25:53.798230 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/987ec26b-c18e-47ec-be15-24059d88e961-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "987ec26b-c18e-47ec-be15-24059d88e961" (UID: "987ec26b-c18e-47ec-be15-24059d88e961"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:25:53 crc kubenswrapper[4955]: I0217 13:25:53.818475 4955 scope.go:117] "RemoveContainer" containerID="4513616fbc9ee9a8b58471bb35d22e68a4a066e423e406eada0ad59580bc24f2" Feb 17 13:25:53 crc kubenswrapper[4955]: E0217 13:25:53.818965 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4513616fbc9ee9a8b58471bb35d22e68a4a066e423e406eada0ad59580bc24f2\": container with ID starting with 4513616fbc9ee9a8b58471bb35d22e68a4a066e423e406eada0ad59580bc24f2 not found: ID does not exist" containerID="4513616fbc9ee9a8b58471bb35d22e68a4a066e423e406eada0ad59580bc24f2" Feb 17 13:25:53 crc kubenswrapper[4955]: I0217 13:25:53.819018 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4513616fbc9ee9a8b58471bb35d22e68a4a066e423e406eada0ad59580bc24f2"} err="failed to get container status \"4513616fbc9ee9a8b58471bb35d22e68a4a066e423e406eada0ad59580bc24f2\": rpc error: code = NotFound desc = could not find container \"4513616fbc9ee9a8b58471bb35d22e68a4a066e423e406eada0ad59580bc24f2\": container with ID starting with 4513616fbc9ee9a8b58471bb35d22e68a4a066e423e406eada0ad59580bc24f2 not found: ID does not exist" Feb 17 13:25:53 crc kubenswrapper[4955]: I0217 13:25:53.819045 4955 scope.go:117] "RemoveContainer" containerID="d0e0c44eab047bf10331abf17b10ea10e7863a066569ee44f2da418f0d8edb45" Feb 17 13:25:53 crc kubenswrapper[4955]: E0217 13:25:53.819436 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0e0c44eab047bf10331abf17b10ea10e7863a066569ee44f2da418f0d8edb45\": container with ID starting with d0e0c44eab047bf10331abf17b10ea10e7863a066569ee44f2da418f0d8edb45 not found: ID does not exist" containerID="d0e0c44eab047bf10331abf17b10ea10e7863a066569ee44f2da418f0d8edb45" Feb 17 13:25:53 crc kubenswrapper[4955]: I0217 13:25:53.819487 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0e0c44eab047bf10331abf17b10ea10e7863a066569ee44f2da418f0d8edb45"} err="failed to get container status \"d0e0c44eab047bf10331abf17b10ea10e7863a066569ee44f2da418f0d8edb45\": rpc error: code = NotFound desc = could not find container \"d0e0c44eab047bf10331abf17b10ea10e7863a066569ee44f2da418f0d8edb45\": container with ID starting with d0e0c44eab047bf10331abf17b10ea10e7863a066569ee44f2da418f0d8edb45 not found: ID does not exist" Feb 17 13:25:53 crc kubenswrapper[4955]: I0217 13:25:53.868108 4955 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/987ec26b-c18e-47ec-be15-24059d88e961-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:53 crc kubenswrapper[4955]: I0217 13:25:53.868172 4955 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/987ec26b-c18e-47ec-be15-24059d88e961-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:54 crc kubenswrapper[4955]: I0217 13:25:54.003141 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-b6cbh"] Feb 17 13:25:54 crc kubenswrapper[4955]: I0217 13:25:54.011154 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-b6cbh"] Feb 17 13:25:54 crc kubenswrapper[4955]: I0217 13:25:54.233371 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="987ec26b-c18e-47ec-be15-24059d88e961" path="/var/lib/kubelet/pods/987ec26b-c18e-47ec-be15-24059d88e961/volumes" Feb 17 13:25:54 crc kubenswrapper[4955]: I0217 13:25:54.672958 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c46f44da-c85d-4a6c-a5db-28e900c054da","Type":"ContainerStarted","Data":"ff7b1ead6331c7ef2375f0e3c54a443375796f7615b4544ab1fdc0165f685e2f"} Feb 17 13:25:54 crc kubenswrapper[4955]: I0217 13:25:54.673285 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 17 13:25:54 crc kubenswrapper[4955]: I0217 13:25:54.697560 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.797774226 podStartE2EDuration="6.697541395s" podCreationTimestamp="2026-02-17 13:25:48 +0000 UTC" firstStartedPulling="2026-02-17 13:25:49.837250621 +0000 UTC m=+1288.359980164" lastFinishedPulling="2026-02-17 13:25:53.73701779 +0000 UTC m=+1292.259747333" observedRunningTime="2026-02-17 13:25:54.695357033 +0000 UTC m=+1293.218086586" watchObservedRunningTime="2026-02-17 13:25:54.697541395 +0000 UTC m=+1293.220270938" Feb 17 13:25:57 crc kubenswrapper[4955]: I0217 13:25:57.706507 4955 generic.go:334] "Generic (PLEG): container finished" podID="f3a87515-64e3-4ed0-bc43-39ea2caa11c0" containerID="bdbc3d9379e7a7d7f836ef2ed615cfbcc7a7d7f6708259816dc806b94dd3b3cd" exitCode=0 Feb 17 13:25:57 crc kubenswrapper[4955]: I0217 13:25:57.706578 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-4g28c" event={"ID":"f3a87515-64e3-4ed0-bc43-39ea2caa11c0","Type":"ContainerDied","Data":"bdbc3d9379e7a7d7f836ef2ed615cfbcc7a7d7f6708259816dc806b94dd3b3cd"} Feb 17 13:25:58 crc kubenswrapper[4955]: I0217 13:25:58.431410 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-bccf8f775-b6cbh" podUID="987ec26b-c18e-47ec-be15-24059d88e961" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.195:5353: i/o timeout" Feb 17 13:25:59 crc kubenswrapper[4955]: I0217 13:25:59.065001 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-4g28c" Feb 17 13:25:59 crc kubenswrapper[4955]: I0217 13:25:59.162699 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3a87515-64e3-4ed0-bc43-39ea2caa11c0-config-data\") pod \"f3a87515-64e3-4ed0-bc43-39ea2caa11c0\" (UID: \"f3a87515-64e3-4ed0-bc43-39ea2caa11c0\") " Feb 17 13:25:59 crc kubenswrapper[4955]: I0217 13:25:59.162844 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3a87515-64e3-4ed0-bc43-39ea2caa11c0-combined-ca-bundle\") pod \"f3a87515-64e3-4ed0-bc43-39ea2caa11c0\" (UID: \"f3a87515-64e3-4ed0-bc43-39ea2caa11c0\") " Feb 17 13:25:59 crc kubenswrapper[4955]: I0217 13:25:59.162924 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cnrps\" (UniqueName: \"kubernetes.io/projected/f3a87515-64e3-4ed0-bc43-39ea2caa11c0-kube-api-access-cnrps\") pod \"f3a87515-64e3-4ed0-bc43-39ea2caa11c0\" (UID: \"f3a87515-64e3-4ed0-bc43-39ea2caa11c0\") " Feb 17 13:25:59 crc kubenswrapper[4955]: I0217 13:25:59.162977 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3a87515-64e3-4ed0-bc43-39ea2caa11c0-scripts\") pod \"f3a87515-64e3-4ed0-bc43-39ea2caa11c0\" (UID: \"f3a87515-64e3-4ed0-bc43-39ea2caa11c0\") " Feb 17 13:25:59 crc kubenswrapper[4955]: I0217 13:25:59.170847 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3a87515-64e3-4ed0-bc43-39ea2caa11c0-kube-api-access-cnrps" (OuterVolumeSpecName: "kube-api-access-cnrps") pod "f3a87515-64e3-4ed0-bc43-39ea2caa11c0" (UID: "f3a87515-64e3-4ed0-bc43-39ea2caa11c0"). InnerVolumeSpecName "kube-api-access-cnrps". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:25:59 crc kubenswrapper[4955]: I0217 13:25:59.175368 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3a87515-64e3-4ed0-bc43-39ea2caa11c0-scripts" (OuterVolumeSpecName: "scripts") pod "f3a87515-64e3-4ed0-bc43-39ea2caa11c0" (UID: "f3a87515-64e3-4ed0-bc43-39ea2caa11c0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:25:59 crc kubenswrapper[4955]: I0217 13:25:59.204253 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3a87515-64e3-4ed0-bc43-39ea2caa11c0-config-data" (OuterVolumeSpecName: "config-data") pod "f3a87515-64e3-4ed0-bc43-39ea2caa11c0" (UID: "f3a87515-64e3-4ed0-bc43-39ea2caa11c0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:25:59 crc kubenswrapper[4955]: I0217 13:25:59.204423 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3a87515-64e3-4ed0-bc43-39ea2caa11c0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f3a87515-64e3-4ed0-bc43-39ea2caa11c0" (UID: "f3a87515-64e3-4ed0-bc43-39ea2caa11c0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:25:59 crc kubenswrapper[4955]: I0217 13:25:59.266453 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3a87515-64e3-4ed0-bc43-39ea2caa11c0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:59 crc kubenswrapper[4955]: I0217 13:25:59.266523 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cnrps\" (UniqueName: \"kubernetes.io/projected/f3a87515-64e3-4ed0-bc43-39ea2caa11c0-kube-api-access-cnrps\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:59 crc kubenswrapper[4955]: I0217 13:25:59.266548 4955 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3a87515-64e3-4ed0-bc43-39ea2caa11c0-scripts\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:59 crc kubenswrapper[4955]: I0217 13:25:59.266563 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3a87515-64e3-4ed0-bc43-39ea2caa11c0-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:25:59 crc kubenswrapper[4955]: I0217 13:25:59.723724 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-4g28c" event={"ID":"f3a87515-64e3-4ed0-bc43-39ea2caa11c0","Type":"ContainerDied","Data":"1bd5102c0d2874e12dd41bf4c88a13b50ea2d30bb954f62efd6121339cd3d9b5"} Feb 17 13:25:59 crc kubenswrapper[4955]: I0217 13:25:59.723764 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1bd5102c0d2874e12dd41bf4c88a13b50ea2d30bb954f62efd6121339cd3d9b5" Feb 17 13:25:59 crc kubenswrapper[4955]: I0217 13:25:59.724179 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-4g28c" Feb 17 13:25:59 crc kubenswrapper[4955]: I0217 13:25:59.911830 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 17 13:25:59 crc kubenswrapper[4955]: I0217 13:25:59.912077 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="6cb6b03f-a3f9-4a85-9db3-c504f959634e" containerName="nova-api-log" containerID="cri-o://dd97550f6e291125acb0ee9d6a3bb7bea3996a670b71e9f34904446a033c0d48" gracePeriod=30 Feb 17 13:25:59 crc kubenswrapper[4955]: I0217 13:25:59.912613 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="6cb6b03f-a3f9-4a85-9db3-c504f959634e" containerName="nova-api-api" containerID="cri-o://9456cf3d2171a35a016c053f0b796677c21a625fcc03cc85a69903a46e9f1888" gracePeriod=30 Feb 17 13:25:59 crc kubenswrapper[4955]: I0217 13:25:59.933726 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 17 13:25:59 crc kubenswrapper[4955]: I0217 13:25:59.933956 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="7a893348-52b9-474c-8031-e67020c445eb" containerName="nova-scheduler-scheduler" containerID="cri-o://5e0370a1ee53a812b14d0cf1a49db5bf79f6f7d91dbbfb8601a5bc8ec1b0f646" gracePeriod=30 Feb 17 13:25:59 crc kubenswrapper[4955]: I0217 13:25:59.948112 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 17 13:25:59 crc kubenswrapper[4955]: I0217 13:25:59.948385 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3fe1264c-d2ce-455a-99b9-88270f4dfb04" containerName="nova-metadata-log" containerID="cri-o://55bcc64e4aacdb624eb82ffeda3b66fe603131d8da1ae9d31139305217a5f598" gracePeriod=30 Feb 17 13:25:59 crc kubenswrapper[4955]: I0217 13:25:59.948539 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3fe1264c-d2ce-455a-99b9-88270f4dfb04" containerName="nova-metadata-metadata" containerID="cri-o://fa8c554eed043625d8a3e30ac076a7f926ace945ba0d86ad179212d29d9bae2e" gracePeriod=30 Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.447629 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.594476 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2pfl\" (UniqueName: \"kubernetes.io/projected/6cb6b03f-a3f9-4a85-9db3-c504f959634e-kube-api-access-q2pfl\") pod \"6cb6b03f-a3f9-4a85-9db3-c504f959634e\" (UID: \"6cb6b03f-a3f9-4a85-9db3-c504f959634e\") " Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.594569 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6cb6b03f-a3f9-4a85-9db3-c504f959634e-logs\") pod \"6cb6b03f-a3f9-4a85-9db3-c504f959634e\" (UID: \"6cb6b03f-a3f9-4a85-9db3-c504f959634e\") " Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.594655 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cb6b03f-a3f9-4a85-9db3-c504f959634e-config-data\") pod \"6cb6b03f-a3f9-4a85-9db3-c504f959634e\" (UID: \"6cb6b03f-a3f9-4a85-9db3-c504f959634e\") " Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.594698 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cb6b03f-a3f9-4a85-9db3-c504f959634e-internal-tls-certs\") pod \"6cb6b03f-a3f9-4a85-9db3-c504f959634e\" (UID: \"6cb6b03f-a3f9-4a85-9db3-c504f959634e\") " Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.594783 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cb6b03f-a3f9-4a85-9db3-c504f959634e-public-tls-certs\") pod \"6cb6b03f-a3f9-4a85-9db3-c504f959634e\" (UID: \"6cb6b03f-a3f9-4a85-9db3-c504f959634e\") " Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.594817 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cb6b03f-a3f9-4a85-9db3-c504f959634e-combined-ca-bundle\") pod \"6cb6b03f-a3f9-4a85-9db3-c504f959634e\" (UID: \"6cb6b03f-a3f9-4a85-9db3-c504f959634e\") " Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.604757 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6cb6b03f-a3f9-4a85-9db3-c504f959634e-logs" (OuterVolumeSpecName: "logs") pod "6cb6b03f-a3f9-4a85-9db3-c504f959634e" (UID: "6cb6b03f-a3f9-4a85-9db3-c504f959634e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.621080 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6cb6b03f-a3f9-4a85-9db3-c504f959634e-kube-api-access-q2pfl" (OuterVolumeSpecName: "kube-api-access-q2pfl") pod "6cb6b03f-a3f9-4a85-9db3-c504f959634e" (UID: "6cb6b03f-a3f9-4a85-9db3-c504f959634e"). InnerVolumeSpecName "kube-api-access-q2pfl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.686019 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cb6b03f-a3f9-4a85-9db3-c504f959634e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6cb6b03f-a3f9-4a85-9db3-c504f959634e" (UID: "6cb6b03f-a3f9-4a85-9db3-c504f959634e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.689143 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cb6b03f-a3f9-4a85-9db3-c504f959634e-config-data" (OuterVolumeSpecName: "config-data") pod "6cb6b03f-a3f9-4a85-9db3-c504f959634e" (UID: "6cb6b03f-a3f9-4a85-9db3-c504f959634e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.697389 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2pfl\" (UniqueName: \"kubernetes.io/projected/6cb6b03f-a3f9-4a85-9db3-c504f959634e-kube-api-access-q2pfl\") on node \"crc\" DevicePath \"\"" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.697423 4955 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6cb6b03f-a3f9-4a85-9db3-c504f959634e-logs\") on node \"crc\" DevicePath \"\"" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.697436 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cb6b03f-a3f9-4a85-9db3-c504f959634e-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.697447 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cb6b03f-a3f9-4a85-9db3-c504f959634e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.698730 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cb6b03f-a3f9-4a85-9db3-c504f959634e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "6cb6b03f-a3f9-4a85-9db3-c504f959634e" (UID: "6cb6b03f-a3f9-4a85-9db3-c504f959634e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.704324 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cb6b03f-a3f9-4a85-9db3-c504f959634e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "6cb6b03f-a3f9-4a85-9db3-c504f959634e" (UID: "6cb6b03f-a3f9-4a85-9db3-c504f959634e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.736011 4955 generic.go:334] "Generic (PLEG): container finished" podID="6cb6b03f-a3f9-4a85-9db3-c504f959634e" containerID="9456cf3d2171a35a016c053f0b796677c21a625fcc03cc85a69903a46e9f1888" exitCode=0 Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.736044 4955 generic.go:334] "Generic (PLEG): container finished" podID="6cb6b03f-a3f9-4a85-9db3-c504f959634e" containerID="dd97550f6e291125acb0ee9d6a3bb7bea3996a670b71e9f34904446a033c0d48" exitCode=143 Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.736069 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.736089 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6cb6b03f-a3f9-4a85-9db3-c504f959634e","Type":"ContainerDied","Data":"9456cf3d2171a35a016c053f0b796677c21a625fcc03cc85a69903a46e9f1888"} Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.736118 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6cb6b03f-a3f9-4a85-9db3-c504f959634e","Type":"ContainerDied","Data":"dd97550f6e291125acb0ee9d6a3bb7bea3996a670b71e9f34904446a033c0d48"} Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.736131 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6cb6b03f-a3f9-4a85-9db3-c504f959634e","Type":"ContainerDied","Data":"4e1768eb4dfc8221cd30f19e3115718e6122d70e484cb3f5a3aa7e8547a3265c"} Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.736147 4955 scope.go:117] "RemoveContainer" containerID="9456cf3d2171a35a016c053f0b796677c21a625fcc03cc85a69903a46e9f1888" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.740091 4955 generic.go:334] "Generic (PLEG): container finished" podID="3fe1264c-d2ce-455a-99b9-88270f4dfb04" containerID="55bcc64e4aacdb624eb82ffeda3b66fe603131d8da1ae9d31139305217a5f598" exitCode=143 Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.740158 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3fe1264c-d2ce-455a-99b9-88270f4dfb04","Type":"ContainerDied","Data":"55bcc64e4aacdb624eb82ffeda3b66fe603131d8da1ae9d31139305217a5f598"} Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.764750 4955 scope.go:117] "RemoveContainer" containerID="dd97550f6e291125acb0ee9d6a3bb7bea3996a670b71e9f34904446a033c0d48" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.769633 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.780642 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.786483 4955 scope.go:117] "RemoveContainer" containerID="9456cf3d2171a35a016c053f0b796677c21a625fcc03cc85a69903a46e9f1888" Feb 17 13:26:00 crc kubenswrapper[4955]: E0217 13:26:00.787064 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9456cf3d2171a35a016c053f0b796677c21a625fcc03cc85a69903a46e9f1888\": container with ID starting with 9456cf3d2171a35a016c053f0b796677c21a625fcc03cc85a69903a46e9f1888 not found: ID does not exist" containerID="9456cf3d2171a35a016c053f0b796677c21a625fcc03cc85a69903a46e9f1888" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.787112 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9456cf3d2171a35a016c053f0b796677c21a625fcc03cc85a69903a46e9f1888"} err="failed to get container status \"9456cf3d2171a35a016c053f0b796677c21a625fcc03cc85a69903a46e9f1888\": rpc error: code = NotFound desc = could not find container \"9456cf3d2171a35a016c053f0b796677c21a625fcc03cc85a69903a46e9f1888\": container with ID starting with 9456cf3d2171a35a016c053f0b796677c21a625fcc03cc85a69903a46e9f1888 not found: ID does not exist" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.787139 4955 scope.go:117] "RemoveContainer" containerID="dd97550f6e291125acb0ee9d6a3bb7bea3996a670b71e9f34904446a033c0d48" Feb 17 13:26:00 crc kubenswrapper[4955]: E0217 13:26:00.787481 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd97550f6e291125acb0ee9d6a3bb7bea3996a670b71e9f34904446a033c0d48\": container with ID starting with dd97550f6e291125acb0ee9d6a3bb7bea3996a670b71e9f34904446a033c0d48 not found: ID does not exist" containerID="dd97550f6e291125acb0ee9d6a3bb7bea3996a670b71e9f34904446a033c0d48" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.787513 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd97550f6e291125acb0ee9d6a3bb7bea3996a670b71e9f34904446a033c0d48"} err="failed to get container status \"dd97550f6e291125acb0ee9d6a3bb7bea3996a670b71e9f34904446a033c0d48\": rpc error: code = NotFound desc = could not find container \"dd97550f6e291125acb0ee9d6a3bb7bea3996a670b71e9f34904446a033c0d48\": container with ID starting with dd97550f6e291125acb0ee9d6a3bb7bea3996a670b71e9f34904446a033c0d48 not found: ID does not exist" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.787538 4955 scope.go:117] "RemoveContainer" containerID="9456cf3d2171a35a016c053f0b796677c21a625fcc03cc85a69903a46e9f1888" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.787975 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9456cf3d2171a35a016c053f0b796677c21a625fcc03cc85a69903a46e9f1888"} err="failed to get container status \"9456cf3d2171a35a016c053f0b796677c21a625fcc03cc85a69903a46e9f1888\": rpc error: code = NotFound desc = could not find container \"9456cf3d2171a35a016c053f0b796677c21a625fcc03cc85a69903a46e9f1888\": container with ID starting with 9456cf3d2171a35a016c053f0b796677c21a625fcc03cc85a69903a46e9f1888 not found: ID does not exist" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.787992 4955 scope.go:117] "RemoveContainer" containerID="dd97550f6e291125acb0ee9d6a3bb7bea3996a670b71e9f34904446a033c0d48" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.788922 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd97550f6e291125acb0ee9d6a3bb7bea3996a670b71e9f34904446a033c0d48"} err="failed to get container status \"dd97550f6e291125acb0ee9d6a3bb7bea3996a670b71e9f34904446a033c0d48\": rpc error: code = NotFound desc = could not find container \"dd97550f6e291125acb0ee9d6a3bb7bea3996a670b71e9f34904446a033c0d48\": container with ID starting with dd97550f6e291125acb0ee9d6a3bb7bea3996a670b71e9f34904446a033c0d48 not found: ID does not exist" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.791949 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 17 13:26:00 crc kubenswrapper[4955]: E0217 13:26:00.792361 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="987ec26b-c18e-47ec-be15-24059d88e961" containerName="init" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.792380 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="987ec26b-c18e-47ec-be15-24059d88e961" containerName="init" Feb 17 13:26:00 crc kubenswrapper[4955]: E0217 13:26:00.792400 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cb6b03f-a3f9-4a85-9db3-c504f959634e" containerName="nova-api-api" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.792406 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cb6b03f-a3f9-4a85-9db3-c504f959634e" containerName="nova-api-api" Feb 17 13:26:00 crc kubenswrapper[4955]: E0217 13:26:00.792421 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3a87515-64e3-4ed0-bc43-39ea2caa11c0" containerName="nova-manage" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.792427 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3a87515-64e3-4ed0-bc43-39ea2caa11c0" containerName="nova-manage" Feb 17 13:26:00 crc kubenswrapper[4955]: E0217 13:26:00.792443 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cb6b03f-a3f9-4a85-9db3-c504f959634e" containerName="nova-api-log" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.792450 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cb6b03f-a3f9-4a85-9db3-c504f959634e" containerName="nova-api-log" Feb 17 13:26:00 crc kubenswrapper[4955]: E0217 13:26:00.792467 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="987ec26b-c18e-47ec-be15-24059d88e961" containerName="dnsmasq-dns" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.792472 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="987ec26b-c18e-47ec-be15-24059d88e961" containerName="dnsmasq-dns" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.792637 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="987ec26b-c18e-47ec-be15-24059d88e961" containerName="dnsmasq-dns" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.792663 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cb6b03f-a3f9-4a85-9db3-c504f959634e" containerName="nova-api-log" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.792674 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cb6b03f-a3f9-4a85-9db3-c504f959634e" containerName="nova-api-api" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.792686 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3a87515-64e3-4ed0-bc43-39ea2caa11c0" containerName="nova-manage" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.793721 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.795894 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.799204 4955 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cb6b03f-a3f9-4a85-9db3-c504f959634e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.799242 4955 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cb6b03f-a3f9-4a85-9db3-c504f959634e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.802818 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.803118 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.814565 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.900561 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a37e38-9f12-49ad-8464-f744f7f5fc2f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a1a37e38-9f12-49ad-8464-f744f7f5fc2f\") " pod="openstack/nova-api-0" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.900626 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1a37e38-9f12-49ad-8464-f744f7f5fc2f-public-tls-certs\") pod \"nova-api-0\" (UID: \"a1a37e38-9f12-49ad-8464-f744f7f5fc2f\") " pod="openstack/nova-api-0" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.900648 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1a37e38-9f12-49ad-8464-f744f7f5fc2f-config-data\") pod \"nova-api-0\" (UID: \"a1a37e38-9f12-49ad-8464-f744f7f5fc2f\") " pod="openstack/nova-api-0" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.900901 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvbzt\" (UniqueName: \"kubernetes.io/projected/a1a37e38-9f12-49ad-8464-f744f7f5fc2f-kube-api-access-rvbzt\") pod \"nova-api-0\" (UID: \"a1a37e38-9f12-49ad-8464-f744f7f5fc2f\") " pod="openstack/nova-api-0" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.901057 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1a37e38-9f12-49ad-8464-f744f7f5fc2f-logs\") pod \"nova-api-0\" (UID: \"a1a37e38-9f12-49ad-8464-f744f7f5fc2f\") " pod="openstack/nova-api-0" Feb 17 13:26:00 crc kubenswrapper[4955]: I0217 13:26:00.901085 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1a37e38-9f12-49ad-8464-f744f7f5fc2f-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a1a37e38-9f12-49ad-8464-f744f7f5fc2f\") " pod="openstack/nova-api-0" Feb 17 13:26:01 crc kubenswrapper[4955]: I0217 13:26:01.004695 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvbzt\" (UniqueName: \"kubernetes.io/projected/a1a37e38-9f12-49ad-8464-f744f7f5fc2f-kube-api-access-rvbzt\") pod \"nova-api-0\" (UID: \"a1a37e38-9f12-49ad-8464-f744f7f5fc2f\") " pod="openstack/nova-api-0" Feb 17 13:26:01 crc kubenswrapper[4955]: I0217 13:26:01.004829 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1a37e38-9f12-49ad-8464-f744f7f5fc2f-logs\") pod \"nova-api-0\" (UID: \"a1a37e38-9f12-49ad-8464-f744f7f5fc2f\") " pod="openstack/nova-api-0" Feb 17 13:26:01 crc kubenswrapper[4955]: I0217 13:26:01.004853 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1a37e38-9f12-49ad-8464-f744f7f5fc2f-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a1a37e38-9f12-49ad-8464-f744f7f5fc2f\") " pod="openstack/nova-api-0" Feb 17 13:26:01 crc kubenswrapper[4955]: I0217 13:26:01.004934 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a37e38-9f12-49ad-8464-f744f7f5fc2f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a1a37e38-9f12-49ad-8464-f744f7f5fc2f\") " pod="openstack/nova-api-0" Feb 17 13:26:01 crc kubenswrapper[4955]: I0217 13:26:01.004970 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1a37e38-9f12-49ad-8464-f744f7f5fc2f-public-tls-certs\") pod \"nova-api-0\" (UID: \"a1a37e38-9f12-49ad-8464-f744f7f5fc2f\") " pod="openstack/nova-api-0" Feb 17 13:26:01 crc kubenswrapper[4955]: I0217 13:26:01.004989 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1a37e38-9f12-49ad-8464-f744f7f5fc2f-config-data\") pod \"nova-api-0\" (UID: \"a1a37e38-9f12-49ad-8464-f744f7f5fc2f\") " pod="openstack/nova-api-0" Feb 17 13:26:01 crc kubenswrapper[4955]: I0217 13:26:01.005199 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1a37e38-9f12-49ad-8464-f744f7f5fc2f-logs\") pod \"nova-api-0\" (UID: \"a1a37e38-9f12-49ad-8464-f744f7f5fc2f\") " pod="openstack/nova-api-0" Feb 17 13:26:01 crc kubenswrapper[4955]: I0217 13:26:01.014879 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1a37e38-9f12-49ad-8464-f744f7f5fc2f-public-tls-certs\") pod \"nova-api-0\" (UID: \"a1a37e38-9f12-49ad-8464-f744f7f5fc2f\") " pod="openstack/nova-api-0" Feb 17 13:26:01 crc kubenswrapper[4955]: I0217 13:26:01.015505 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a37e38-9f12-49ad-8464-f744f7f5fc2f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a1a37e38-9f12-49ad-8464-f744f7f5fc2f\") " pod="openstack/nova-api-0" Feb 17 13:26:01 crc kubenswrapper[4955]: I0217 13:26:01.015624 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1a37e38-9f12-49ad-8464-f744f7f5fc2f-config-data\") pod \"nova-api-0\" (UID: \"a1a37e38-9f12-49ad-8464-f744f7f5fc2f\") " pod="openstack/nova-api-0" Feb 17 13:26:01 crc kubenswrapper[4955]: I0217 13:26:01.016581 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1a37e38-9f12-49ad-8464-f744f7f5fc2f-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a1a37e38-9f12-49ad-8464-f744f7f5fc2f\") " pod="openstack/nova-api-0" Feb 17 13:26:01 crc kubenswrapper[4955]: I0217 13:26:01.032114 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvbzt\" (UniqueName: \"kubernetes.io/projected/a1a37e38-9f12-49ad-8464-f744f7f5fc2f-kube-api-access-rvbzt\") pod \"nova-api-0\" (UID: \"a1a37e38-9f12-49ad-8464-f744f7f5fc2f\") " pod="openstack/nova-api-0" Feb 17 13:26:01 crc kubenswrapper[4955]: I0217 13:26:01.129554 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 17 13:26:01 crc kubenswrapper[4955]: E0217 13:26:01.511555 4955 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5e0370a1ee53a812b14d0cf1a49db5bf79f6f7d91dbbfb8601a5bc8ec1b0f646" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 17 13:26:01 crc kubenswrapper[4955]: E0217 13:26:01.513409 4955 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5e0370a1ee53a812b14d0cf1a49db5bf79f6f7d91dbbfb8601a5bc8ec1b0f646" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 17 13:26:01 crc kubenswrapper[4955]: E0217 13:26:01.515149 4955 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5e0370a1ee53a812b14d0cf1a49db5bf79f6f7d91dbbfb8601a5bc8ec1b0f646" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 17 13:26:01 crc kubenswrapper[4955]: E0217 13:26:01.515233 4955 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="7a893348-52b9-474c-8031-e67020c445eb" containerName="nova-scheduler-scheduler" Feb 17 13:26:01 crc kubenswrapper[4955]: I0217 13:26:01.612114 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 17 13:26:01 crc kubenswrapper[4955]: W0217 13:26:01.625601 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda1a37e38_9f12_49ad_8464_f744f7f5fc2f.slice/crio-aacc3e3796cdae049d1514748f710d99d17bac55c08e780afefad31ca12f6c41 WatchSource:0}: Error finding container aacc3e3796cdae049d1514748f710d99d17bac55c08e780afefad31ca12f6c41: Status 404 returned error can't find the container with id aacc3e3796cdae049d1514748f710d99d17bac55c08e780afefad31ca12f6c41 Feb 17 13:26:01 crc kubenswrapper[4955]: I0217 13:26:01.753254 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a1a37e38-9f12-49ad-8464-f744f7f5fc2f","Type":"ContainerStarted","Data":"aacc3e3796cdae049d1514748f710d99d17bac55c08e780afefad31ca12f6c41"} Feb 17 13:26:02 crc kubenswrapper[4955]: I0217 13:26:02.236594 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6cb6b03f-a3f9-4a85-9db3-c504f959634e" path="/var/lib/kubelet/pods/6cb6b03f-a3f9-4a85-9db3-c504f959634e/volumes" Feb 17 13:26:02 crc kubenswrapper[4955]: I0217 13:26:02.765222 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a1a37e38-9f12-49ad-8464-f744f7f5fc2f","Type":"ContainerStarted","Data":"8cfdf4eba30afb32da947d8ea6ce20a6458d354f218708cc0cfa89b01bd5e056"} Feb 17 13:26:02 crc kubenswrapper[4955]: I0217 13:26:02.765549 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a1a37e38-9f12-49ad-8464-f744f7f5fc2f","Type":"ContainerStarted","Data":"0ab9898f7493382327ec1a8ef9e26603a85a679722ad3199fd91a4ef06546f26"} Feb 17 13:26:02 crc kubenswrapper[4955]: I0217 13:26:02.801186 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.801155533 podStartE2EDuration="2.801155533s" podCreationTimestamp="2026-02-17 13:26:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:26:02.785074732 +0000 UTC m=+1301.307804285" watchObservedRunningTime="2026-02-17 13:26:02.801155533 +0000 UTC m=+1301.323885116" Feb 17 13:26:03 crc kubenswrapper[4955]: I0217 13:26:03.544596 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 17 13:26:03 crc kubenswrapper[4955]: I0217 13:26:03.652189 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxckz\" (UniqueName: \"kubernetes.io/projected/3fe1264c-d2ce-455a-99b9-88270f4dfb04-kube-api-access-vxckz\") pod \"3fe1264c-d2ce-455a-99b9-88270f4dfb04\" (UID: \"3fe1264c-d2ce-455a-99b9-88270f4dfb04\") " Feb 17 13:26:03 crc kubenswrapper[4955]: I0217 13:26:03.652266 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3fe1264c-d2ce-455a-99b9-88270f4dfb04-nova-metadata-tls-certs\") pod \"3fe1264c-d2ce-455a-99b9-88270f4dfb04\" (UID: \"3fe1264c-d2ce-455a-99b9-88270f4dfb04\") " Feb 17 13:26:03 crc kubenswrapper[4955]: I0217 13:26:03.652346 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3fe1264c-d2ce-455a-99b9-88270f4dfb04-logs\") pod \"3fe1264c-d2ce-455a-99b9-88270f4dfb04\" (UID: \"3fe1264c-d2ce-455a-99b9-88270f4dfb04\") " Feb 17 13:26:03 crc kubenswrapper[4955]: I0217 13:26:03.652370 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fe1264c-d2ce-455a-99b9-88270f4dfb04-config-data\") pod \"3fe1264c-d2ce-455a-99b9-88270f4dfb04\" (UID: \"3fe1264c-d2ce-455a-99b9-88270f4dfb04\") " Feb 17 13:26:03 crc kubenswrapper[4955]: I0217 13:26:03.652473 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fe1264c-d2ce-455a-99b9-88270f4dfb04-combined-ca-bundle\") pod \"3fe1264c-d2ce-455a-99b9-88270f4dfb04\" (UID: \"3fe1264c-d2ce-455a-99b9-88270f4dfb04\") " Feb 17 13:26:03 crc kubenswrapper[4955]: I0217 13:26:03.653294 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3fe1264c-d2ce-455a-99b9-88270f4dfb04-logs" (OuterVolumeSpecName: "logs") pod "3fe1264c-d2ce-455a-99b9-88270f4dfb04" (UID: "3fe1264c-d2ce-455a-99b9-88270f4dfb04"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:26:03 crc kubenswrapper[4955]: I0217 13:26:03.657698 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fe1264c-d2ce-455a-99b9-88270f4dfb04-kube-api-access-vxckz" (OuterVolumeSpecName: "kube-api-access-vxckz") pod "3fe1264c-d2ce-455a-99b9-88270f4dfb04" (UID: "3fe1264c-d2ce-455a-99b9-88270f4dfb04"). InnerVolumeSpecName "kube-api-access-vxckz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:26:03 crc kubenswrapper[4955]: I0217 13:26:03.678678 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fe1264c-d2ce-455a-99b9-88270f4dfb04-config-data" (OuterVolumeSpecName: "config-data") pod "3fe1264c-d2ce-455a-99b9-88270f4dfb04" (UID: "3fe1264c-d2ce-455a-99b9-88270f4dfb04"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:26:03 crc kubenswrapper[4955]: I0217 13:26:03.700163 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fe1264c-d2ce-455a-99b9-88270f4dfb04-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3fe1264c-d2ce-455a-99b9-88270f4dfb04" (UID: "3fe1264c-d2ce-455a-99b9-88270f4dfb04"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:26:03 crc kubenswrapper[4955]: I0217 13:26:03.703293 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fe1264c-d2ce-455a-99b9-88270f4dfb04-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "3fe1264c-d2ce-455a-99b9-88270f4dfb04" (UID: "3fe1264c-d2ce-455a-99b9-88270f4dfb04"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:26:03 crc kubenswrapper[4955]: I0217 13:26:03.754399 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fe1264c-d2ce-455a-99b9-88270f4dfb04-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:26:03 crc kubenswrapper[4955]: I0217 13:26:03.754438 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxckz\" (UniqueName: \"kubernetes.io/projected/3fe1264c-d2ce-455a-99b9-88270f4dfb04-kube-api-access-vxckz\") on node \"crc\" DevicePath \"\"" Feb 17 13:26:03 crc kubenswrapper[4955]: I0217 13:26:03.754449 4955 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3fe1264c-d2ce-455a-99b9-88270f4dfb04-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 17 13:26:03 crc kubenswrapper[4955]: I0217 13:26:03.754459 4955 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3fe1264c-d2ce-455a-99b9-88270f4dfb04-logs\") on node \"crc\" DevicePath \"\"" Feb 17 13:26:03 crc kubenswrapper[4955]: I0217 13:26:03.754467 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fe1264c-d2ce-455a-99b9-88270f4dfb04-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:26:03 crc kubenswrapper[4955]: I0217 13:26:03.781227 4955 generic.go:334] "Generic (PLEG): container finished" podID="3fe1264c-d2ce-455a-99b9-88270f4dfb04" containerID="fa8c554eed043625d8a3e30ac076a7f926ace945ba0d86ad179212d29d9bae2e" exitCode=0 Feb 17 13:26:03 crc kubenswrapper[4955]: I0217 13:26:03.781279 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3fe1264c-d2ce-455a-99b9-88270f4dfb04","Type":"ContainerDied","Data":"fa8c554eed043625d8a3e30ac076a7f926ace945ba0d86ad179212d29d9bae2e"} Feb 17 13:26:03 crc kubenswrapper[4955]: I0217 13:26:03.781344 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3fe1264c-d2ce-455a-99b9-88270f4dfb04","Type":"ContainerDied","Data":"e29f177fa16272579c55e170117e853d9f3bfbfd1a79199137846ee8ef65ec82"} Feb 17 13:26:03 crc kubenswrapper[4955]: I0217 13:26:03.781340 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 17 13:26:03 crc kubenswrapper[4955]: I0217 13:26:03.781398 4955 scope.go:117] "RemoveContainer" containerID="fa8c554eed043625d8a3e30ac076a7f926ace945ba0d86ad179212d29d9bae2e" Feb 17 13:26:03 crc kubenswrapper[4955]: I0217 13:26:03.809589 4955 scope.go:117] "RemoveContainer" containerID="55bcc64e4aacdb624eb82ffeda3b66fe603131d8da1ae9d31139305217a5f598" Feb 17 13:26:03 crc kubenswrapper[4955]: I0217 13:26:03.817456 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 17 13:26:03 crc kubenswrapper[4955]: I0217 13:26:03.836116 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 17 13:26:03 crc kubenswrapper[4955]: I0217 13:26:03.844882 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 17 13:26:03 crc kubenswrapper[4955]: E0217 13:26:03.854708 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fe1264c-d2ce-455a-99b9-88270f4dfb04" containerName="nova-metadata-metadata" Feb 17 13:26:03 crc kubenswrapper[4955]: I0217 13:26:03.854765 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fe1264c-d2ce-455a-99b9-88270f4dfb04" containerName="nova-metadata-metadata" Feb 17 13:26:03 crc kubenswrapper[4955]: E0217 13:26:03.854849 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fe1264c-d2ce-455a-99b9-88270f4dfb04" containerName="nova-metadata-log" Feb 17 13:26:03 crc kubenswrapper[4955]: I0217 13:26:03.854860 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fe1264c-d2ce-455a-99b9-88270f4dfb04" containerName="nova-metadata-log" Feb 17 13:26:03 crc kubenswrapper[4955]: I0217 13:26:03.855298 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fe1264c-d2ce-455a-99b9-88270f4dfb04" containerName="nova-metadata-log" Feb 17 13:26:03 crc kubenswrapper[4955]: I0217 13:26:03.855352 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fe1264c-d2ce-455a-99b9-88270f4dfb04" containerName="nova-metadata-metadata" Feb 17 13:26:03 crc kubenswrapper[4955]: I0217 13:26:03.855665 4955 scope.go:117] "RemoveContainer" containerID="fa8c554eed043625d8a3e30ac076a7f926ace945ba0d86ad179212d29d9bae2e" Feb 17 13:26:03 crc kubenswrapper[4955]: E0217 13:26:03.856670 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa8c554eed043625d8a3e30ac076a7f926ace945ba0d86ad179212d29d9bae2e\": container with ID starting with fa8c554eed043625d8a3e30ac076a7f926ace945ba0d86ad179212d29d9bae2e not found: ID does not exist" containerID="fa8c554eed043625d8a3e30ac076a7f926ace945ba0d86ad179212d29d9bae2e" Feb 17 13:26:03 crc kubenswrapper[4955]: I0217 13:26:03.856717 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa8c554eed043625d8a3e30ac076a7f926ace945ba0d86ad179212d29d9bae2e"} err="failed to get container status \"fa8c554eed043625d8a3e30ac076a7f926ace945ba0d86ad179212d29d9bae2e\": rpc error: code = NotFound desc = could not find container \"fa8c554eed043625d8a3e30ac076a7f926ace945ba0d86ad179212d29d9bae2e\": container with ID starting with fa8c554eed043625d8a3e30ac076a7f926ace945ba0d86ad179212d29d9bae2e not found: ID does not exist" Feb 17 13:26:03 crc kubenswrapper[4955]: I0217 13:26:03.856753 4955 scope.go:117] "RemoveContainer" containerID="55bcc64e4aacdb624eb82ffeda3b66fe603131d8da1ae9d31139305217a5f598" Feb 17 13:26:03 crc kubenswrapper[4955]: I0217 13:26:03.856853 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 17 13:26:03 crc kubenswrapper[4955]: E0217 13:26:03.857439 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55bcc64e4aacdb624eb82ffeda3b66fe603131d8da1ae9d31139305217a5f598\": container with ID starting with 55bcc64e4aacdb624eb82ffeda3b66fe603131d8da1ae9d31139305217a5f598 not found: ID does not exist" containerID="55bcc64e4aacdb624eb82ffeda3b66fe603131d8da1ae9d31139305217a5f598" Feb 17 13:26:03 crc kubenswrapper[4955]: I0217 13:26:03.857570 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55bcc64e4aacdb624eb82ffeda3b66fe603131d8da1ae9d31139305217a5f598"} err="failed to get container status \"55bcc64e4aacdb624eb82ffeda3b66fe603131d8da1ae9d31139305217a5f598\": rpc error: code = NotFound desc = could not find container \"55bcc64e4aacdb624eb82ffeda3b66fe603131d8da1ae9d31139305217a5f598\": container with ID starting with 55bcc64e4aacdb624eb82ffeda3b66fe603131d8da1ae9d31139305217a5f598 not found: ID does not exist" Feb 17 13:26:03 crc kubenswrapper[4955]: I0217 13:26:03.859191 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 17 13:26:03 crc kubenswrapper[4955]: I0217 13:26:03.859421 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 17 13:26:03 crc kubenswrapper[4955]: I0217 13:26:03.859877 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 17 13:26:03 crc kubenswrapper[4955]: I0217 13:26:03.958281 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff800864-d970-46b0-bb2e-eb213d423da4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ff800864-d970-46b0-bb2e-eb213d423da4\") " pod="openstack/nova-metadata-0" Feb 17 13:26:03 crc kubenswrapper[4955]: I0217 13:26:03.958343 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlcpp\" (UniqueName: \"kubernetes.io/projected/ff800864-d970-46b0-bb2e-eb213d423da4-kube-api-access-qlcpp\") pod \"nova-metadata-0\" (UID: \"ff800864-d970-46b0-bb2e-eb213d423da4\") " pod="openstack/nova-metadata-0" Feb 17 13:26:03 crc kubenswrapper[4955]: I0217 13:26:03.958403 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff800864-d970-46b0-bb2e-eb213d423da4-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ff800864-d970-46b0-bb2e-eb213d423da4\") " pod="openstack/nova-metadata-0" Feb 17 13:26:03 crc kubenswrapper[4955]: I0217 13:26:03.958516 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff800864-d970-46b0-bb2e-eb213d423da4-logs\") pod \"nova-metadata-0\" (UID: \"ff800864-d970-46b0-bb2e-eb213d423da4\") " pod="openstack/nova-metadata-0" Feb 17 13:26:03 crc kubenswrapper[4955]: I0217 13:26:03.958564 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff800864-d970-46b0-bb2e-eb213d423da4-config-data\") pod \"nova-metadata-0\" (UID: \"ff800864-d970-46b0-bb2e-eb213d423da4\") " pod="openstack/nova-metadata-0" Feb 17 13:26:04 crc kubenswrapper[4955]: I0217 13:26:04.060059 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff800864-d970-46b0-bb2e-eb213d423da4-logs\") pod \"nova-metadata-0\" (UID: \"ff800864-d970-46b0-bb2e-eb213d423da4\") " pod="openstack/nova-metadata-0" Feb 17 13:26:04 crc kubenswrapper[4955]: I0217 13:26:04.060390 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff800864-d970-46b0-bb2e-eb213d423da4-config-data\") pod \"nova-metadata-0\" (UID: \"ff800864-d970-46b0-bb2e-eb213d423da4\") " pod="openstack/nova-metadata-0" Feb 17 13:26:04 crc kubenswrapper[4955]: I0217 13:26:04.060559 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff800864-d970-46b0-bb2e-eb213d423da4-logs\") pod \"nova-metadata-0\" (UID: \"ff800864-d970-46b0-bb2e-eb213d423da4\") " pod="openstack/nova-metadata-0" Feb 17 13:26:04 crc kubenswrapper[4955]: I0217 13:26:04.060738 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff800864-d970-46b0-bb2e-eb213d423da4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ff800864-d970-46b0-bb2e-eb213d423da4\") " pod="openstack/nova-metadata-0" Feb 17 13:26:04 crc kubenswrapper[4955]: I0217 13:26:04.060939 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlcpp\" (UniqueName: \"kubernetes.io/projected/ff800864-d970-46b0-bb2e-eb213d423da4-kube-api-access-qlcpp\") pod \"nova-metadata-0\" (UID: \"ff800864-d970-46b0-bb2e-eb213d423da4\") " pod="openstack/nova-metadata-0" Feb 17 13:26:04 crc kubenswrapper[4955]: I0217 13:26:04.061076 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff800864-d970-46b0-bb2e-eb213d423da4-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ff800864-d970-46b0-bb2e-eb213d423da4\") " pod="openstack/nova-metadata-0" Feb 17 13:26:04 crc kubenswrapper[4955]: I0217 13:26:04.064373 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff800864-d970-46b0-bb2e-eb213d423da4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ff800864-d970-46b0-bb2e-eb213d423da4\") " pod="openstack/nova-metadata-0" Feb 17 13:26:04 crc kubenswrapper[4955]: I0217 13:26:04.064455 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff800864-d970-46b0-bb2e-eb213d423da4-config-data\") pod \"nova-metadata-0\" (UID: \"ff800864-d970-46b0-bb2e-eb213d423da4\") " pod="openstack/nova-metadata-0" Feb 17 13:26:04 crc kubenswrapper[4955]: I0217 13:26:04.065216 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff800864-d970-46b0-bb2e-eb213d423da4-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ff800864-d970-46b0-bb2e-eb213d423da4\") " pod="openstack/nova-metadata-0" Feb 17 13:26:04 crc kubenswrapper[4955]: I0217 13:26:04.080499 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlcpp\" (UniqueName: \"kubernetes.io/projected/ff800864-d970-46b0-bb2e-eb213d423da4-kube-api-access-qlcpp\") pod \"nova-metadata-0\" (UID: \"ff800864-d970-46b0-bb2e-eb213d423da4\") " pod="openstack/nova-metadata-0" Feb 17 13:26:04 crc kubenswrapper[4955]: I0217 13:26:04.179043 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 17 13:26:04 crc kubenswrapper[4955]: I0217 13:26:04.237238 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3fe1264c-d2ce-455a-99b9-88270f4dfb04" path="/var/lib/kubelet/pods/3fe1264c-d2ce-455a-99b9-88270f4dfb04/volumes" Feb 17 13:26:04 crc kubenswrapper[4955]: I0217 13:26:04.574819 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 13:26:04 crc kubenswrapper[4955]: I0217 13:26:04.575218 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 13:26:04 crc kubenswrapper[4955]: I0217 13:26:04.685942 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 17 13:26:04 crc kubenswrapper[4955]: I0217 13:26:04.790704 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ff800864-d970-46b0-bb2e-eb213d423da4","Type":"ContainerStarted","Data":"0e82244bfaf48a0fb483c6585cbebd0747ad30de727902709abe25c3fb1f42ae"} Feb 17 13:26:05 crc kubenswrapper[4955]: I0217 13:26:05.535889 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 17 13:26:05 crc kubenswrapper[4955]: I0217 13:26:05.592299 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kl7jf\" (UniqueName: \"kubernetes.io/projected/7a893348-52b9-474c-8031-e67020c445eb-kube-api-access-kl7jf\") pod \"7a893348-52b9-474c-8031-e67020c445eb\" (UID: \"7a893348-52b9-474c-8031-e67020c445eb\") " Feb 17 13:26:05 crc kubenswrapper[4955]: I0217 13:26:05.592445 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a893348-52b9-474c-8031-e67020c445eb-combined-ca-bundle\") pod \"7a893348-52b9-474c-8031-e67020c445eb\" (UID: \"7a893348-52b9-474c-8031-e67020c445eb\") " Feb 17 13:26:05 crc kubenswrapper[4955]: I0217 13:26:05.592526 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a893348-52b9-474c-8031-e67020c445eb-config-data\") pod \"7a893348-52b9-474c-8031-e67020c445eb\" (UID: \"7a893348-52b9-474c-8031-e67020c445eb\") " Feb 17 13:26:05 crc kubenswrapper[4955]: I0217 13:26:05.599069 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a893348-52b9-474c-8031-e67020c445eb-kube-api-access-kl7jf" (OuterVolumeSpecName: "kube-api-access-kl7jf") pod "7a893348-52b9-474c-8031-e67020c445eb" (UID: "7a893348-52b9-474c-8031-e67020c445eb"). InnerVolumeSpecName "kube-api-access-kl7jf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:26:05 crc kubenswrapper[4955]: I0217 13:26:05.625395 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a893348-52b9-474c-8031-e67020c445eb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7a893348-52b9-474c-8031-e67020c445eb" (UID: "7a893348-52b9-474c-8031-e67020c445eb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:26:05 crc kubenswrapper[4955]: I0217 13:26:05.626969 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a893348-52b9-474c-8031-e67020c445eb-config-data" (OuterVolumeSpecName: "config-data") pod "7a893348-52b9-474c-8031-e67020c445eb" (UID: "7a893348-52b9-474c-8031-e67020c445eb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:26:05 crc kubenswrapper[4955]: I0217 13:26:05.695000 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a893348-52b9-474c-8031-e67020c445eb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:26:05 crc kubenswrapper[4955]: I0217 13:26:05.695033 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a893348-52b9-474c-8031-e67020c445eb-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:26:05 crc kubenswrapper[4955]: I0217 13:26:05.695043 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kl7jf\" (UniqueName: \"kubernetes.io/projected/7a893348-52b9-474c-8031-e67020c445eb-kube-api-access-kl7jf\") on node \"crc\" DevicePath \"\"" Feb 17 13:26:05 crc kubenswrapper[4955]: I0217 13:26:05.802947 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ff800864-d970-46b0-bb2e-eb213d423da4","Type":"ContainerStarted","Data":"d833444b521365a6adb2eb3d56191cb752594976ee101dabb36d2b99f6a2501b"} Feb 17 13:26:05 crc kubenswrapper[4955]: I0217 13:26:05.802999 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ff800864-d970-46b0-bb2e-eb213d423da4","Type":"ContainerStarted","Data":"032ae00422c9646aaf136d7427de6c4556f8c6171d89097ee22b50f9db853bea"} Feb 17 13:26:05 crc kubenswrapper[4955]: I0217 13:26:05.804532 4955 generic.go:334] "Generic (PLEG): container finished" podID="7a893348-52b9-474c-8031-e67020c445eb" containerID="5e0370a1ee53a812b14d0cf1a49db5bf79f6f7d91dbbfb8601a5bc8ec1b0f646" exitCode=0 Feb 17 13:26:05 crc kubenswrapper[4955]: I0217 13:26:05.804562 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7a893348-52b9-474c-8031-e67020c445eb","Type":"ContainerDied","Data":"5e0370a1ee53a812b14d0cf1a49db5bf79f6f7d91dbbfb8601a5bc8ec1b0f646"} Feb 17 13:26:05 crc kubenswrapper[4955]: I0217 13:26:05.804579 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7a893348-52b9-474c-8031-e67020c445eb","Type":"ContainerDied","Data":"fe1283912db36870d1dfdadb33f8eb25481b222b96a63fd493c39b168d051535"} Feb 17 13:26:05 crc kubenswrapper[4955]: I0217 13:26:05.804597 4955 scope.go:117] "RemoveContainer" containerID="5e0370a1ee53a812b14d0cf1a49db5bf79f6f7d91dbbfb8601a5bc8ec1b0f646" Feb 17 13:26:05 crc kubenswrapper[4955]: I0217 13:26:05.804599 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 17 13:26:05 crc kubenswrapper[4955]: I0217 13:26:05.823364 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.823351564 podStartE2EDuration="2.823351564s" podCreationTimestamp="2026-02-17 13:26:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:26:05.82005301 +0000 UTC m=+1304.342782573" watchObservedRunningTime="2026-02-17 13:26:05.823351564 +0000 UTC m=+1304.346081107" Feb 17 13:26:05 crc kubenswrapper[4955]: I0217 13:26:05.827562 4955 scope.go:117] "RemoveContainer" containerID="5e0370a1ee53a812b14d0cf1a49db5bf79f6f7d91dbbfb8601a5bc8ec1b0f646" Feb 17 13:26:05 crc kubenswrapper[4955]: E0217 13:26:05.828153 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e0370a1ee53a812b14d0cf1a49db5bf79f6f7d91dbbfb8601a5bc8ec1b0f646\": container with ID starting with 5e0370a1ee53a812b14d0cf1a49db5bf79f6f7d91dbbfb8601a5bc8ec1b0f646 not found: ID does not exist" containerID="5e0370a1ee53a812b14d0cf1a49db5bf79f6f7d91dbbfb8601a5bc8ec1b0f646" Feb 17 13:26:05 crc kubenswrapper[4955]: I0217 13:26:05.828189 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e0370a1ee53a812b14d0cf1a49db5bf79f6f7d91dbbfb8601a5bc8ec1b0f646"} err="failed to get container status \"5e0370a1ee53a812b14d0cf1a49db5bf79f6f7d91dbbfb8601a5bc8ec1b0f646\": rpc error: code = NotFound desc = could not find container \"5e0370a1ee53a812b14d0cf1a49db5bf79f6f7d91dbbfb8601a5bc8ec1b0f646\": container with ID starting with 5e0370a1ee53a812b14d0cf1a49db5bf79f6f7d91dbbfb8601a5bc8ec1b0f646 not found: ID does not exist" Feb 17 13:26:05 crc kubenswrapper[4955]: I0217 13:26:05.854477 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 17 13:26:05 crc kubenswrapper[4955]: I0217 13:26:05.872800 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Feb 17 13:26:05 crc kubenswrapper[4955]: I0217 13:26:05.881190 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 17 13:26:05 crc kubenswrapper[4955]: E0217 13:26:05.881716 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a893348-52b9-474c-8031-e67020c445eb" containerName="nova-scheduler-scheduler" Feb 17 13:26:05 crc kubenswrapper[4955]: I0217 13:26:05.881732 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a893348-52b9-474c-8031-e67020c445eb" containerName="nova-scheduler-scheduler" Feb 17 13:26:05 crc kubenswrapper[4955]: I0217 13:26:05.882047 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a893348-52b9-474c-8031-e67020c445eb" containerName="nova-scheduler-scheduler" Feb 17 13:26:05 crc kubenswrapper[4955]: I0217 13:26:05.882853 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 17 13:26:05 crc kubenswrapper[4955]: I0217 13:26:05.885897 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 17 13:26:05 crc kubenswrapper[4955]: I0217 13:26:05.888751 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 17 13:26:05 crc kubenswrapper[4955]: I0217 13:26:05.899135 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca665950-2b74-4ac5-8d48-eacdac1f65ae-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ca665950-2b74-4ac5-8d48-eacdac1f65ae\") " pod="openstack/nova-scheduler-0" Feb 17 13:26:05 crc kubenswrapper[4955]: I0217 13:26:05.899238 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ss7v\" (UniqueName: \"kubernetes.io/projected/ca665950-2b74-4ac5-8d48-eacdac1f65ae-kube-api-access-7ss7v\") pod \"nova-scheduler-0\" (UID: \"ca665950-2b74-4ac5-8d48-eacdac1f65ae\") " pod="openstack/nova-scheduler-0" Feb 17 13:26:05 crc kubenswrapper[4955]: I0217 13:26:05.899410 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca665950-2b74-4ac5-8d48-eacdac1f65ae-config-data\") pod \"nova-scheduler-0\" (UID: \"ca665950-2b74-4ac5-8d48-eacdac1f65ae\") " pod="openstack/nova-scheduler-0" Feb 17 13:26:06 crc kubenswrapper[4955]: I0217 13:26:06.001169 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca665950-2b74-4ac5-8d48-eacdac1f65ae-config-data\") pod \"nova-scheduler-0\" (UID: \"ca665950-2b74-4ac5-8d48-eacdac1f65ae\") " pod="openstack/nova-scheduler-0" Feb 17 13:26:06 crc kubenswrapper[4955]: I0217 13:26:06.001812 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca665950-2b74-4ac5-8d48-eacdac1f65ae-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ca665950-2b74-4ac5-8d48-eacdac1f65ae\") " pod="openstack/nova-scheduler-0" Feb 17 13:26:06 crc kubenswrapper[4955]: I0217 13:26:06.001871 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ss7v\" (UniqueName: \"kubernetes.io/projected/ca665950-2b74-4ac5-8d48-eacdac1f65ae-kube-api-access-7ss7v\") pod \"nova-scheduler-0\" (UID: \"ca665950-2b74-4ac5-8d48-eacdac1f65ae\") " pod="openstack/nova-scheduler-0" Feb 17 13:26:06 crc kubenswrapper[4955]: I0217 13:26:06.008973 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca665950-2b74-4ac5-8d48-eacdac1f65ae-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ca665950-2b74-4ac5-8d48-eacdac1f65ae\") " pod="openstack/nova-scheduler-0" Feb 17 13:26:06 crc kubenswrapper[4955]: I0217 13:26:06.009042 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca665950-2b74-4ac5-8d48-eacdac1f65ae-config-data\") pod \"nova-scheduler-0\" (UID: \"ca665950-2b74-4ac5-8d48-eacdac1f65ae\") " pod="openstack/nova-scheduler-0" Feb 17 13:26:06 crc kubenswrapper[4955]: I0217 13:26:06.017776 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ss7v\" (UniqueName: \"kubernetes.io/projected/ca665950-2b74-4ac5-8d48-eacdac1f65ae-kube-api-access-7ss7v\") pod \"nova-scheduler-0\" (UID: \"ca665950-2b74-4ac5-8d48-eacdac1f65ae\") " pod="openstack/nova-scheduler-0" Feb 17 13:26:06 crc kubenswrapper[4955]: I0217 13:26:06.205476 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 17 13:26:06 crc kubenswrapper[4955]: I0217 13:26:06.235811 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a893348-52b9-474c-8031-e67020c445eb" path="/var/lib/kubelet/pods/7a893348-52b9-474c-8031-e67020c445eb/volumes" Feb 17 13:26:06 crc kubenswrapper[4955]: W0217 13:26:06.649232 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca665950_2b74_4ac5_8d48_eacdac1f65ae.slice/crio-dbac1080f2c9d9cb5f345e89c8f267ee8236812afb493afc6f73fa91de772845 WatchSource:0}: Error finding container dbac1080f2c9d9cb5f345e89c8f267ee8236812afb493afc6f73fa91de772845: Status 404 returned error can't find the container with id dbac1080f2c9d9cb5f345e89c8f267ee8236812afb493afc6f73fa91de772845 Feb 17 13:26:06 crc kubenswrapper[4955]: I0217 13:26:06.651331 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 17 13:26:06 crc kubenswrapper[4955]: I0217 13:26:06.815881 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ca665950-2b74-4ac5-8d48-eacdac1f65ae","Type":"ContainerStarted","Data":"dbac1080f2c9d9cb5f345e89c8f267ee8236812afb493afc6f73fa91de772845"} Feb 17 13:26:07 crc kubenswrapper[4955]: I0217 13:26:07.832663 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ca665950-2b74-4ac5-8d48-eacdac1f65ae","Type":"ContainerStarted","Data":"803c690b07e9116807b060610813f81f214d2591b0e80e482ba07bb5a379beeb"} Feb 17 13:26:07 crc kubenswrapper[4955]: I0217 13:26:07.853622 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.853601339 podStartE2EDuration="2.853601339s" podCreationTimestamp="2026-02-17 13:26:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:26:07.851946702 +0000 UTC m=+1306.374676265" watchObservedRunningTime="2026-02-17 13:26:07.853601339 +0000 UTC m=+1306.376330872" Feb 17 13:26:09 crc kubenswrapper[4955]: I0217 13:26:09.179225 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 17 13:26:09 crc kubenswrapper[4955]: I0217 13:26:09.179501 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 17 13:26:11 crc kubenswrapper[4955]: I0217 13:26:11.130830 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 17 13:26:11 crc kubenswrapper[4955]: I0217 13:26:11.131274 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 17 13:26:11 crc kubenswrapper[4955]: I0217 13:26:11.206033 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 17 13:26:12 crc kubenswrapper[4955]: I0217 13:26:12.140977 4955 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a1a37e38-9f12-49ad-8464-f744f7f5fc2f" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.208:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 17 13:26:12 crc kubenswrapper[4955]: I0217 13:26:12.141052 4955 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a1a37e38-9f12-49ad-8464-f744f7f5fc2f" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.208:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 17 13:26:14 crc kubenswrapper[4955]: I0217 13:26:14.180053 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 17 13:26:14 crc kubenswrapper[4955]: I0217 13:26:14.181522 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 17 13:26:15 crc kubenswrapper[4955]: I0217 13:26:15.193922 4955 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="ff800864-d970-46b0-bb2e-eb213d423da4" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.209:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 17 13:26:15 crc kubenswrapper[4955]: I0217 13:26:15.193965 4955 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="ff800864-d970-46b0-bb2e-eb213d423da4" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.209:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 17 13:26:16 crc kubenswrapper[4955]: I0217 13:26:16.205725 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 17 13:26:16 crc kubenswrapper[4955]: I0217 13:26:16.239461 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 17 13:26:16 crc kubenswrapper[4955]: I0217 13:26:16.953433 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 17 13:26:19 crc kubenswrapper[4955]: I0217 13:26:19.286500 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 17 13:26:21 crc kubenswrapper[4955]: I0217 13:26:21.138053 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 17 13:26:21 crc kubenswrapper[4955]: I0217 13:26:21.138716 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 17 13:26:21 crc kubenswrapper[4955]: I0217 13:26:21.139763 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 17 13:26:21 crc kubenswrapper[4955]: I0217 13:26:21.146450 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 17 13:26:21 crc kubenswrapper[4955]: I0217 13:26:21.964377 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 17 13:26:21 crc kubenswrapper[4955]: I0217 13:26:21.970255 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 17 13:26:24 crc kubenswrapper[4955]: I0217 13:26:24.296638 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 17 13:26:24 crc kubenswrapper[4955]: I0217 13:26:24.300439 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 17 13:26:24 crc kubenswrapper[4955]: I0217 13:26:24.311280 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 17 13:26:24 crc kubenswrapper[4955]: I0217 13:26:24.995759 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 17 13:26:32 crc kubenswrapper[4955]: I0217 13:26:32.489358 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 17 13:26:33 crc kubenswrapper[4955]: I0217 13:26:33.600117 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 17 13:26:34 crc kubenswrapper[4955]: I0217 13:26:34.574743 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 13:26:34 crc kubenswrapper[4955]: I0217 13:26:34.574833 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 13:26:37 crc kubenswrapper[4955]: I0217 13:26:37.605968 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="34552e31-68c3-4156-addf-37ea6fdf0b6c" containerName="rabbitmq" containerID="cri-o://f07eeaceb5a628e2e825faa765c2135da249f0bd9b80274d2fdeafb27d148d57" gracePeriod=604795 Feb 17 13:26:37 crc kubenswrapper[4955]: I0217 13:26:37.915048 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="c241677c-9f24-4cc0-8d15-e166174e4cd0" containerName="rabbitmq" containerID="cri-o://30e8197d8a6e3cdbe4f6414abf7e655a8860a1ac94d7c201ed6d3f58a7f1ecf2" gracePeriod=604796 Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.189341 4955 generic.go:334] "Generic (PLEG): container finished" podID="34552e31-68c3-4156-addf-37ea6fdf0b6c" containerID="f07eeaceb5a628e2e825faa765c2135da249f0bd9b80274d2fdeafb27d148d57" exitCode=0 Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.189424 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"34552e31-68c3-4156-addf-37ea6fdf0b6c","Type":"ContainerDied","Data":"f07eeaceb5a628e2e825faa765c2135da249f0bd9b80274d2fdeafb27d148d57"} Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.192372 4955 generic.go:334] "Generic (PLEG): container finished" podID="c241677c-9f24-4cc0-8d15-e166174e4cd0" containerID="30e8197d8a6e3cdbe4f6414abf7e655a8860a1ac94d7c201ed6d3f58a7f1ecf2" exitCode=0 Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.192423 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c241677c-9f24-4cc0-8d15-e166174e4cd0","Type":"ContainerDied","Data":"30e8197d8a6e3cdbe4f6414abf7e655a8860a1ac94d7c201ed6d3f58a7f1ecf2"} Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.298514 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.462515 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/34552e31-68c3-4156-addf-37ea6fdf0b6c-rabbitmq-erlang-cookie\") pod \"34552e31-68c3-4156-addf-37ea6fdf0b6c\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.462571 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/34552e31-68c3-4156-addf-37ea6fdf0b6c-rabbitmq-confd\") pod \"34552e31-68c3-4156-addf-37ea6fdf0b6c\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.462629 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/34552e31-68c3-4156-addf-37ea6fdf0b6c-rabbitmq-plugins\") pod \"34552e31-68c3-4156-addf-37ea6fdf0b6c\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.462683 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/34552e31-68c3-4156-addf-37ea6fdf0b6c-pod-info\") pod \"34552e31-68c3-4156-addf-37ea6fdf0b6c\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.462725 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/34552e31-68c3-4156-addf-37ea6fdf0b6c-rabbitmq-tls\") pod \"34552e31-68c3-4156-addf-37ea6fdf0b6c\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.462837 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/34552e31-68c3-4156-addf-37ea6fdf0b6c-server-conf\") pod \"34552e31-68c3-4156-addf-37ea6fdf0b6c\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.462873 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/34552e31-68c3-4156-addf-37ea6fdf0b6c-plugins-conf\") pod \"34552e31-68c3-4156-addf-37ea6fdf0b6c\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.462905 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"34552e31-68c3-4156-addf-37ea6fdf0b6c\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.462926 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/34552e31-68c3-4156-addf-37ea6fdf0b6c-erlang-cookie-secret\") pod \"34552e31-68c3-4156-addf-37ea6fdf0b6c\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.462956 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nbxhr\" (UniqueName: \"kubernetes.io/projected/34552e31-68c3-4156-addf-37ea6fdf0b6c-kube-api-access-nbxhr\") pod \"34552e31-68c3-4156-addf-37ea6fdf0b6c\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.462988 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/34552e31-68c3-4156-addf-37ea6fdf0b6c-config-data\") pod \"34552e31-68c3-4156-addf-37ea6fdf0b6c\" (UID: \"34552e31-68c3-4156-addf-37ea6fdf0b6c\") " Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.473149 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34552e31-68c3-4156-addf-37ea6fdf0b6c-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "34552e31-68c3-4156-addf-37ea6fdf0b6c" (UID: "34552e31-68c3-4156-addf-37ea6fdf0b6c"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.473367 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "persistence") pod "34552e31-68c3-4156-addf-37ea6fdf0b6c" (UID: "34552e31-68c3-4156-addf-37ea6fdf0b6c"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.477181 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34552e31-68c3-4156-addf-37ea6fdf0b6c-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "34552e31-68c3-4156-addf-37ea6fdf0b6c" (UID: "34552e31-68c3-4156-addf-37ea6fdf0b6c"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.477232 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34552e31-68c3-4156-addf-37ea6fdf0b6c-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "34552e31-68c3-4156-addf-37ea6fdf0b6c" (UID: "34552e31-68c3-4156-addf-37ea6fdf0b6c"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.480534 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34552e31-68c3-4156-addf-37ea6fdf0b6c-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "34552e31-68c3-4156-addf-37ea6fdf0b6c" (UID: "34552e31-68c3-4156-addf-37ea6fdf0b6c"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.480562 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34552e31-68c3-4156-addf-37ea6fdf0b6c-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "34552e31-68c3-4156-addf-37ea6fdf0b6c" (UID: "34552e31-68c3-4156-addf-37ea6fdf0b6c"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.483674 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/34552e31-68c3-4156-addf-37ea6fdf0b6c-pod-info" (OuterVolumeSpecName: "pod-info") pod "34552e31-68c3-4156-addf-37ea6fdf0b6c" (UID: "34552e31-68c3-4156-addf-37ea6fdf0b6c"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.490135 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34552e31-68c3-4156-addf-37ea6fdf0b6c-kube-api-access-nbxhr" (OuterVolumeSpecName: "kube-api-access-nbxhr") pod "34552e31-68c3-4156-addf-37ea6fdf0b6c" (UID: "34552e31-68c3-4156-addf-37ea6fdf0b6c"). InnerVolumeSpecName "kube-api-access-nbxhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.566797 4955 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/34552e31-68c3-4156-addf-37ea6fdf0b6c-plugins-conf\") on node \"crc\" DevicePath \"\"" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.566847 4955 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.566864 4955 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/34552e31-68c3-4156-addf-37ea6fdf0b6c-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.566879 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nbxhr\" (UniqueName: \"kubernetes.io/projected/34552e31-68c3-4156-addf-37ea6fdf0b6c-kube-api-access-nbxhr\") on node \"crc\" DevicePath \"\"" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.566896 4955 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/34552e31-68c3-4156-addf-37ea6fdf0b6c-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.566909 4955 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/34552e31-68c3-4156-addf-37ea6fdf0b6c-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.566921 4955 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/34552e31-68c3-4156-addf-37ea6fdf0b6c-pod-info\") on node \"crc\" DevicePath \"\"" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.566940 4955 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/34552e31-68c3-4156-addf-37ea6fdf0b6c-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.628493 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34552e31-68c3-4156-addf-37ea6fdf0b6c-server-conf" (OuterVolumeSpecName: "server-conf") pod "34552e31-68c3-4156-addf-37ea6fdf0b6c" (UID: "34552e31-68c3-4156-addf-37ea6fdf0b6c"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.642538 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34552e31-68c3-4156-addf-37ea6fdf0b6c-config-data" (OuterVolumeSpecName: "config-data") pod "34552e31-68c3-4156-addf-37ea6fdf0b6c" (UID: "34552e31-68c3-4156-addf-37ea6fdf0b6c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.645839 4955 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.668076 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.669042 4955 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/34552e31-68c3-4156-addf-37ea6fdf0b6c-server-conf\") on node \"crc\" DevicePath \"\"" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.669063 4955 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.669072 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/34552e31-68c3-4156-addf-37ea6fdf0b6c-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.752859 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34552e31-68c3-4156-addf-37ea6fdf0b6c-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "34552e31-68c3-4156-addf-37ea6fdf0b6c" (UID: "34552e31-68c3-4156-addf-37ea6fdf0b6c"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.770382 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c241677c-9f24-4cc0-8d15-e166174e4cd0-rabbitmq-confd\") pod \"c241677c-9f24-4cc0-8d15-e166174e4cd0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.770438 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c241677c-9f24-4cc0-8d15-e166174e4cd0-erlang-cookie-secret\") pod \"c241677c-9f24-4cc0-8d15-e166174e4cd0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.770500 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c241677c-9f24-4cc0-8d15-e166174e4cd0-config-data\") pod \"c241677c-9f24-4cc0-8d15-e166174e4cd0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.770592 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"c241677c-9f24-4cc0-8d15-e166174e4cd0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.770707 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c241677c-9f24-4cc0-8d15-e166174e4cd0-rabbitmq-tls\") pod \"c241677c-9f24-4cc0-8d15-e166174e4cd0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.770744 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c241677c-9f24-4cc0-8d15-e166174e4cd0-plugins-conf\") pod \"c241677c-9f24-4cc0-8d15-e166174e4cd0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.770922 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c241677c-9f24-4cc0-8d15-e166174e4cd0-pod-info\") pod \"c241677c-9f24-4cc0-8d15-e166174e4cd0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.770963 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c241677c-9f24-4cc0-8d15-e166174e4cd0-rabbitmq-plugins\") pod \"c241677c-9f24-4cc0-8d15-e166174e4cd0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.771017 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c241677c-9f24-4cc0-8d15-e166174e4cd0-server-conf\") pod \"c241677c-9f24-4cc0-8d15-e166174e4cd0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.771047 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c241677c-9f24-4cc0-8d15-e166174e4cd0-rabbitmq-erlang-cookie\") pod \"c241677c-9f24-4cc0-8d15-e166174e4cd0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.771072 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5sdlw\" (UniqueName: \"kubernetes.io/projected/c241677c-9f24-4cc0-8d15-e166174e4cd0-kube-api-access-5sdlw\") pod \"c241677c-9f24-4cc0-8d15-e166174e4cd0\" (UID: \"c241677c-9f24-4cc0-8d15-e166174e4cd0\") " Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.771525 4955 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/34552e31-68c3-4156-addf-37ea6fdf0b6c-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.779077 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c241677c-9f24-4cc0-8d15-e166174e4cd0-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "c241677c-9f24-4cc0-8d15-e166174e4cd0" (UID: "c241677c-9f24-4cc0-8d15-e166174e4cd0"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.781601 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c241677c-9f24-4cc0-8d15-e166174e4cd0-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "c241677c-9f24-4cc0-8d15-e166174e4cd0" (UID: "c241677c-9f24-4cc0-8d15-e166174e4cd0"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.783753 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c241677c-9f24-4cc0-8d15-e166174e4cd0-kube-api-access-5sdlw" (OuterVolumeSpecName: "kube-api-access-5sdlw") pod "c241677c-9f24-4cc0-8d15-e166174e4cd0" (UID: "c241677c-9f24-4cc0-8d15-e166174e4cd0"). InnerVolumeSpecName "kube-api-access-5sdlw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.784880 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c241677c-9f24-4cc0-8d15-e166174e4cd0-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "c241677c-9f24-4cc0-8d15-e166174e4cd0" (UID: "c241677c-9f24-4cc0-8d15-e166174e4cd0"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.787232 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c241677c-9f24-4cc0-8d15-e166174e4cd0-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "c241677c-9f24-4cc0-8d15-e166174e4cd0" (UID: "c241677c-9f24-4cc0-8d15-e166174e4cd0"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.791167 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c241677c-9f24-4cc0-8d15-e166174e4cd0-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "c241677c-9f24-4cc0-8d15-e166174e4cd0" (UID: "c241677c-9f24-4cc0-8d15-e166174e4cd0"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.791483 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/c241677c-9f24-4cc0-8d15-e166174e4cd0-pod-info" (OuterVolumeSpecName: "pod-info") pod "c241677c-9f24-4cc0-8d15-e166174e4cd0" (UID: "c241677c-9f24-4cc0-8d15-e166174e4cd0"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.797836 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "persistence") pod "c241677c-9f24-4cc0-8d15-e166174e4cd0" (UID: "c241677c-9f24-4cc0-8d15-e166174e4cd0"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.816827 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c241677c-9f24-4cc0-8d15-e166174e4cd0-config-data" (OuterVolumeSpecName: "config-data") pod "c241677c-9f24-4cc0-8d15-e166174e4cd0" (UID: "c241677c-9f24-4cc0-8d15-e166174e4cd0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.869739 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c241677c-9f24-4cc0-8d15-e166174e4cd0-server-conf" (OuterVolumeSpecName: "server-conf") pod "c241677c-9f24-4cc0-8d15-e166174e4cd0" (UID: "c241677c-9f24-4cc0-8d15-e166174e4cd0"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.873084 4955 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c241677c-9f24-4cc0-8d15-e166174e4cd0-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.873114 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c241677c-9f24-4cc0-8d15-e166174e4cd0-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.873144 4955 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.873154 4955 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c241677c-9f24-4cc0-8d15-e166174e4cd0-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.873164 4955 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c241677c-9f24-4cc0-8d15-e166174e4cd0-plugins-conf\") on node \"crc\" DevicePath \"\"" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.873172 4955 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c241677c-9f24-4cc0-8d15-e166174e4cd0-pod-info\") on node \"crc\" DevicePath \"\"" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.873180 4955 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c241677c-9f24-4cc0-8d15-e166174e4cd0-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.873188 4955 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c241677c-9f24-4cc0-8d15-e166174e4cd0-server-conf\") on node \"crc\" DevicePath \"\"" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.873197 4955 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c241677c-9f24-4cc0-8d15-e166174e4cd0-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.873205 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5sdlw\" (UniqueName: \"kubernetes.io/projected/c241677c-9f24-4cc0-8d15-e166174e4cd0-kube-api-access-5sdlw\") on node \"crc\" DevicePath \"\"" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.909476 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c241677c-9f24-4cc0-8d15-e166174e4cd0-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "c241677c-9f24-4cc0-8d15-e166174e4cd0" (UID: "c241677c-9f24-4cc0-8d15-e166174e4cd0"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.914271 4955 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.975264 4955 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Feb 17 13:26:44 crc kubenswrapper[4955]: I0217 13:26:44.975302 4955 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c241677c-9f24-4cc0-8d15-e166174e4cd0-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.202574 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c241677c-9f24-4cc0-8d15-e166174e4cd0","Type":"ContainerDied","Data":"4513e349aa0229a921459de11450edded87f95558f22921e8e613e999f0e2cf4"} Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.202643 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.202654 4955 scope.go:117] "RemoveContainer" containerID="30e8197d8a6e3cdbe4f6414abf7e655a8860a1ac94d7c201ed6d3f58a7f1ecf2" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.205148 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"34552e31-68c3-4156-addf-37ea6fdf0b6c","Type":"ContainerDied","Data":"d03b368ffd4e5b84e52261415ca43843774000447301c378956fb0f752fff984"} Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.205216 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.229349 4955 scope.go:117] "RemoveContainer" containerID="cd19a92beba58ac2a112c23c7364fb146fdc4af6d6854544bdce5d7e774494a2" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.264833 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.280650 4955 scope.go:117] "RemoveContainer" containerID="f07eeaceb5a628e2e825faa765c2135da249f0bd9b80274d2fdeafb27d148d57" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.280855 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.294625 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.322861 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.332294 4955 scope.go:117] "RemoveContainer" containerID="a8a6c6d9a2920a8907fcd1815d0cabffa7db57ffe67f73f31657d130b21005d5" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.358245 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 17 13:26:45 crc kubenswrapper[4955]: E0217 13:26:45.359312 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c241677c-9f24-4cc0-8d15-e166174e4cd0" containerName="setup-container" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.359340 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="c241677c-9f24-4cc0-8d15-e166174e4cd0" containerName="setup-container" Feb 17 13:26:45 crc kubenswrapper[4955]: E0217 13:26:45.359379 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34552e31-68c3-4156-addf-37ea6fdf0b6c" containerName="setup-container" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.359387 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="34552e31-68c3-4156-addf-37ea6fdf0b6c" containerName="setup-container" Feb 17 13:26:45 crc kubenswrapper[4955]: E0217 13:26:45.359445 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c241677c-9f24-4cc0-8d15-e166174e4cd0" containerName="rabbitmq" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.359455 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="c241677c-9f24-4cc0-8d15-e166174e4cd0" containerName="rabbitmq" Feb 17 13:26:45 crc kubenswrapper[4955]: E0217 13:26:45.359494 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34552e31-68c3-4156-addf-37ea6fdf0b6c" containerName="rabbitmq" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.359507 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="34552e31-68c3-4156-addf-37ea6fdf0b6c" containerName="rabbitmq" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.359973 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="c241677c-9f24-4cc0-8d15-e166174e4cd0" containerName="rabbitmq" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.359994 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="34552e31-68c3-4156-addf-37ea6fdf0b6c" containerName="rabbitmq" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.362026 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.365247 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.367308 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.367478 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.367522 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.367482 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.369361 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.374022 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-86vw7" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.415817 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.419217 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.423098 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-nt4nx" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.423508 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.424123 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.424385 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.424622 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.424828 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.425175 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.431675 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.446240 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.508334 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cd70e7eb-fb12-4b99-ab83-67d4abb9392b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"cd70e7eb-fb12-4b99-ab83-67d4abb9392b\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.508426 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cd70e7eb-fb12-4b99-ab83-67d4abb9392b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"cd70e7eb-fb12-4b99-ab83-67d4abb9392b\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.508464 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cd70e7eb-fb12-4b99-ab83-67d4abb9392b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"cd70e7eb-fb12-4b99-ab83-67d4abb9392b\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.508489 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cd70e7eb-fb12-4b99-ab83-67d4abb9392b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"cd70e7eb-fb12-4b99-ab83-67d4abb9392b\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.508519 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"cd70e7eb-fb12-4b99-ab83-67d4abb9392b\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.508543 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cd70e7eb-fb12-4b99-ab83-67d4abb9392b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cd70e7eb-fb12-4b99-ab83-67d4abb9392b\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.508564 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cd70e7eb-fb12-4b99-ab83-67d4abb9392b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cd70e7eb-fb12-4b99-ab83-67d4abb9392b\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.508616 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cd70e7eb-fb12-4b99-ab83-67d4abb9392b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"cd70e7eb-fb12-4b99-ab83-67d4abb9392b\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.508651 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cd70e7eb-fb12-4b99-ab83-67d4abb9392b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"cd70e7eb-fb12-4b99-ab83-67d4abb9392b\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.508683 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cd70e7eb-fb12-4b99-ab83-67d4abb9392b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"cd70e7eb-fb12-4b99-ab83-67d4abb9392b\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.508751 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7p97\" (UniqueName: \"kubernetes.io/projected/cd70e7eb-fb12-4b99-ab83-67d4abb9392b-kube-api-access-j7p97\") pod \"rabbitmq-cell1-server-0\" (UID: \"cd70e7eb-fb12-4b99-ab83-67d4abb9392b\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.610485 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c8c91914-fefa-479e-b497-de0f1f2978a2-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"c8c91914-fefa-479e-b497-de0f1f2978a2\") " pod="openstack/rabbitmq-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.610547 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7p97\" (UniqueName: \"kubernetes.io/projected/cd70e7eb-fb12-4b99-ab83-67d4abb9392b-kube-api-access-j7p97\") pod \"rabbitmq-cell1-server-0\" (UID: \"cd70e7eb-fb12-4b99-ab83-67d4abb9392b\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.610566 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c8c91914-fefa-479e-b497-de0f1f2978a2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c8c91914-fefa-479e-b497-de0f1f2978a2\") " pod="openstack/rabbitmq-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.610588 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c8c91914-fefa-479e-b497-de0f1f2978a2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c8c91914-fefa-479e-b497-de0f1f2978a2\") " pod="openstack/rabbitmq-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.610835 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t65j6\" (UniqueName: \"kubernetes.io/projected/c8c91914-fefa-479e-b497-de0f1f2978a2-kube-api-access-t65j6\") pod \"rabbitmq-server-0\" (UID: \"c8c91914-fefa-479e-b497-de0f1f2978a2\") " pod="openstack/rabbitmq-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.610891 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c8c91914-fefa-479e-b497-de0f1f2978a2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c8c91914-fefa-479e-b497-de0f1f2978a2\") " pod="openstack/rabbitmq-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.610968 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cd70e7eb-fb12-4b99-ab83-67d4abb9392b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"cd70e7eb-fb12-4b99-ab83-67d4abb9392b\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.611061 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cd70e7eb-fb12-4b99-ab83-67d4abb9392b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"cd70e7eb-fb12-4b99-ab83-67d4abb9392b\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.611078 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c8c91914-fefa-479e-b497-de0f1f2978a2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c8c91914-fefa-479e-b497-de0f1f2978a2\") " pod="openstack/rabbitmq-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.611117 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cd70e7eb-fb12-4b99-ab83-67d4abb9392b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"cd70e7eb-fb12-4b99-ab83-67d4abb9392b\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.611149 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cd70e7eb-fb12-4b99-ab83-67d4abb9392b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"cd70e7eb-fb12-4b99-ab83-67d4abb9392b\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.611242 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"cd70e7eb-fb12-4b99-ab83-67d4abb9392b\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.611289 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cd70e7eb-fb12-4b99-ab83-67d4abb9392b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cd70e7eb-fb12-4b99-ab83-67d4abb9392b\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.611319 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cd70e7eb-fb12-4b99-ab83-67d4abb9392b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cd70e7eb-fb12-4b99-ab83-67d4abb9392b\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.611381 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c8c91914-fefa-479e-b497-de0f1f2978a2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"c8c91914-fefa-479e-b497-de0f1f2978a2\") " pod="openstack/rabbitmq-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.611446 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c8c91914-fefa-479e-b497-de0f1f2978a2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c8c91914-fefa-479e-b497-de0f1f2978a2\") " pod="openstack/rabbitmq-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.611457 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cd70e7eb-fb12-4b99-ab83-67d4abb9392b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"cd70e7eb-fb12-4b99-ab83-67d4abb9392b\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.611560 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c8c91914-fefa-479e-b497-de0f1f2978a2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c8c91914-fefa-479e-b497-de0f1f2978a2\") " pod="openstack/rabbitmq-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.611617 4955 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"cd70e7eb-fb12-4b99-ab83-67d4abb9392b\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.611827 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cd70e7eb-fb12-4b99-ab83-67d4abb9392b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"cd70e7eb-fb12-4b99-ab83-67d4abb9392b\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.611625 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cd70e7eb-fb12-4b99-ab83-67d4abb9392b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"cd70e7eb-fb12-4b99-ab83-67d4abb9392b\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.611906 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"c8c91914-fefa-479e-b497-de0f1f2978a2\") " pod="openstack/rabbitmq-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.611942 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cd70e7eb-fb12-4b99-ab83-67d4abb9392b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"cd70e7eb-fb12-4b99-ab83-67d4abb9392b\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.611977 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cd70e7eb-fb12-4b99-ab83-67d4abb9392b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"cd70e7eb-fb12-4b99-ab83-67d4abb9392b\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.612011 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c8c91914-fefa-479e-b497-de0f1f2978a2-config-data\") pod \"rabbitmq-server-0\" (UID: \"c8c91914-fefa-479e-b497-de0f1f2978a2\") " pod="openstack/rabbitmq-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.612419 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cd70e7eb-fb12-4b99-ab83-67d4abb9392b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cd70e7eb-fb12-4b99-ab83-67d4abb9392b\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.612451 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cd70e7eb-fb12-4b99-ab83-67d4abb9392b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"cd70e7eb-fb12-4b99-ab83-67d4abb9392b\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.612712 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cd70e7eb-fb12-4b99-ab83-67d4abb9392b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cd70e7eb-fb12-4b99-ab83-67d4abb9392b\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.615960 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cd70e7eb-fb12-4b99-ab83-67d4abb9392b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"cd70e7eb-fb12-4b99-ab83-67d4abb9392b\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.616512 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cd70e7eb-fb12-4b99-ab83-67d4abb9392b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"cd70e7eb-fb12-4b99-ab83-67d4abb9392b\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.616821 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cd70e7eb-fb12-4b99-ab83-67d4abb9392b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"cd70e7eb-fb12-4b99-ab83-67d4abb9392b\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.617493 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cd70e7eb-fb12-4b99-ab83-67d4abb9392b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"cd70e7eb-fb12-4b99-ab83-67d4abb9392b\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.652147 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"cd70e7eb-fb12-4b99-ab83-67d4abb9392b\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.654065 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7p97\" (UniqueName: \"kubernetes.io/projected/cd70e7eb-fb12-4b99-ab83-67d4abb9392b-kube-api-access-j7p97\") pod \"rabbitmq-cell1-server-0\" (UID: \"cd70e7eb-fb12-4b99-ab83-67d4abb9392b\") " pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.713807 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c8c91914-fefa-479e-b497-de0f1f2978a2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c8c91914-fefa-479e-b497-de0f1f2978a2\") " pod="openstack/rabbitmq-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.713901 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c8c91914-fefa-479e-b497-de0f1f2978a2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"c8c91914-fefa-479e-b497-de0f1f2978a2\") " pod="openstack/rabbitmq-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.713930 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c8c91914-fefa-479e-b497-de0f1f2978a2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c8c91914-fefa-479e-b497-de0f1f2978a2\") " pod="openstack/rabbitmq-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.713964 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c8c91914-fefa-479e-b497-de0f1f2978a2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c8c91914-fefa-479e-b497-de0f1f2978a2\") " pod="openstack/rabbitmq-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.714002 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"c8c91914-fefa-479e-b497-de0f1f2978a2\") " pod="openstack/rabbitmq-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.714047 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c8c91914-fefa-479e-b497-de0f1f2978a2-config-data\") pod \"rabbitmq-server-0\" (UID: \"c8c91914-fefa-479e-b497-de0f1f2978a2\") " pod="openstack/rabbitmq-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.714086 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c8c91914-fefa-479e-b497-de0f1f2978a2-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"c8c91914-fefa-479e-b497-de0f1f2978a2\") " pod="openstack/rabbitmq-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.714118 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c8c91914-fefa-479e-b497-de0f1f2978a2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c8c91914-fefa-479e-b497-de0f1f2978a2\") " pod="openstack/rabbitmq-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.714136 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c8c91914-fefa-479e-b497-de0f1f2978a2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c8c91914-fefa-479e-b497-de0f1f2978a2\") " pod="openstack/rabbitmq-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.714163 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t65j6\" (UniqueName: \"kubernetes.io/projected/c8c91914-fefa-479e-b497-de0f1f2978a2-kube-api-access-t65j6\") pod \"rabbitmq-server-0\" (UID: \"c8c91914-fefa-479e-b497-de0f1f2978a2\") " pod="openstack/rabbitmq-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.714178 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c8c91914-fefa-479e-b497-de0f1f2978a2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c8c91914-fefa-479e-b497-de0f1f2978a2\") " pod="openstack/rabbitmq-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.714960 4955 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"c8c91914-fefa-479e-b497-de0f1f2978a2\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.715104 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c8c91914-fefa-479e-b497-de0f1f2978a2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c8c91914-fefa-479e-b497-de0f1f2978a2\") " pod="openstack/rabbitmq-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.715509 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c8c91914-fefa-479e-b497-de0f1f2978a2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"c8c91914-fefa-479e-b497-de0f1f2978a2\") " pod="openstack/rabbitmq-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.715668 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c8c91914-fefa-479e-b497-de0f1f2978a2-config-data\") pod \"rabbitmq-server-0\" (UID: \"c8c91914-fefa-479e-b497-de0f1f2978a2\") " pod="openstack/rabbitmq-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.715725 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c8c91914-fefa-479e-b497-de0f1f2978a2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c8c91914-fefa-479e-b497-de0f1f2978a2\") " pod="openstack/rabbitmq-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.717166 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c8c91914-fefa-479e-b497-de0f1f2978a2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c8c91914-fefa-479e-b497-de0f1f2978a2\") " pod="openstack/rabbitmq-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.718404 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c8c91914-fefa-479e-b497-de0f1f2978a2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c8c91914-fefa-479e-b497-de0f1f2978a2\") " pod="openstack/rabbitmq-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.718605 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c8c91914-fefa-479e-b497-de0f1f2978a2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c8c91914-fefa-479e-b497-de0f1f2978a2\") " pod="openstack/rabbitmq-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.718686 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c8c91914-fefa-479e-b497-de0f1f2978a2-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"c8c91914-fefa-479e-b497-de0f1f2978a2\") " pod="openstack/rabbitmq-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.719119 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c8c91914-fefa-479e-b497-de0f1f2978a2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c8c91914-fefa-479e-b497-de0f1f2978a2\") " pod="openstack/rabbitmq-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.731002 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.733472 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t65j6\" (UniqueName: \"kubernetes.io/projected/c8c91914-fefa-479e-b497-de0f1f2978a2-kube-api-access-t65j6\") pod \"rabbitmq-server-0\" (UID: \"c8c91914-fefa-479e-b497-de0f1f2978a2\") " pod="openstack/rabbitmq-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.753153 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"c8c91914-fefa-479e-b497-de0f1f2978a2\") " pod="openstack/rabbitmq-server-0" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.980283 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-d558885bc-mqt6k"] Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.981839 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-mqt6k" Feb 17 13:26:45 crc kubenswrapper[4955]: I0217 13:26:45.985955 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Feb 17 13:26:46 crc kubenswrapper[4955]: I0217 13:26:46.012426 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-mqt6k"] Feb 17 13:26:46 crc kubenswrapper[4955]: I0217 13:26:46.054983 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 17 13:26:46 crc kubenswrapper[4955]: I0217 13:26:46.121443 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c31031a2-1485-4507-8a90-d06913010865-dns-svc\") pod \"dnsmasq-dns-d558885bc-mqt6k\" (UID: \"c31031a2-1485-4507-8a90-d06913010865\") " pod="openstack/dnsmasq-dns-d558885bc-mqt6k" Feb 17 13:26:46 crc kubenswrapper[4955]: I0217 13:26:46.121501 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c31031a2-1485-4507-8a90-d06913010865-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-mqt6k\" (UID: \"c31031a2-1485-4507-8a90-d06913010865\") " pod="openstack/dnsmasq-dns-d558885bc-mqt6k" Feb 17 13:26:46 crc kubenswrapper[4955]: I0217 13:26:46.121651 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c31031a2-1485-4507-8a90-d06913010865-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-mqt6k\" (UID: \"c31031a2-1485-4507-8a90-d06913010865\") " pod="openstack/dnsmasq-dns-d558885bc-mqt6k" Feb 17 13:26:46 crc kubenswrapper[4955]: I0217 13:26:46.121766 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/c31031a2-1485-4507-8a90-d06913010865-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-mqt6k\" (UID: \"c31031a2-1485-4507-8a90-d06913010865\") " pod="openstack/dnsmasq-dns-d558885bc-mqt6k" Feb 17 13:26:46 crc kubenswrapper[4955]: I0217 13:26:46.121900 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c31031a2-1485-4507-8a90-d06913010865-config\") pod \"dnsmasq-dns-d558885bc-mqt6k\" (UID: \"c31031a2-1485-4507-8a90-d06913010865\") " pod="openstack/dnsmasq-dns-d558885bc-mqt6k" Feb 17 13:26:46 crc kubenswrapper[4955]: I0217 13:26:46.121988 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c31031a2-1485-4507-8a90-d06913010865-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-mqt6k\" (UID: \"c31031a2-1485-4507-8a90-d06913010865\") " pod="openstack/dnsmasq-dns-d558885bc-mqt6k" Feb 17 13:26:46 crc kubenswrapper[4955]: I0217 13:26:46.122182 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jw7dh\" (UniqueName: \"kubernetes.io/projected/c31031a2-1485-4507-8a90-d06913010865-kube-api-access-jw7dh\") pod \"dnsmasq-dns-d558885bc-mqt6k\" (UID: \"c31031a2-1485-4507-8a90-d06913010865\") " pod="openstack/dnsmasq-dns-d558885bc-mqt6k" Feb 17 13:26:46 crc kubenswrapper[4955]: I0217 13:26:46.222115 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 17 13:26:46 crc kubenswrapper[4955]: I0217 13:26:46.223556 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c31031a2-1485-4507-8a90-d06913010865-dns-svc\") pod \"dnsmasq-dns-d558885bc-mqt6k\" (UID: \"c31031a2-1485-4507-8a90-d06913010865\") " pod="openstack/dnsmasq-dns-d558885bc-mqt6k" Feb 17 13:26:46 crc kubenswrapper[4955]: I0217 13:26:46.223616 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c31031a2-1485-4507-8a90-d06913010865-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-mqt6k\" (UID: \"c31031a2-1485-4507-8a90-d06913010865\") " pod="openstack/dnsmasq-dns-d558885bc-mqt6k" Feb 17 13:26:46 crc kubenswrapper[4955]: I0217 13:26:46.223681 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c31031a2-1485-4507-8a90-d06913010865-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-mqt6k\" (UID: \"c31031a2-1485-4507-8a90-d06913010865\") " pod="openstack/dnsmasq-dns-d558885bc-mqt6k" Feb 17 13:26:46 crc kubenswrapper[4955]: I0217 13:26:46.223724 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/c31031a2-1485-4507-8a90-d06913010865-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-mqt6k\" (UID: \"c31031a2-1485-4507-8a90-d06913010865\") " pod="openstack/dnsmasq-dns-d558885bc-mqt6k" Feb 17 13:26:46 crc kubenswrapper[4955]: I0217 13:26:46.223771 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c31031a2-1485-4507-8a90-d06913010865-config\") pod \"dnsmasq-dns-d558885bc-mqt6k\" (UID: \"c31031a2-1485-4507-8a90-d06913010865\") " pod="openstack/dnsmasq-dns-d558885bc-mqt6k" Feb 17 13:26:46 crc kubenswrapper[4955]: I0217 13:26:46.223835 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c31031a2-1485-4507-8a90-d06913010865-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-mqt6k\" (UID: \"c31031a2-1485-4507-8a90-d06913010865\") " pod="openstack/dnsmasq-dns-d558885bc-mqt6k" Feb 17 13:26:46 crc kubenswrapper[4955]: I0217 13:26:46.223916 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jw7dh\" (UniqueName: \"kubernetes.io/projected/c31031a2-1485-4507-8a90-d06913010865-kube-api-access-jw7dh\") pod \"dnsmasq-dns-d558885bc-mqt6k\" (UID: \"c31031a2-1485-4507-8a90-d06913010865\") " pod="openstack/dnsmasq-dns-d558885bc-mqt6k" Feb 17 13:26:46 crc kubenswrapper[4955]: I0217 13:26:46.227827 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c31031a2-1485-4507-8a90-d06913010865-dns-svc\") pod \"dnsmasq-dns-d558885bc-mqt6k\" (UID: \"c31031a2-1485-4507-8a90-d06913010865\") " pod="openstack/dnsmasq-dns-d558885bc-mqt6k" Feb 17 13:26:46 crc kubenswrapper[4955]: I0217 13:26:46.228545 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c31031a2-1485-4507-8a90-d06913010865-config\") pod \"dnsmasq-dns-d558885bc-mqt6k\" (UID: \"c31031a2-1485-4507-8a90-d06913010865\") " pod="openstack/dnsmasq-dns-d558885bc-mqt6k" Feb 17 13:26:46 crc kubenswrapper[4955]: I0217 13:26:46.229072 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c31031a2-1485-4507-8a90-d06913010865-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-mqt6k\" (UID: \"c31031a2-1485-4507-8a90-d06913010865\") " pod="openstack/dnsmasq-dns-d558885bc-mqt6k" Feb 17 13:26:46 crc kubenswrapper[4955]: I0217 13:26:46.229754 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c31031a2-1485-4507-8a90-d06913010865-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-mqt6k\" (UID: \"c31031a2-1485-4507-8a90-d06913010865\") " pod="openstack/dnsmasq-dns-d558885bc-mqt6k" Feb 17 13:26:46 crc kubenswrapper[4955]: I0217 13:26:46.230321 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c31031a2-1485-4507-8a90-d06913010865-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-mqt6k\" (UID: \"c31031a2-1485-4507-8a90-d06913010865\") " pod="openstack/dnsmasq-dns-d558885bc-mqt6k" Feb 17 13:26:46 crc kubenswrapper[4955]: I0217 13:26:46.243812 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/c31031a2-1485-4507-8a90-d06913010865-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-mqt6k\" (UID: \"c31031a2-1485-4507-8a90-d06913010865\") " pod="openstack/dnsmasq-dns-d558885bc-mqt6k" Feb 17 13:26:46 crc kubenswrapper[4955]: I0217 13:26:46.244960 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jw7dh\" (UniqueName: \"kubernetes.io/projected/c31031a2-1485-4507-8a90-d06913010865-kube-api-access-jw7dh\") pod \"dnsmasq-dns-d558885bc-mqt6k\" (UID: \"c31031a2-1485-4507-8a90-d06913010865\") " pod="openstack/dnsmasq-dns-d558885bc-mqt6k" Feb 17 13:26:46 crc kubenswrapper[4955]: I0217 13:26:46.273949 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34552e31-68c3-4156-addf-37ea6fdf0b6c" path="/var/lib/kubelet/pods/34552e31-68c3-4156-addf-37ea6fdf0b6c/volumes" Feb 17 13:26:46 crc kubenswrapper[4955]: I0217 13:26:46.275290 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c241677c-9f24-4cc0-8d15-e166174e4cd0" path="/var/lib/kubelet/pods/c241677c-9f24-4cc0-8d15-e166174e4cd0/volumes" Feb 17 13:26:46 crc kubenswrapper[4955]: I0217 13:26:46.319331 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-mqt6k" Feb 17 13:26:47 crc kubenswrapper[4955]: I0217 13:26:46.562988 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 17 13:26:47 crc kubenswrapper[4955]: W0217 13:26:46.567190 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc8c91914_fefa_479e_b497_de0f1f2978a2.slice/crio-e5271a6dddd48823782ab61a728a2c2b5e4b2cceffe91462be2dac4d29de7aa2 WatchSource:0}: Error finding container e5271a6dddd48823782ab61a728a2c2b5e4b2cceffe91462be2dac4d29de7aa2: Status 404 returned error can't find the container with id e5271a6dddd48823782ab61a728a2c2b5e4b2cceffe91462be2dac4d29de7aa2 Feb 17 13:26:47 crc kubenswrapper[4955]: I0217 13:26:47.271438 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"cd70e7eb-fb12-4b99-ab83-67d4abb9392b","Type":"ContainerStarted","Data":"3c285fe3ce21c57996919bba17b6084d994a2b1cf161ae53805b9cbb338f2306"} Feb 17 13:26:47 crc kubenswrapper[4955]: I0217 13:26:47.272839 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c8c91914-fefa-479e-b497-de0f1f2978a2","Type":"ContainerStarted","Data":"e5271a6dddd48823782ab61a728a2c2b5e4b2cceffe91462be2dac4d29de7aa2"} Feb 17 13:26:47 crc kubenswrapper[4955]: I0217 13:26:47.869406 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-mqt6k"] Feb 17 13:26:47 crc kubenswrapper[4955]: W0217 13:26:47.892138 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc31031a2_1485_4507_8a90_d06913010865.slice/crio-5ae0132e3fff7b2d7610acb3549b234d39d4781cda4c5118ffee1658689e8b7e WatchSource:0}: Error finding container 5ae0132e3fff7b2d7610acb3549b234d39d4781cda4c5118ffee1658689e8b7e: Status 404 returned error can't find the container with id 5ae0132e3fff7b2d7610acb3549b234d39d4781cda4c5118ffee1658689e8b7e Feb 17 13:26:48 crc kubenswrapper[4955]: I0217 13:26:48.281345 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-mqt6k" event={"ID":"c31031a2-1485-4507-8a90-d06913010865","Type":"ContainerStarted","Data":"5ae0132e3fff7b2d7610acb3549b234d39d4781cda4c5118ffee1658689e8b7e"} Feb 17 13:26:48 crc kubenswrapper[4955]: I0217 13:26:48.283148 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"cd70e7eb-fb12-4b99-ab83-67d4abb9392b","Type":"ContainerStarted","Data":"4493ad91867b6ad99901129eeaef053f033bb9685824695228eda14fa0ac124d"} Feb 17 13:26:49 crc kubenswrapper[4955]: I0217 13:26:49.295709 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c8c91914-fefa-479e-b497-de0f1f2978a2","Type":"ContainerStarted","Data":"dd90a1bfd0db84aa04a7f0fd77caaa9bb7fe5384e795186784679c7475b3abfd"} Feb 17 13:26:49 crc kubenswrapper[4955]: I0217 13:26:49.298373 4955 generic.go:334] "Generic (PLEG): container finished" podID="c31031a2-1485-4507-8a90-d06913010865" containerID="06a07877376e2fb02c7296215cfcdc234914b8b090638fa038a611fbdddd137f" exitCode=0 Feb 17 13:26:49 crc kubenswrapper[4955]: I0217 13:26:49.298472 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-mqt6k" event={"ID":"c31031a2-1485-4507-8a90-d06913010865","Type":"ContainerDied","Data":"06a07877376e2fb02c7296215cfcdc234914b8b090638fa038a611fbdddd137f"} Feb 17 13:26:50 crc kubenswrapper[4955]: I0217 13:26:50.307917 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-mqt6k" event={"ID":"c31031a2-1485-4507-8a90-d06913010865","Type":"ContainerStarted","Data":"75ea6a58f6a0018b69ba7b7aa87832742c54b55ee898a837323941bb6ed92575"} Feb 17 13:26:50 crc kubenswrapper[4955]: I0217 13:26:50.308323 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-d558885bc-mqt6k" Feb 17 13:26:50 crc kubenswrapper[4955]: I0217 13:26:50.332684 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-d558885bc-mqt6k" podStartSLOduration=5.332668204 podStartE2EDuration="5.332668204s" podCreationTimestamp="2026-02-17 13:26:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:26:50.331544002 +0000 UTC m=+1348.854273545" watchObservedRunningTime="2026-02-17 13:26:50.332668204 +0000 UTC m=+1348.855397747" Feb 17 13:26:56 crc kubenswrapper[4955]: I0217 13:26:56.321940 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-d558885bc-mqt6k" Feb 17 13:26:56 crc kubenswrapper[4955]: I0217 13:26:56.412042 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-9fmsq"] Feb 17 13:26:56 crc kubenswrapper[4955]: I0217 13:26:56.412256 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cd5cbd7b9-9fmsq" podUID="3ea144e9-32f5-4f5d-85fd-a6ec6e73add3" containerName="dnsmasq-dns" containerID="cri-o://8b42b497e8dd742535e0e73c9eb2c77ffc626ae2e0cc6c6f483d434219efb640" gracePeriod=10 Feb 17 13:26:56 crc kubenswrapper[4955]: I0217 13:26:56.691595 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78c64bc9c5-hktkl"] Feb 17 13:26:56 crc kubenswrapper[4955]: I0217 13:26:56.693878 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78c64bc9c5-hktkl" Feb 17 13:26:56 crc kubenswrapper[4955]: I0217 13:26:56.705044 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78c64bc9c5-hktkl"] Feb 17 13:26:56 crc kubenswrapper[4955]: I0217 13:26:56.820317 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d64b1ace-43dd-409e-a0ca-f7573676d9b3-openstack-edpm-ipam\") pod \"dnsmasq-dns-78c64bc9c5-hktkl\" (UID: \"d64b1ace-43dd-409e-a0ca-f7573676d9b3\") " pod="openstack/dnsmasq-dns-78c64bc9c5-hktkl" Feb 17 13:26:56 crc kubenswrapper[4955]: I0217 13:26:56.820386 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hddj\" (UniqueName: \"kubernetes.io/projected/d64b1ace-43dd-409e-a0ca-f7573676d9b3-kube-api-access-7hddj\") pod \"dnsmasq-dns-78c64bc9c5-hktkl\" (UID: \"d64b1ace-43dd-409e-a0ca-f7573676d9b3\") " pod="openstack/dnsmasq-dns-78c64bc9c5-hktkl" Feb 17 13:26:56 crc kubenswrapper[4955]: I0217 13:26:56.820482 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d64b1ace-43dd-409e-a0ca-f7573676d9b3-ovsdbserver-sb\") pod \"dnsmasq-dns-78c64bc9c5-hktkl\" (UID: \"d64b1ace-43dd-409e-a0ca-f7573676d9b3\") " pod="openstack/dnsmasq-dns-78c64bc9c5-hktkl" Feb 17 13:26:56 crc kubenswrapper[4955]: I0217 13:26:56.820561 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d64b1ace-43dd-409e-a0ca-f7573676d9b3-dns-swift-storage-0\") pod \"dnsmasq-dns-78c64bc9c5-hktkl\" (UID: \"d64b1ace-43dd-409e-a0ca-f7573676d9b3\") " pod="openstack/dnsmasq-dns-78c64bc9c5-hktkl" Feb 17 13:26:56 crc kubenswrapper[4955]: I0217 13:26:56.820600 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d64b1ace-43dd-409e-a0ca-f7573676d9b3-config\") pod \"dnsmasq-dns-78c64bc9c5-hktkl\" (UID: \"d64b1ace-43dd-409e-a0ca-f7573676d9b3\") " pod="openstack/dnsmasq-dns-78c64bc9c5-hktkl" Feb 17 13:26:56 crc kubenswrapper[4955]: I0217 13:26:56.820678 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d64b1ace-43dd-409e-a0ca-f7573676d9b3-dns-svc\") pod \"dnsmasq-dns-78c64bc9c5-hktkl\" (UID: \"d64b1ace-43dd-409e-a0ca-f7573676d9b3\") " pod="openstack/dnsmasq-dns-78c64bc9c5-hktkl" Feb 17 13:26:56 crc kubenswrapper[4955]: I0217 13:26:56.820716 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d64b1ace-43dd-409e-a0ca-f7573676d9b3-ovsdbserver-nb\") pod \"dnsmasq-dns-78c64bc9c5-hktkl\" (UID: \"d64b1ace-43dd-409e-a0ca-f7573676d9b3\") " pod="openstack/dnsmasq-dns-78c64bc9c5-hktkl" Feb 17 13:26:56 crc kubenswrapper[4955]: I0217 13:26:56.922355 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d64b1ace-43dd-409e-a0ca-f7573676d9b3-ovsdbserver-sb\") pod \"dnsmasq-dns-78c64bc9c5-hktkl\" (UID: \"d64b1ace-43dd-409e-a0ca-f7573676d9b3\") " pod="openstack/dnsmasq-dns-78c64bc9c5-hktkl" Feb 17 13:26:56 crc kubenswrapper[4955]: I0217 13:26:56.922431 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d64b1ace-43dd-409e-a0ca-f7573676d9b3-dns-swift-storage-0\") pod \"dnsmasq-dns-78c64bc9c5-hktkl\" (UID: \"d64b1ace-43dd-409e-a0ca-f7573676d9b3\") " pod="openstack/dnsmasq-dns-78c64bc9c5-hktkl" Feb 17 13:26:56 crc kubenswrapper[4955]: I0217 13:26:56.922461 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d64b1ace-43dd-409e-a0ca-f7573676d9b3-config\") pod \"dnsmasq-dns-78c64bc9c5-hktkl\" (UID: \"d64b1ace-43dd-409e-a0ca-f7573676d9b3\") " pod="openstack/dnsmasq-dns-78c64bc9c5-hktkl" Feb 17 13:26:56 crc kubenswrapper[4955]: I0217 13:26:56.922504 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d64b1ace-43dd-409e-a0ca-f7573676d9b3-dns-svc\") pod \"dnsmasq-dns-78c64bc9c5-hktkl\" (UID: \"d64b1ace-43dd-409e-a0ca-f7573676d9b3\") " pod="openstack/dnsmasq-dns-78c64bc9c5-hktkl" Feb 17 13:26:56 crc kubenswrapper[4955]: I0217 13:26:56.922524 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d64b1ace-43dd-409e-a0ca-f7573676d9b3-ovsdbserver-nb\") pod \"dnsmasq-dns-78c64bc9c5-hktkl\" (UID: \"d64b1ace-43dd-409e-a0ca-f7573676d9b3\") " pod="openstack/dnsmasq-dns-78c64bc9c5-hktkl" Feb 17 13:26:56 crc kubenswrapper[4955]: I0217 13:26:56.922590 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d64b1ace-43dd-409e-a0ca-f7573676d9b3-openstack-edpm-ipam\") pod \"dnsmasq-dns-78c64bc9c5-hktkl\" (UID: \"d64b1ace-43dd-409e-a0ca-f7573676d9b3\") " pod="openstack/dnsmasq-dns-78c64bc9c5-hktkl" Feb 17 13:26:56 crc kubenswrapper[4955]: I0217 13:26:56.922616 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hddj\" (UniqueName: \"kubernetes.io/projected/d64b1ace-43dd-409e-a0ca-f7573676d9b3-kube-api-access-7hddj\") pod \"dnsmasq-dns-78c64bc9c5-hktkl\" (UID: \"d64b1ace-43dd-409e-a0ca-f7573676d9b3\") " pod="openstack/dnsmasq-dns-78c64bc9c5-hktkl" Feb 17 13:26:56 crc kubenswrapper[4955]: I0217 13:26:56.923818 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d64b1ace-43dd-409e-a0ca-f7573676d9b3-dns-svc\") pod \"dnsmasq-dns-78c64bc9c5-hktkl\" (UID: \"d64b1ace-43dd-409e-a0ca-f7573676d9b3\") " pod="openstack/dnsmasq-dns-78c64bc9c5-hktkl" Feb 17 13:26:56 crc kubenswrapper[4955]: I0217 13:26:56.923878 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d64b1ace-43dd-409e-a0ca-f7573676d9b3-ovsdbserver-sb\") pod \"dnsmasq-dns-78c64bc9c5-hktkl\" (UID: \"d64b1ace-43dd-409e-a0ca-f7573676d9b3\") " pod="openstack/dnsmasq-dns-78c64bc9c5-hktkl" Feb 17 13:26:56 crc kubenswrapper[4955]: I0217 13:26:56.924110 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d64b1ace-43dd-409e-a0ca-f7573676d9b3-openstack-edpm-ipam\") pod \"dnsmasq-dns-78c64bc9c5-hktkl\" (UID: \"d64b1ace-43dd-409e-a0ca-f7573676d9b3\") " pod="openstack/dnsmasq-dns-78c64bc9c5-hktkl" Feb 17 13:26:56 crc kubenswrapper[4955]: I0217 13:26:56.924116 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d64b1ace-43dd-409e-a0ca-f7573676d9b3-dns-swift-storage-0\") pod \"dnsmasq-dns-78c64bc9c5-hktkl\" (UID: \"d64b1ace-43dd-409e-a0ca-f7573676d9b3\") " pod="openstack/dnsmasq-dns-78c64bc9c5-hktkl" Feb 17 13:26:56 crc kubenswrapper[4955]: I0217 13:26:56.924635 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d64b1ace-43dd-409e-a0ca-f7573676d9b3-config\") pod \"dnsmasq-dns-78c64bc9c5-hktkl\" (UID: \"d64b1ace-43dd-409e-a0ca-f7573676d9b3\") " pod="openstack/dnsmasq-dns-78c64bc9c5-hktkl" Feb 17 13:26:56 crc kubenswrapper[4955]: I0217 13:26:56.924715 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d64b1ace-43dd-409e-a0ca-f7573676d9b3-ovsdbserver-nb\") pod \"dnsmasq-dns-78c64bc9c5-hktkl\" (UID: \"d64b1ace-43dd-409e-a0ca-f7573676d9b3\") " pod="openstack/dnsmasq-dns-78c64bc9c5-hktkl" Feb 17 13:26:56 crc kubenswrapper[4955]: I0217 13:26:56.950116 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hddj\" (UniqueName: \"kubernetes.io/projected/d64b1ace-43dd-409e-a0ca-f7573676d9b3-kube-api-access-7hddj\") pod \"dnsmasq-dns-78c64bc9c5-hktkl\" (UID: \"d64b1ace-43dd-409e-a0ca-f7573676d9b3\") " pod="openstack/dnsmasq-dns-78c64bc9c5-hktkl" Feb 17 13:26:57 crc kubenswrapper[4955]: I0217 13:26:57.017096 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78c64bc9c5-hktkl" Feb 17 13:26:57 crc kubenswrapper[4955]: I0217 13:26:57.037621 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-9fmsq" Feb 17 13:26:57 crc kubenswrapper[4955]: I0217 13:26:57.231836 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ea144e9-32f5-4f5d-85fd-a6ec6e73add3-ovsdbserver-nb\") pod \"3ea144e9-32f5-4f5d-85fd-a6ec6e73add3\" (UID: \"3ea144e9-32f5-4f5d-85fd-a6ec6e73add3\") " Feb 17 13:26:57 crc kubenswrapper[4955]: I0217 13:26:57.231926 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ea144e9-32f5-4f5d-85fd-a6ec6e73add3-dns-svc\") pod \"3ea144e9-32f5-4f5d-85fd-a6ec6e73add3\" (UID: \"3ea144e9-32f5-4f5d-85fd-a6ec6e73add3\") " Feb 17 13:26:57 crc kubenswrapper[4955]: I0217 13:26:57.231961 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ea144e9-32f5-4f5d-85fd-a6ec6e73add3-config\") pod \"3ea144e9-32f5-4f5d-85fd-a6ec6e73add3\" (UID: \"3ea144e9-32f5-4f5d-85fd-a6ec6e73add3\") " Feb 17 13:26:57 crc kubenswrapper[4955]: I0217 13:26:57.232907 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ea144e9-32f5-4f5d-85fd-a6ec6e73add3-ovsdbserver-sb\") pod \"3ea144e9-32f5-4f5d-85fd-a6ec6e73add3\" (UID: \"3ea144e9-32f5-4f5d-85fd-a6ec6e73add3\") " Feb 17 13:26:57 crc kubenswrapper[4955]: I0217 13:26:57.233282 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3ea144e9-32f5-4f5d-85fd-a6ec6e73add3-dns-swift-storage-0\") pod \"3ea144e9-32f5-4f5d-85fd-a6ec6e73add3\" (UID: \"3ea144e9-32f5-4f5d-85fd-a6ec6e73add3\") " Feb 17 13:26:57 crc kubenswrapper[4955]: I0217 13:26:57.233356 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdq8t\" (UniqueName: \"kubernetes.io/projected/3ea144e9-32f5-4f5d-85fd-a6ec6e73add3-kube-api-access-zdq8t\") pod \"3ea144e9-32f5-4f5d-85fd-a6ec6e73add3\" (UID: \"3ea144e9-32f5-4f5d-85fd-a6ec6e73add3\") " Feb 17 13:26:57 crc kubenswrapper[4955]: I0217 13:26:57.241125 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ea144e9-32f5-4f5d-85fd-a6ec6e73add3-kube-api-access-zdq8t" (OuterVolumeSpecName: "kube-api-access-zdq8t") pod "3ea144e9-32f5-4f5d-85fd-a6ec6e73add3" (UID: "3ea144e9-32f5-4f5d-85fd-a6ec6e73add3"). InnerVolumeSpecName "kube-api-access-zdq8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:26:57 crc kubenswrapper[4955]: I0217 13:26:57.285889 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ea144e9-32f5-4f5d-85fd-a6ec6e73add3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3ea144e9-32f5-4f5d-85fd-a6ec6e73add3" (UID: "3ea144e9-32f5-4f5d-85fd-a6ec6e73add3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:26:57 crc kubenswrapper[4955]: I0217 13:26:57.293993 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ea144e9-32f5-4f5d-85fd-a6ec6e73add3-config" (OuterVolumeSpecName: "config") pod "3ea144e9-32f5-4f5d-85fd-a6ec6e73add3" (UID: "3ea144e9-32f5-4f5d-85fd-a6ec6e73add3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:26:57 crc kubenswrapper[4955]: I0217 13:26:57.302478 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ea144e9-32f5-4f5d-85fd-a6ec6e73add3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3ea144e9-32f5-4f5d-85fd-a6ec6e73add3" (UID: "3ea144e9-32f5-4f5d-85fd-a6ec6e73add3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:26:57 crc kubenswrapper[4955]: I0217 13:26:57.303763 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ea144e9-32f5-4f5d-85fd-a6ec6e73add3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3ea144e9-32f5-4f5d-85fd-a6ec6e73add3" (UID: "3ea144e9-32f5-4f5d-85fd-a6ec6e73add3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:26:57 crc kubenswrapper[4955]: I0217 13:26:57.304403 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ea144e9-32f5-4f5d-85fd-a6ec6e73add3-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3ea144e9-32f5-4f5d-85fd-a6ec6e73add3" (UID: "3ea144e9-32f5-4f5d-85fd-a6ec6e73add3"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:26:57 crc kubenswrapper[4955]: I0217 13:26:57.336930 4955 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3ea144e9-32f5-4f5d-85fd-a6ec6e73add3-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 17 13:26:57 crc kubenswrapper[4955]: I0217 13:26:57.336968 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdq8t\" (UniqueName: \"kubernetes.io/projected/3ea144e9-32f5-4f5d-85fd-a6ec6e73add3-kube-api-access-zdq8t\") on node \"crc\" DevicePath \"\"" Feb 17 13:26:57 crc kubenswrapper[4955]: I0217 13:26:57.336983 4955 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ea144e9-32f5-4f5d-85fd-a6ec6e73add3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 17 13:26:57 crc kubenswrapper[4955]: I0217 13:26:57.336999 4955 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ea144e9-32f5-4f5d-85fd-a6ec6e73add3-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 17 13:26:57 crc kubenswrapper[4955]: I0217 13:26:57.337012 4955 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ea144e9-32f5-4f5d-85fd-a6ec6e73add3-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:26:57 crc kubenswrapper[4955]: I0217 13:26:57.337024 4955 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ea144e9-32f5-4f5d-85fd-a6ec6e73add3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 17 13:26:57 crc kubenswrapper[4955]: I0217 13:26:57.391890 4955 generic.go:334] "Generic (PLEG): container finished" podID="3ea144e9-32f5-4f5d-85fd-a6ec6e73add3" containerID="8b42b497e8dd742535e0e73c9eb2c77ffc626ae2e0cc6c6f483d434219efb640" exitCode=0 Feb 17 13:26:57 crc kubenswrapper[4955]: I0217 13:26:57.391969 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-9fmsq" Feb 17 13:26:57 crc kubenswrapper[4955]: I0217 13:26:57.391977 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-9fmsq" event={"ID":"3ea144e9-32f5-4f5d-85fd-a6ec6e73add3","Type":"ContainerDied","Data":"8b42b497e8dd742535e0e73c9eb2c77ffc626ae2e0cc6c6f483d434219efb640"} Feb 17 13:26:57 crc kubenswrapper[4955]: I0217 13:26:57.392406 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-9fmsq" event={"ID":"3ea144e9-32f5-4f5d-85fd-a6ec6e73add3","Type":"ContainerDied","Data":"b877b588b2d967a7f359cfb8338966090fa0ab03a837b925937fe7e9e3b3562d"} Feb 17 13:26:57 crc kubenswrapper[4955]: I0217 13:26:57.392430 4955 scope.go:117] "RemoveContainer" containerID="8b42b497e8dd742535e0e73c9eb2c77ffc626ae2e0cc6c6f483d434219efb640" Feb 17 13:26:57 crc kubenswrapper[4955]: I0217 13:26:57.418131 4955 scope.go:117] "RemoveContainer" containerID="22214ad7f31c951e9c4565fdbc15df40fec663e16d96a70ca906278b9bd1a11a" Feb 17 13:26:57 crc kubenswrapper[4955]: I0217 13:26:57.442069 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-9fmsq"] Feb 17 13:26:57 crc kubenswrapper[4955]: I0217 13:26:57.445656 4955 scope.go:117] "RemoveContainer" containerID="8b42b497e8dd742535e0e73c9eb2c77ffc626ae2e0cc6c6f483d434219efb640" Feb 17 13:26:57 crc kubenswrapper[4955]: E0217 13:26:57.446225 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b42b497e8dd742535e0e73c9eb2c77ffc626ae2e0cc6c6f483d434219efb640\": container with ID starting with 8b42b497e8dd742535e0e73c9eb2c77ffc626ae2e0cc6c6f483d434219efb640 not found: ID does not exist" containerID="8b42b497e8dd742535e0e73c9eb2c77ffc626ae2e0cc6c6f483d434219efb640" Feb 17 13:26:57 crc kubenswrapper[4955]: I0217 13:26:57.446327 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b42b497e8dd742535e0e73c9eb2c77ffc626ae2e0cc6c6f483d434219efb640"} err="failed to get container status \"8b42b497e8dd742535e0e73c9eb2c77ffc626ae2e0cc6c6f483d434219efb640\": rpc error: code = NotFound desc = could not find container \"8b42b497e8dd742535e0e73c9eb2c77ffc626ae2e0cc6c6f483d434219efb640\": container with ID starting with 8b42b497e8dd742535e0e73c9eb2c77ffc626ae2e0cc6c6f483d434219efb640 not found: ID does not exist" Feb 17 13:26:57 crc kubenswrapper[4955]: I0217 13:26:57.446517 4955 scope.go:117] "RemoveContainer" containerID="22214ad7f31c951e9c4565fdbc15df40fec663e16d96a70ca906278b9bd1a11a" Feb 17 13:26:57 crc kubenswrapper[4955]: E0217 13:26:57.447029 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22214ad7f31c951e9c4565fdbc15df40fec663e16d96a70ca906278b9bd1a11a\": container with ID starting with 22214ad7f31c951e9c4565fdbc15df40fec663e16d96a70ca906278b9bd1a11a not found: ID does not exist" containerID="22214ad7f31c951e9c4565fdbc15df40fec663e16d96a70ca906278b9bd1a11a" Feb 17 13:26:57 crc kubenswrapper[4955]: I0217 13:26:57.447123 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22214ad7f31c951e9c4565fdbc15df40fec663e16d96a70ca906278b9bd1a11a"} err="failed to get container status \"22214ad7f31c951e9c4565fdbc15df40fec663e16d96a70ca906278b9bd1a11a\": rpc error: code = NotFound desc = could not find container \"22214ad7f31c951e9c4565fdbc15df40fec663e16d96a70ca906278b9bd1a11a\": container with ID starting with 22214ad7f31c951e9c4565fdbc15df40fec663e16d96a70ca906278b9bd1a11a not found: ID does not exist" Feb 17 13:26:57 crc kubenswrapper[4955]: I0217 13:26:57.451497 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-9fmsq"] Feb 17 13:26:57 crc kubenswrapper[4955]: I0217 13:26:57.512147 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78c64bc9c5-hktkl"] Feb 17 13:26:58 crc kubenswrapper[4955]: I0217 13:26:58.241829 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ea144e9-32f5-4f5d-85fd-a6ec6e73add3" path="/var/lib/kubelet/pods/3ea144e9-32f5-4f5d-85fd-a6ec6e73add3/volumes" Feb 17 13:26:58 crc kubenswrapper[4955]: I0217 13:26:58.404159 4955 generic.go:334] "Generic (PLEG): container finished" podID="d64b1ace-43dd-409e-a0ca-f7573676d9b3" containerID="9e0d8aeb55dd550ab6b6d351c0756895471b2152ad8462d8be606c306d3106f7" exitCode=0 Feb 17 13:26:58 crc kubenswrapper[4955]: I0217 13:26:58.404211 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78c64bc9c5-hktkl" event={"ID":"d64b1ace-43dd-409e-a0ca-f7573676d9b3","Type":"ContainerDied","Data":"9e0d8aeb55dd550ab6b6d351c0756895471b2152ad8462d8be606c306d3106f7"} Feb 17 13:26:58 crc kubenswrapper[4955]: I0217 13:26:58.404265 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78c64bc9c5-hktkl" event={"ID":"d64b1ace-43dd-409e-a0ca-f7573676d9b3","Type":"ContainerStarted","Data":"17e33e19489188f9ef9bf8a7c981320a8adde5d7ebd9a7a5b24a02a0d420c2bf"} Feb 17 13:26:59 crc kubenswrapper[4955]: I0217 13:26:59.417198 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78c64bc9c5-hktkl" event={"ID":"d64b1ace-43dd-409e-a0ca-f7573676d9b3","Type":"ContainerStarted","Data":"8b6c6aac833fa7f8a1a0a91f5957907ca6e9db7a59b67d9f3db763ec52614482"} Feb 17 13:26:59 crc kubenswrapper[4955]: I0217 13:26:59.417982 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-78c64bc9c5-hktkl" Feb 17 13:26:59 crc kubenswrapper[4955]: I0217 13:26:59.442580 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-78c64bc9c5-hktkl" podStartSLOduration=3.442564709 podStartE2EDuration="3.442564709s" podCreationTimestamp="2026-02-17 13:26:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:26:59.440038916 +0000 UTC m=+1357.962768449" watchObservedRunningTime="2026-02-17 13:26:59.442564709 +0000 UTC m=+1357.965294242" Feb 17 13:27:04 crc kubenswrapper[4955]: I0217 13:27:04.574510 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 13:27:04 crc kubenswrapper[4955]: I0217 13:27:04.575051 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 13:27:04 crc kubenswrapper[4955]: I0217 13:27:04.575103 4955 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" Feb 17 13:27:04 crc kubenswrapper[4955]: I0217 13:27:04.575848 4955 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2a2497139f3e8754b75d83788b3f1744445e509419949dab14b01c6f81278a2c"} pod="openshift-machine-config-operator/machine-config-daemon-29qxq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 17 13:27:04 crc kubenswrapper[4955]: I0217 13:27:04.575894 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" containerID="cri-o://2a2497139f3e8754b75d83788b3f1744445e509419949dab14b01c6f81278a2c" gracePeriod=600 Feb 17 13:27:05 crc kubenswrapper[4955]: I0217 13:27:05.474668 4955 generic.go:334] "Generic (PLEG): container finished" podID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerID="2a2497139f3e8754b75d83788b3f1744445e509419949dab14b01c6f81278a2c" exitCode=0 Feb 17 13:27:05 crc kubenswrapper[4955]: I0217 13:27:05.474748 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" event={"ID":"5bf36887-2a2b-4193-abdb-a15beaf46300","Type":"ContainerDied","Data":"2a2497139f3e8754b75d83788b3f1744445e509419949dab14b01c6f81278a2c"} Feb 17 13:27:05 crc kubenswrapper[4955]: I0217 13:27:05.475158 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" event={"ID":"5bf36887-2a2b-4193-abdb-a15beaf46300","Type":"ContainerStarted","Data":"969807d9ca63cdcc1aa7cc07aa21a249802167bf58b0b6150149361ce19b14ca"} Feb 17 13:27:05 crc kubenswrapper[4955]: I0217 13:27:05.475175 4955 scope.go:117] "RemoveContainer" containerID="3481e1de1949e001dc1ec5b32fb42f82ef040a09d687f51932cebedf9288edf6" Feb 17 13:27:07 crc kubenswrapper[4955]: I0217 13:27:07.020150 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-78c64bc9c5-hktkl" Feb 17 13:27:07 crc kubenswrapper[4955]: I0217 13:27:07.102363 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-mqt6k"] Feb 17 13:27:07 crc kubenswrapper[4955]: I0217 13:27:07.102687 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-d558885bc-mqt6k" podUID="c31031a2-1485-4507-8a90-d06913010865" containerName="dnsmasq-dns" containerID="cri-o://75ea6a58f6a0018b69ba7b7aa87832742c54b55ee898a837323941bb6ed92575" gracePeriod=10 Feb 17 13:27:07 crc kubenswrapper[4955]: I0217 13:27:07.501941 4955 generic.go:334] "Generic (PLEG): container finished" podID="c31031a2-1485-4507-8a90-d06913010865" containerID="75ea6a58f6a0018b69ba7b7aa87832742c54b55ee898a837323941bb6ed92575" exitCode=0 Feb 17 13:27:07 crc kubenswrapper[4955]: I0217 13:27:07.501993 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-mqt6k" event={"ID":"c31031a2-1485-4507-8a90-d06913010865","Type":"ContainerDied","Data":"75ea6a58f6a0018b69ba7b7aa87832742c54b55ee898a837323941bb6ed92575"} Feb 17 13:27:07 crc kubenswrapper[4955]: I0217 13:27:07.502028 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-mqt6k" event={"ID":"c31031a2-1485-4507-8a90-d06913010865","Type":"ContainerDied","Data":"5ae0132e3fff7b2d7610acb3549b234d39d4781cda4c5118ffee1658689e8b7e"} Feb 17 13:27:07 crc kubenswrapper[4955]: I0217 13:27:07.502043 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ae0132e3fff7b2d7610acb3549b234d39d4781cda4c5118ffee1658689e8b7e" Feb 17 13:27:07 crc kubenswrapper[4955]: I0217 13:27:07.545267 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-mqt6k" Feb 17 13:27:07 crc kubenswrapper[4955]: I0217 13:27:07.646260 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c31031a2-1485-4507-8a90-d06913010865-ovsdbserver-sb\") pod \"c31031a2-1485-4507-8a90-d06913010865\" (UID: \"c31031a2-1485-4507-8a90-d06913010865\") " Feb 17 13:27:07 crc kubenswrapper[4955]: I0217 13:27:07.646632 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c31031a2-1485-4507-8a90-d06913010865-config\") pod \"c31031a2-1485-4507-8a90-d06913010865\" (UID: \"c31031a2-1485-4507-8a90-d06913010865\") " Feb 17 13:27:07 crc kubenswrapper[4955]: I0217 13:27:07.646691 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jw7dh\" (UniqueName: \"kubernetes.io/projected/c31031a2-1485-4507-8a90-d06913010865-kube-api-access-jw7dh\") pod \"c31031a2-1485-4507-8a90-d06913010865\" (UID: \"c31031a2-1485-4507-8a90-d06913010865\") " Feb 17 13:27:07 crc kubenswrapper[4955]: I0217 13:27:07.646842 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c31031a2-1485-4507-8a90-d06913010865-dns-svc\") pod \"c31031a2-1485-4507-8a90-d06913010865\" (UID: \"c31031a2-1485-4507-8a90-d06913010865\") " Feb 17 13:27:07 crc kubenswrapper[4955]: I0217 13:27:07.646881 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c31031a2-1485-4507-8a90-d06913010865-ovsdbserver-nb\") pod \"c31031a2-1485-4507-8a90-d06913010865\" (UID: \"c31031a2-1485-4507-8a90-d06913010865\") " Feb 17 13:27:07 crc kubenswrapper[4955]: I0217 13:27:07.646955 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/c31031a2-1485-4507-8a90-d06913010865-openstack-edpm-ipam\") pod \"c31031a2-1485-4507-8a90-d06913010865\" (UID: \"c31031a2-1485-4507-8a90-d06913010865\") " Feb 17 13:27:07 crc kubenswrapper[4955]: I0217 13:27:07.647009 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c31031a2-1485-4507-8a90-d06913010865-dns-swift-storage-0\") pod \"c31031a2-1485-4507-8a90-d06913010865\" (UID: \"c31031a2-1485-4507-8a90-d06913010865\") " Feb 17 13:27:07 crc kubenswrapper[4955]: I0217 13:27:07.654202 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c31031a2-1485-4507-8a90-d06913010865-kube-api-access-jw7dh" (OuterVolumeSpecName: "kube-api-access-jw7dh") pod "c31031a2-1485-4507-8a90-d06913010865" (UID: "c31031a2-1485-4507-8a90-d06913010865"). InnerVolumeSpecName "kube-api-access-jw7dh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:27:07 crc kubenswrapper[4955]: I0217 13:27:07.696431 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c31031a2-1485-4507-8a90-d06913010865-config" (OuterVolumeSpecName: "config") pod "c31031a2-1485-4507-8a90-d06913010865" (UID: "c31031a2-1485-4507-8a90-d06913010865"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:27:07 crc kubenswrapper[4955]: I0217 13:27:07.696488 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c31031a2-1485-4507-8a90-d06913010865-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "c31031a2-1485-4507-8a90-d06913010865" (UID: "c31031a2-1485-4507-8a90-d06913010865"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:27:07 crc kubenswrapper[4955]: I0217 13:27:07.700505 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c31031a2-1485-4507-8a90-d06913010865-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c31031a2-1485-4507-8a90-d06913010865" (UID: "c31031a2-1485-4507-8a90-d06913010865"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:27:07 crc kubenswrapper[4955]: I0217 13:27:07.713965 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c31031a2-1485-4507-8a90-d06913010865-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c31031a2-1485-4507-8a90-d06913010865" (UID: "c31031a2-1485-4507-8a90-d06913010865"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:27:07 crc kubenswrapper[4955]: I0217 13:27:07.714929 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c31031a2-1485-4507-8a90-d06913010865-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c31031a2-1485-4507-8a90-d06913010865" (UID: "c31031a2-1485-4507-8a90-d06913010865"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:27:07 crc kubenswrapper[4955]: I0217 13:27:07.723066 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c31031a2-1485-4507-8a90-d06913010865-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c31031a2-1485-4507-8a90-d06913010865" (UID: "c31031a2-1485-4507-8a90-d06913010865"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:27:07 crc kubenswrapper[4955]: I0217 13:27:07.749847 4955 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c31031a2-1485-4507-8a90-d06913010865-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:27:07 crc kubenswrapper[4955]: I0217 13:27:07.749881 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jw7dh\" (UniqueName: \"kubernetes.io/projected/c31031a2-1485-4507-8a90-d06913010865-kube-api-access-jw7dh\") on node \"crc\" DevicePath \"\"" Feb 17 13:27:07 crc kubenswrapper[4955]: I0217 13:27:07.749892 4955 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c31031a2-1485-4507-8a90-d06913010865-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 17 13:27:07 crc kubenswrapper[4955]: I0217 13:27:07.749903 4955 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c31031a2-1485-4507-8a90-d06913010865-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 17 13:27:07 crc kubenswrapper[4955]: I0217 13:27:07.749912 4955 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/c31031a2-1485-4507-8a90-d06913010865-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 17 13:27:07 crc kubenswrapper[4955]: I0217 13:27:07.749920 4955 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c31031a2-1485-4507-8a90-d06913010865-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 17 13:27:07 crc kubenswrapper[4955]: I0217 13:27:07.749929 4955 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c31031a2-1485-4507-8a90-d06913010865-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 17 13:27:08 crc kubenswrapper[4955]: I0217 13:27:08.522155 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-mqt6k" Feb 17 13:27:08 crc kubenswrapper[4955]: I0217 13:27:08.546595 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-mqt6k"] Feb 17 13:27:08 crc kubenswrapper[4955]: I0217 13:27:08.556770 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-mqt6k"] Feb 17 13:27:10 crc kubenswrapper[4955]: I0217 13:27:10.235625 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c31031a2-1485-4507-8a90-d06913010865" path="/var/lib/kubelet/pods/c31031a2-1485-4507-8a90-d06913010865/volumes" Feb 17 13:27:15 crc kubenswrapper[4955]: I0217 13:27:15.686433 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g2nhl"] Feb 17 13:27:15 crc kubenswrapper[4955]: E0217 13:27:15.687352 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c31031a2-1485-4507-8a90-d06913010865" containerName="init" Feb 17 13:27:15 crc kubenswrapper[4955]: I0217 13:27:15.687367 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="c31031a2-1485-4507-8a90-d06913010865" containerName="init" Feb 17 13:27:15 crc kubenswrapper[4955]: E0217 13:27:15.687379 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c31031a2-1485-4507-8a90-d06913010865" containerName="dnsmasq-dns" Feb 17 13:27:15 crc kubenswrapper[4955]: I0217 13:27:15.687385 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="c31031a2-1485-4507-8a90-d06913010865" containerName="dnsmasq-dns" Feb 17 13:27:15 crc kubenswrapper[4955]: E0217 13:27:15.687405 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ea144e9-32f5-4f5d-85fd-a6ec6e73add3" containerName="init" Feb 17 13:27:15 crc kubenswrapper[4955]: I0217 13:27:15.687411 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ea144e9-32f5-4f5d-85fd-a6ec6e73add3" containerName="init" Feb 17 13:27:15 crc kubenswrapper[4955]: E0217 13:27:15.687421 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ea144e9-32f5-4f5d-85fd-a6ec6e73add3" containerName="dnsmasq-dns" Feb 17 13:27:15 crc kubenswrapper[4955]: I0217 13:27:15.687426 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ea144e9-32f5-4f5d-85fd-a6ec6e73add3" containerName="dnsmasq-dns" Feb 17 13:27:15 crc kubenswrapper[4955]: I0217 13:27:15.687605 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ea144e9-32f5-4f5d-85fd-a6ec6e73add3" containerName="dnsmasq-dns" Feb 17 13:27:15 crc kubenswrapper[4955]: I0217 13:27:15.687633 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="c31031a2-1485-4507-8a90-d06913010865" containerName="dnsmasq-dns" Feb 17 13:27:15 crc kubenswrapper[4955]: I0217 13:27:15.688197 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g2nhl" Feb 17 13:27:15 crc kubenswrapper[4955]: I0217 13:27:15.689986 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 17 13:27:15 crc kubenswrapper[4955]: I0217 13:27:15.691030 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 17 13:27:15 crc kubenswrapper[4955]: I0217 13:27:15.691111 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 17 13:27:15 crc kubenswrapper[4955]: I0217 13:27:15.691187 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-t9lgc" Feb 17 13:27:15 crc kubenswrapper[4955]: I0217 13:27:15.712433 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g2nhl"] Feb 17 13:27:15 crc kubenswrapper[4955]: I0217 13:27:15.748638 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e1bffb7a-a33a-466d-b382-a6896e83d8d9-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g2nhl\" (UID: \"e1bffb7a-a33a-466d-b382-a6896e83d8d9\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g2nhl" Feb 17 13:27:15 crc kubenswrapper[4955]: I0217 13:27:15.748725 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1bffb7a-a33a-466d-b382-a6896e83d8d9-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g2nhl\" (UID: \"e1bffb7a-a33a-466d-b382-a6896e83d8d9\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g2nhl" Feb 17 13:27:15 crc kubenswrapper[4955]: I0217 13:27:15.748854 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbggb\" (UniqueName: \"kubernetes.io/projected/e1bffb7a-a33a-466d-b382-a6896e83d8d9-kube-api-access-jbggb\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g2nhl\" (UID: \"e1bffb7a-a33a-466d-b382-a6896e83d8d9\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g2nhl" Feb 17 13:27:15 crc kubenswrapper[4955]: I0217 13:27:15.748968 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e1bffb7a-a33a-466d-b382-a6896e83d8d9-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g2nhl\" (UID: \"e1bffb7a-a33a-466d-b382-a6896e83d8d9\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g2nhl" Feb 17 13:27:15 crc kubenswrapper[4955]: I0217 13:27:15.850357 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbggb\" (UniqueName: \"kubernetes.io/projected/e1bffb7a-a33a-466d-b382-a6896e83d8d9-kube-api-access-jbggb\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g2nhl\" (UID: \"e1bffb7a-a33a-466d-b382-a6896e83d8d9\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g2nhl" Feb 17 13:27:15 crc kubenswrapper[4955]: I0217 13:27:15.850438 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e1bffb7a-a33a-466d-b382-a6896e83d8d9-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g2nhl\" (UID: \"e1bffb7a-a33a-466d-b382-a6896e83d8d9\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g2nhl" Feb 17 13:27:15 crc kubenswrapper[4955]: I0217 13:27:15.850518 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e1bffb7a-a33a-466d-b382-a6896e83d8d9-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g2nhl\" (UID: \"e1bffb7a-a33a-466d-b382-a6896e83d8d9\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g2nhl" Feb 17 13:27:15 crc kubenswrapper[4955]: I0217 13:27:15.850553 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1bffb7a-a33a-466d-b382-a6896e83d8d9-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g2nhl\" (UID: \"e1bffb7a-a33a-466d-b382-a6896e83d8d9\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g2nhl" Feb 17 13:27:15 crc kubenswrapper[4955]: I0217 13:27:15.857100 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e1bffb7a-a33a-466d-b382-a6896e83d8d9-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g2nhl\" (UID: \"e1bffb7a-a33a-466d-b382-a6896e83d8d9\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g2nhl" Feb 17 13:27:15 crc kubenswrapper[4955]: I0217 13:27:15.857815 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e1bffb7a-a33a-466d-b382-a6896e83d8d9-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g2nhl\" (UID: \"e1bffb7a-a33a-466d-b382-a6896e83d8d9\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g2nhl" Feb 17 13:27:15 crc kubenswrapper[4955]: I0217 13:27:15.858587 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1bffb7a-a33a-466d-b382-a6896e83d8d9-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g2nhl\" (UID: \"e1bffb7a-a33a-466d-b382-a6896e83d8d9\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g2nhl" Feb 17 13:27:15 crc kubenswrapper[4955]: I0217 13:27:15.869828 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbggb\" (UniqueName: \"kubernetes.io/projected/e1bffb7a-a33a-466d-b382-a6896e83d8d9-kube-api-access-jbggb\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g2nhl\" (UID: \"e1bffb7a-a33a-466d-b382-a6896e83d8d9\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g2nhl" Feb 17 13:27:16 crc kubenswrapper[4955]: I0217 13:27:16.011575 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g2nhl" Feb 17 13:27:16 crc kubenswrapper[4955]: I0217 13:27:16.557729 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g2nhl"] Feb 17 13:27:16 crc kubenswrapper[4955]: I0217 13:27:16.633642 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g2nhl" event={"ID":"e1bffb7a-a33a-466d-b382-a6896e83d8d9","Type":"ContainerStarted","Data":"b6180b12f7dc882baace2bb4318354b22234697c51930a409be4dda5eba608d9"} Feb 17 13:27:20 crc kubenswrapper[4955]: I0217 13:27:20.680087 4955 generic.go:334] "Generic (PLEG): container finished" podID="cd70e7eb-fb12-4b99-ab83-67d4abb9392b" containerID="4493ad91867b6ad99901129eeaef053f033bb9685824695228eda14fa0ac124d" exitCode=0 Feb 17 13:27:20 crc kubenswrapper[4955]: I0217 13:27:20.680259 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"cd70e7eb-fb12-4b99-ab83-67d4abb9392b","Type":"ContainerDied","Data":"4493ad91867b6ad99901129eeaef053f033bb9685824695228eda14fa0ac124d"} Feb 17 13:27:20 crc kubenswrapper[4955]: I0217 13:27:20.682809 4955 generic.go:334] "Generic (PLEG): container finished" podID="c8c91914-fefa-479e-b497-de0f1f2978a2" containerID="dd90a1bfd0db84aa04a7f0fd77caaa9bb7fe5384e795186784679c7475b3abfd" exitCode=0 Feb 17 13:27:20 crc kubenswrapper[4955]: I0217 13:27:20.682845 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c8c91914-fefa-479e-b497-de0f1f2978a2","Type":"ContainerDied","Data":"dd90a1bfd0db84aa04a7f0fd77caaa9bb7fe5384e795186784679c7475b3abfd"} Feb 17 13:27:24 crc kubenswrapper[4955]: I0217 13:27:24.801512 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 17 13:27:25 crc kubenswrapper[4955]: I0217 13:27:25.738059 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c8c91914-fefa-479e-b497-de0f1f2978a2","Type":"ContainerStarted","Data":"5e583c0dfec999b033596dcb59baf195d9d56a4178133fa07454affaac72f643"} Feb 17 13:27:25 crc kubenswrapper[4955]: I0217 13:27:25.738745 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Feb 17 13:27:25 crc kubenswrapper[4955]: I0217 13:27:25.740149 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"cd70e7eb-fb12-4b99-ab83-67d4abb9392b","Type":"ContainerStarted","Data":"9ab38f643cfd2d5b60c24e8c3008ab61db6b57f83f3bb385d47d7e0e3e4df854"} Feb 17 13:27:25 crc kubenswrapper[4955]: I0217 13:27:25.740362 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:27:25 crc kubenswrapper[4955]: I0217 13:27:25.744217 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g2nhl" event={"ID":"e1bffb7a-a33a-466d-b382-a6896e83d8d9","Type":"ContainerStarted","Data":"a8399689018950fb006b903c2c310e9040951c55fb61b77bbfb78d950be9a476"} Feb 17 13:27:25 crc kubenswrapper[4955]: I0217 13:27:25.766626 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=40.766605671 podStartE2EDuration="40.766605671s" podCreationTimestamp="2026-02-17 13:26:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:27:25.759381784 +0000 UTC m=+1384.282111337" watchObservedRunningTime="2026-02-17 13:27:25.766605671 +0000 UTC m=+1384.289335214" Feb 17 13:27:25 crc kubenswrapper[4955]: I0217 13:27:25.792378 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=40.792357721 podStartE2EDuration="40.792357721s" podCreationTimestamp="2026-02-17 13:26:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:27:25.78918773 +0000 UTC m=+1384.311917283" watchObservedRunningTime="2026-02-17 13:27:25.792357721 +0000 UTC m=+1384.315087264" Feb 17 13:27:25 crc kubenswrapper[4955]: I0217 13:27:25.823710 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g2nhl" podStartSLOduration=2.643027068 podStartE2EDuration="10.82368913s" podCreationTimestamp="2026-02-17 13:27:15 +0000 UTC" firstStartedPulling="2026-02-17 13:27:16.617681964 +0000 UTC m=+1375.140411507" lastFinishedPulling="2026-02-17 13:27:24.798344016 +0000 UTC m=+1383.321073569" observedRunningTime="2026-02-17 13:27:25.816350749 +0000 UTC m=+1384.339080292" watchObservedRunningTime="2026-02-17 13:27:25.82368913 +0000 UTC m=+1384.346418673" Feb 17 13:27:27 crc kubenswrapper[4955]: I0217 13:27:27.917229 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jqmrl"] Feb 17 13:27:27 crc kubenswrapper[4955]: I0217 13:27:27.919206 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jqmrl" Feb 17 13:27:27 crc kubenswrapper[4955]: I0217 13:27:27.935535 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jqmrl"] Feb 17 13:27:28 crc kubenswrapper[4955]: I0217 13:27:28.034700 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c485bbf2-8d96-4c64-a2c8-37c76c557d4f-catalog-content\") pod \"certified-operators-jqmrl\" (UID: \"c485bbf2-8d96-4c64-a2c8-37c76c557d4f\") " pod="openshift-marketplace/certified-operators-jqmrl" Feb 17 13:27:28 crc kubenswrapper[4955]: I0217 13:27:28.034758 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c485bbf2-8d96-4c64-a2c8-37c76c557d4f-utilities\") pod \"certified-operators-jqmrl\" (UID: \"c485bbf2-8d96-4c64-a2c8-37c76c557d4f\") " pod="openshift-marketplace/certified-operators-jqmrl" Feb 17 13:27:28 crc kubenswrapper[4955]: I0217 13:27:28.034813 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nrsq\" (UniqueName: \"kubernetes.io/projected/c485bbf2-8d96-4c64-a2c8-37c76c557d4f-kube-api-access-2nrsq\") pod \"certified-operators-jqmrl\" (UID: \"c485bbf2-8d96-4c64-a2c8-37c76c557d4f\") " pod="openshift-marketplace/certified-operators-jqmrl" Feb 17 13:27:28 crc kubenswrapper[4955]: I0217 13:27:28.137118 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c485bbf2-8d96-4c64-a2c8-37c76c557d4f-catalog-content\") pod \"certified-operators-jqmrl\" (UID: \"c485bbf2-8d96-4c64-a2c8-37c76c557d4f\") " pod="openshift-marketplace/certified-operators-jqmrl" Feb 17 13:27:28 crc kubenswrapper[4955]: I0217 13:27:28.137187 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c485bbf2-8d96-4c64-a2c8-37c76c557d4f-utilities\") pod \"certified-operators-jqmrl\" (UID: \"c485bbf2-8d96-4c64-a2c8-37c76c557d4f\") " pod="openshift-marketplace/certified-operators-jqmrl" Feb 17 13:27:28 crc kubenswrapper[4955]: I0217 13:27:28.137225 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nrsq\" (UniqueName: \"kubernetes.io/projected/c485bbf2-8d96-4c64-a2c8-37c76c557d4f-kube-api-access-2nrsq\") pod \"certified-operators-jqmrl\" (UID: \"c485bbf2-8d96-4c64-a2c8-37c76c557d4f\") " pod="openshift-marketplace/certified-operators-jqmrl" Feb 17 13:27:28 crc kubenswrapper[4955]: I0217 13:27:28.137985 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c485bbf2-8d96-4c64-a2c8-37c76c557d4f-utilities\") pod \"certified-operators-jqmrl\" (UID: \"c485bbf2-8d96-4c64-a2c8-37c76c557d4f\") " pod="openshift-marketplace/certified-operators-jqmrl" Feb 17 13:27:28 crc kubenswrapper[4955]: I0217 13:27:28.138023 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c485bbf2-8d96-4c64-a2c8-37c76c557d4f-catalog-content\") pod \"certified-operators-jqmrl\" (UID: \"c485bbf2-8d96-4c64-a2c8-37c76c557d4f\") " pod="openshift-marketplace/certified-operators-jqmrl" Feb 17 13:27:28 crc kubenswrapper[4955]: I0217 13:27:28.164959 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nrsq\" (UniqueName: \"kubernetes.io/projected/c485bbf2-8d96-4c64-a2c8-37c76c557d4f-kube-api-access-2nrsq\") pod \"certified-operators-jqmrl\" (UID: \"c485bbf2-8d96-4c64-a2c8-37c76c557d4f\") " pod="openshift-marketplace/certified-operators-jqmrl" Feb 17 13:27:28 crc kubenswrapper[4955]: I0217 13:27:28.239158 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jqmrl" Feb 17 13:27:28 crc kubenswrapper[4955]: I0217 13:27:28.865260 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jqmrl"] Feb 17 13:27:29 crc kubenswrapper[4955]: I0217 13:27:29.816090 4955 generic.go:334] "Generic (PLEG): container finished" podID="c485bbf2-8d96-4c64-a2c8-37c76c557d4f" containerID="6e4966ac31e08553542beabcacb077ab467cbb2f7eb62826dce3f010f81b8c3f" exitCode=0 Feb 17 13:27:29 crc kubenswrapper[4955]: I0217 13:27:29.816212 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jqmrl" event={"ID":"c485bbf2-8d96-4c64-a2c8-37c76c557d4f","Type":"ContainerDied","Data":"6e4966ac31e08553542beabcacb077ab467cbb2f7eb62826dce3f010f81b8c3f"} Feb 17 13:27:29 crc kubenswrapper[4955]: I0217 13:27:29.816654 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jqmrl" event={"ID":"c485bbf2-8d96-4c64-a2c8-37c76c557d4f","Type":"ContainerStarted","Data":"66e37d9c78566a74b52756288f79204aaaa35449e79304f6e468e207b93dd154"} Feb 17 13:27:31 crc kubenswrapper[4955]: I0217 13:27:31.842512 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jqmrl" event={"ID":"c485bbf2-8d96-4c64-a2c8-37c76c557d4f","Type":"ContainerStarted","Data":"b5ea2958904dc0f2d17fad897b4bcc0032985cb7fd2e026fb041a3718bd91384"} Feb 17 13:27:32 crc kubenswrapper[4955]: I0217 13:27:32.855869 4955 generic.go:334] "Generic (PLEG): container finished" podID="c485bbf2-8d96-4c64-a2c8-37c76c557d4f" containerID="b5ea2958904dc0f2d17fad897b4bcc0032985cb7fd2e026fb041a3718bd91384" exitCode=0 Feb 17 13:27:32 crc kubenswrapper[4955]: I0217 13:27:32.855995 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jqmrl" event={"ID":"c485bbf2-8d96-4c64-a2c8-37c76c557d4f","Type":"ContainerDied","Data":"b5ea2958904dc0f2d17fad897b4bcc0032985cb7fd2e026fb041a3718bd91384"} Feb 17 13:27:33 crc kubenswrapper[4955]: I0217 13:27:33.870041 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jqmrl" event={"ID":"c485bbf2-8d96-4c64-a2c8-37c76c557d4f","Type":"ContainerStarted","Data":"1ca5a8bff6032c823bd77a4179eb8c586e2958762a10dc9dd08e01fd48e6a663"} Feb 17 13:27:33 crc kubenswrapper[4955]: I0217 13:27:33.887357 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jqmrl" podStartSLOduration=3.46536587 podStartE2EDuration="6.887317629s" podCreationTimestamp="2026-02-17 13:27:27 +0000 UTC" firstStartedPulling="2026-02-17 13:27:29.817776091 +0000 UTC m=+1388.340505644" lastFinishedPulling="2026-02-17 13:27:33.23972785 +0000 UTC m=+1391.762457403" observedRunningTime="2026-02-17 13:27:33.886693631 +0000 UTC m=+1392.409423184" watchObservedRunningTime="2026-02-17 13:27:33.887317629 +0000 UTC m=+1392.410047172" Feb 17 13:27:35 crc kubenswrapper[4955]: I0217 13:27:35.734998 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Feb 17 13:27:36 crc kubenswrapper[4955]: I0217 13:27:36.084296 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Feb 17 13:27:36 crc kubenswrapper[4955]: I0217 13:27:36.898414 4955 generic.go:334] "Generic (PLEG): container finished" podID="e1bffb7a-a33a-466d-b382-a6896e83d8d9" containerID="a8399689018950fb006b903c2c310e9040951c55fb61b77bbfb78d950be9a476" exitCode=0 Feb 17 13:27:36 crc kubenswrapper[4955]: I0217 13:27:36.898473 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g2nhl" event={"ID":"e1bffb7a-a33a-466d-b382-a6896e83d8d9","Type":"ContainerDied","Data":"a8399689018950fb006b903c2c310e9040951c55fb61b77bbfb78d950be9a476"} Feb 17 13:27:38 crc kubenswrapper[4955]: I0217 13:27:38.239341 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jqmrl" Feb 17 13:27:38 crc kubenswrapper[4955]: I0217 13:27:38.239669 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jqmrl" Feb 17 13:27:38 crc kubenswrapper[4955]: I0217 13:27:38.321562 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g2nhl" Feb 17 13:27:38 crc kubenswrapper[4955]: I0217 13:27:38.355718 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1bffb7a-a33a-466d-b382-a6896e83d8d9-repo-setup-combined-ca-bundle\") pod \"e1bffb7a-a33a-466d-b382-a6896e83d8d9\" (UID: \"e1bffb7a-a33a-466d-b382-a6896e83d8d9\") " Feb 17 13:27:38 crc kubenswrapper[4955]: I0217 13:27:38.355849 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e1bffb7a-a33a-466d-b382-a6896e83d8d9-inventory\") pod \"e1bffb7a-a33a-466d-b382-a6896e83d8d9\" (UID: \"e1bffb7a-a33a-466d-b382-a6896e83d8d9\") " Feb 17 13:27:38 crc kubenswrapper[4955]: I0217 13:27:38.355875 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e1bffb7a-a33a-466d-b382-a6896e83d8d9-ssh-key-openstack-edpm-ipam\") pod \"e1bffb7a-a33a-466d-b382-a6896e83d8d9\" (UID: \"e1bffb7a-a33a-466d-b382-a6896e83d8d9\") " Feb 17 13:27:38 crc kubenswrapper[4955]: I0217 13:27:38.358971 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbggb\" (UniqueName: \"kubernetes.io/projected/e1bffb7a-a33a-466d-b382-a6896e83d8d9-kube-api-access-jbggb\") pod \"e1bffb7a-a33a-466d-b382-a6896e83d8d9\" (UID: \"e1bffb7a-a33a-466d-b382-a6896e83d8d9\") " Feb 17 13:27:38 crc kubenswrapper[4955]: I0217 13:27:38.365900 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1bffb7a-a33a-466d-b382-a6896e83d8d9-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "e1bffb7a-a33a-466d-b382-a6896e83d8d9" (UID: "e1bffb7a-a33a-466d-b382-a6896e83d8d9"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:27:38 crc kubenswrapper[4955]: I0217 13:27:38.366293 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1bffb7a-a33a-466d-b382-a6896e83d8d9-kube-api-access-jbggb" (OuterVolumeSpecName: "kube-api-access-jbggb") pod "e1bffb7a-a33a-466d-b382-a6896e83d8d9" (UID: "e1bffb7a-a33a-466d-b382-a6896e83d8d9"). InnerVolumeSpecName "kube-api-access-jbggb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:27:38 crc kubenswrapper[4955]: I0217 13:27:38.388642 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1bffb7a-a33a-466d-b382-a6896e83d8d9-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "e1bffb7a-a33a-466d-b382-a6896e83d8d9" (UID: "e1bffb7a-a33a-466d-b382-a6896e83d8d9"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:27:38 crc kubenswrapper[4955]: I0217 13:27:38.391631 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1bffb7a-a33a-466d-b382-a6896e83d8d9-inventory" (OuterVolumeSpecName: "inventory") pod "e1bffb7a-a33a-466d-b382-a6896e83d8d9" (UID: "e1bffb7a-a33a-466d-b382-a6896e83d8d9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:27:38 crc kubenswrapper[4955]: I0217 13:27:38.461852 4955 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1bffb7a-a33a-466d-b382-a6896e83d8d9-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:27:38 crc kubenswrapper[4955]: I0217 13:27:38.461889 4955 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e1bffb7a-a33a-466d-b382-a6896e83d8d9-inventory\") on node \"crc\" DevicePath \"\"" Feb 17 13:27:38 crc kubenswrapper[4955]: I0217 13:27:38.461901 4955 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e1bffb7a-a33a-466d-b382-a6896e83d8d9-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 17 13:27:38 crc kubenswrapper[4955]: I0217 13:27:38.461910 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbggb\" (UniqueName: \"kubernetes.io/projected/e1bffb7a-a33a-466d-b382-a6896e83d8d9-kube-api-access-jbggb\") on node \"crc\" DevicePath \"\"" Feb 17 13:27:38 crc kubenswrapper[4955]: I0217 13:27:38.914876 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g2nhl" event={"ID":"e1bffb7a-a33a-466d-b382-a6896e83d8d9","Type":"ContainerDied","Data":"b6180b12f7dc882baace2bb4318354b22234697c51930a409be4dda5eba608d9"} Feb 17 13:27:38 crc kubenswrapper[4955]: I0217 13:27:38.914926 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b6180b12f7dc882baace2bb4318354b22234697c51930a409be4dda5eba608d9" Feb 17 13:27:38 crc kubenswrapper[4955]: I0217 13:27:38.914932 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g2nhl" Feb 17 13:27:39 crc kubenswrapper[4955]: I0217 13:27:39.007290 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-cv8cz"] Feb 17 13:27:39 crc kubenswrapper[4955]: E0217 13:27:39.007692 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1bffb7a-a33a-466d-b382-a6896e83d8d9" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 17 13:27:39 crc kubenswrapper[4955]: I0217 13:27:39.007710 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1bffb7a-a33a-466d-b382-a6896e83d8d9" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 17 13:27:39 crc kubenswrapper[4955]: I0217 13:27:39.007923 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1bffb7a-a33a-466d-b382-a6896e83d8d9" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 17 13:27:39 crc kubenswrapper[4955]: I0217 13:27:39.008624 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-cv8cz" Feb 17 13:27:39 crc kubenswrapper[4955]: I0217 13:27:39.012436 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 17 13:27:39 crc kubenswrapper[4955]: I0217 13:27:39.012489 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 17 13:27:39 crc kubenswrapper[4955]: I0217 13:27:39.012509 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-t9lgc" Feb 17 13:27:39 crc kubenswrapper[4955]: I0217 13:27:39.012661 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 17 13:27:39 crc kubenswrapper[4955]: I0217 13:27:39.024743 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-cv8cz"] Feb 17 13:27:39 crc kubenswrapper[4955]: I0217 13:27:39.073507 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8e826189-c723-4c9e-8ab8-3e60c06b7e42-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-cv8cz\" (UID: \"8e826189-c723-4c9e-8ab8-3e60c06b7e42\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-cv8cz" Feb 17 13:27:39 crc kubenswrapper[4955]: I0217 13:27:39.073853 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8e826189-c723-4c9e-8ab8-3e60c06b7e42-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-cv8cz\" (UID: \"8e826189-c723-4c9e-8ab8-3e60c06b7e42\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-cv8cz" Feb 17 13:27:39 crc kubenswrapper[4955]: I0217 13:27:39.073999 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8x7ld\" (UniqueName: \"kubernetes.io/projected/8e826189-c723-4c9e-8ab8-3e60c06b7e42-kube-api-access-8x7ld\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-cv8cz\" (UID: \"8e826189-c723-4c9e-8ab8-3e60c06b7e42\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-cv8cz" Feb 17 13:27:39 crc kubenswrapper[4955]: I0217 13:27:39.176129 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8e826189-c723-4c9e-8ab8-3e60c06b7e42-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-cv8cz\" (UID: \"8e826189-c723-4c9e-8ab8-3e60c06b7e42\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-cv8cz" Feb 17 13:27:39 crc kubenswrapper[4955]: I0217 13:27:39.176365 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8e826189-c723-4c9e-8ab8-3e60c06b7e42-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-cv8cz\" (UID: \"8e826189-c723-4c9e-8ab8-3e60c06b7e42\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-cv8cz" Feb 17 13:27:39 crc kubenswrapper[4955]: I0217 13:27:39.176426 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8x7ld\" (UniqueName: \"kubernetes.io/projected/8e826189-c723-4c9e-8ab8-3e60c06b7e42-kube-api-access-8x7ld\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-cv8cz\" (UID: \"8e826189-c723-4c9e-8ab8-3e60c06b7e42\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-cv8cz" Feb 17 13:27:39 crc kubenswrapper[4955]: I0217 13:27:39.180585 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8e826189-c723-4c9e-8ab8-3e60c06b7e42-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-cv8cz\" (UID: \"8e826189-c723-4c9e-8ab8-3e60c06b7e42\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-cv8cz" Feb 17 13:27:39 crc kubenswrapper[4955]: I0217 13:27:39.180682 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8e826189-c723-4c9e-8ab8-3e60c06b7e42-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-cv8cz\" (UID: \"8e826189-c723-4c9e-8ab8-3e60c06b7e42\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-cv8cz" Feb 17 13:27:39 crc kubenswrapper[4955]: I0217 13:27:39.194806 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8x7ld\" (UniqueName: \"kubernetes.io/projected/8e826189-c723-4c9e-8ab8-3e60c06b7e42-kube-api-access-8x7ld\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-cv8cz\" (UID: \"8e826189-c723-4c9e-8ab8-3e60c06b7e42\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-cv8cz" Feb 17 13:27:39 crc kubenswrapper[4955]: I0217 13:27:39.287007 4955 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-jqmrl" podUID="c485bbf2-8d96-4c64-a2c8-37c76c557d4f" containerName="registry-server" probeResult="failure" output=< Feb 17 13:27:39 crc kubenswrapper[4955]: timeout: failed to connect service ":50051" within 1s Feb 17 13:27:39 crc kubenswrapper[4955]: > Feb 17 13:27:39 crc kubenswrapper[4955]: I0217 13:27:39.324682 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-cv8cz" Feb 17 13:27:39 crc kubenswrapper[4955]: I0217 13:27:39.916800 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-cv8cz"] Feb 17 13:27:39 crc kubenswrapper[4955]: W0217 13:27:39.918018 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8e826189_c723_4c9e_8ab8_3e60c06b7e42.slice/crio-dc4bca4bfddbffb763d9f1d95ed0a9fd30746fd57502d2a7c59afa323d36af34 WatchSource:0}: Error finding container dc4bca4bfddbffb763d9f1d95ed0a9fd30746fd57502d2a7c59afa323d36af34: Status 404 returned error can't find the container with id dc4bca4bfddbffb763d9f1d95ed0a9fd30746fd57502d2a7c59afa323d36af34 Feb 17 13:27:40 crc kubenswrapper[4955]: I0217 13:27:40.932535 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-cv8cz" event={"ID":"8e826189-c723-4c9e-8ab8-3e60c06b7e42","Type":"ContainerStarted","Data":"5d812d4237bd92e1562e27a6cf92f0da7fba7fc172d08027f56e5c077fa0f76f"} Feb 17 13:27:40 crc kubenswrapper[4955]: I0217 13:27:40.934138 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-cv8cz" event={"ID":"8e826189-c723-4c9e-8ab8-3e60c06b7e42","Type":"ContainerStarted","Data":"dc4bca4bfddbffb763d9f1d95ed0a9fd30746fd57502d2a7c59afa323d36af34"} Feb 17 13:27:40 crc kubenswrapper[4955]: I0217 13:27:40.956340 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-cv8cz" podStartSLOduration=2.4302499060000002 podStartE2EDuration="2.956323398s" podCreationTimestamp="2026-02-17 13:27:38 +0000 UTC" firstStartedPulling="2026-02-17 13:27:39.922593944 +0000 UTC m=+1398.445323487" lastFinishedPulling="2026-02-17 13:27:40.448667396 +0000 UTC m=+1398.971396979" observedRunningTime="2026-02-17 13:27:40.948285048 +0000 UTC m=+1399.471014591" watchObservedRunningTime="2026-02-17 13:27:40.956323398 +0000 UTC m=+1399.479052941" Feb 17 13:27:41 crc kubenswrapper[4955]: I0217 13:27:41.307202 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dwd4j"] Feb 17 13:27:41 crc kubenswrapper[4955]: I0217 13:27:41.309673 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dwd4j" Feb 17 13:27:41 crc kubenswrapper[4955]: I0217 13:27:41.318949 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dwd4j"] Feb 17 13:27:41 crc kubenswrapper[4955]: I0217 13:27:41.415187 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6f8w\" (UniqueName: \"kubernetes.io/projected/512586cb-2c3c-4332-9f24-875632c4f22e-kube-api-access-h6f8w\") pod \"redhat-marketplace-dwd4j\" (UID: \"512586cb-2c3c-4332-9f24-875632c4f22e\") " pod="openshift-marketplace/redhat-marketplace-dwd4j" Feb 17 13:27:41 crc kubenswrapper[4955]: I0217 13:27:41.415276 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/512586cb-2c3c-4332-9f24-875632c4f22e-catalog-content\") pod \"redhat-marketplace-dwd4j\" (UID: \"512586cb-2c3c-4332-9f24-875632c4f22e\") " pod="openshift-marketplace/redhat-marketplace-dwd4j" Feb 17 13:27:41 crc kubenswrapper[4955]: I0217 13:27:41.415326 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/512586cb-2c3c-4332-9f24-875632c4f22e-utilities\") pod \"redhat-marketplace-dwd4j\" (UID: \"512586cb-2c3c-4332-9f24-875632c4f22e\") " pod="openshift-marketplace/redhat-marketplace-dwd4j" Feb 17 13:27:41 crc kubenswrapper[4955]: I0217 13:27:41.516344 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6f8w\" (UniqueName: \"kubernetes.io/projected/512586cb-2c3c-4332-9f24-875632c4f22e-kube-api-access-h6f8w\") pod \"redhat-marketplace-dwd4j\" (UID: \"512586cb-2c3c-4332-9f24-875632c4f22e\") " pod="openshift-marketplace/redhat-marketplace-dwd4j" Feb 17 13:27:41 crc kubenswrapper[4955]: I0217 13:27:41.516403 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/512586cb-2c3c-4332-9f24-875632c4f22e-catalog-content\") pod \"redhat-marketplace-dwd4j\" (UID: \"512586cb-2c3c-4332-9f24-875632c4f22e\") " pod="openshift-marketplace/redhat-marketplace-dwd4j" Feb 17 13:27:41 crc kubenswrapper[4955]: I0217 13:27:41.516446 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/512586cb-2c3c-4332-9f24-875632c4f22e-utilities\") pod \"redhat-marketplace-dwd4j\" (UID: \"512586cb-2c3c-4332-9f24-875632c4f22e\") " pod="openshift-marketplace/redhat-marketplace-dwd4j" Feb 17 13:27:41 crc kubenswrapper[4955]: I0217 13:27:41.516881 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/512586cb-2c3c-4332-9f24-875632c4f22e-catalog-content\") pod \"redhat-marketplace-dwd4j\" (UID: \"512586cb-2c3c-4332-9f24-875632c4f22e\") " pod="openshift-marketplace/redhat-marketplace-dwd4j" Feb 17 13:27:41 crc kubenswrapper[4955]: I0217 13:27:41.518402 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/512586cb-2c3c-4332-9f24-875632c4f22e-utilities\") pod \"redhat-marketplace-dwd4j\" (UID: \"512586cb-2c3c-4332-9f24-875632c4f22e\") " pod="openshift-marketplace/redhat-marketplace-dwd4j" Feb 17 13:27:41 crc kubenswrapper[4955]: I0217 13:27:41.535364 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6f8w\" (UniqueName: \"kubernetes.io/projected/512586cb-2c3c-4332-9f24-875632c4f22e-kube-api-access-h6f8w\") pod \"redhat-marketplace-dwd4j\" (UID: \"512586cb-2c3c-4332-9f24-875632c4f22e\") " pod="openshift-marketplace/redhat-marketplace-dwd4j" Feb 17 13:27:41 crc kubenswrapper[4955]: I0217 13:27:41.644722 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dwd4j" Feb 17 13:27:42 crc kubenswrapper[4955]: I0217 13:27:42.127222 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dwd4j"] Feb 17 13:27:42 crc kubenswrapper[4955]: I0217 13:27:42.956510 4955 generic.go:334] "Generic (PLEG): container finished" podID="512586cb-2c3c-4332-9f24-875632c4f22e" containerID="e6874c28d872efe0e26677b3f2192516abca9d04f247e129d8e0334acd576089" exitCode=0 Feb 17 13:27:42 crc kubenswrapper[4955]: I0217 13:27:42.956583 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dwd4j" event={"ID":"512586cb-2c3c-4332-9f24-875632c4f22e","Type":"ContainerDied","Data":"e6874c28d872efe0e26677b3f2192516abca9d04f247e129d8e0334acd576089"} Feb 17 13:27:42 crc kubenswrapper[4955]: I0217 13:27:42.956882 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dwd4j" event={"ID":"512586cb-2c3c-4332-9f24-875632c4f22e","Type":"ContainerStarted","Data":"b0b63b33cef043b0b4390f8c8b86faa3b8e7a37f1dd53afc3839b03775de858f"} Feb 17 13:27:43 crc kubenswrapper[4955]: I0217 13:27:43.970041 4955 generic.go:334] "Generic (PLEG): container finished" podID="8e826189-c723-4c9e-8ab8-3e60c06b7e42" containerID="5d812d4237bd92e1562e27a6cf92f0da7fba7fc172d08027f56e5c077fa0f76f" exitCode=0 Feb 17 13:27:43 crc kubenswrapper[4955]: I0217 13:27:43.970303 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-cv8cz" event={"ID":"8e826189-c723-4c9e-8ab8-3e60c06b7e42","Type":"ContainerDied","Data":"5d812d4237bd92e1562e27a6cf92f0da7fba7fc172d08027f56e5c077fa0f76f"} Feb 17 13:27:44 crc kubenswrapper[4955]: I0217 13:27:44.982722 4955 generic.go:334] "Generic (PLEG): container finished" podID="512586cb-2c3c-4332-9f24-875632c4f22e" containerID="2a48bb5a87a4190a6c784c8700b0ea717d264f941af9c823cfec070240e05545" exitCode=0 Feb 17 13:27:44 crc kubenswrapper[4955]: I0217 13:27:44.982827 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dwd4j" event={"ID":"512586cb-2c3c-4332-9f24-875632c4f22e","Type":"ContainerDied","Data":"2a48bb5a87a4190a6c784c8700b0ea717d264f941af9c823cfec070240e05545"} Feb 17 13:27:45 crc kubenswrapper[4955]: I0217 13:27:45.490587 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-cv8cz" Feb 17 13:27:45 crc kubenswrapper[4955]: I0217 13:27:45.591296 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8e826189-c723-4c9e-8ab8-3e60c06b7e42-ssh-key-openstack-edpm-ipam\") pod \"8e826189-c723-4c9e-8ab8-3e60c06b7e42\" (UID: \"8e826189-c723-4c9e-8ab8-3e60c06b7e42\") " Feb 17 13:27:45 crc kubenswrapper[4955]: I0217 13:27:45.591425 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8x7ld\" (UniqueName: \"kubernetes.io/projected/8e826189-c723-4c9e-8ab8-3e60c06b7e42-kube-api-access-8x7ld\") pod \"8e826189-c723-4c9e-8ab8-3e60c06b7e42\" (UID: \"8e826189-c723-4c9e-8ab8-3e60c06b7e42\") " Feb 17 13:27:45 crc kubenswrapper[4955]: I0217 13:27:45.591497 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8e826189-c723-4c9e-8ab8-3e60c06b7e42-inventory\") pod \"8e826189-c723-4c9e-8ab8-3e60c06b7e42\" (UID: \"8e826189-c723-4c9e-8ab8-3e60c06b7e42\") " Feb 17 13:27:45 crc kubenswrapper[4955]: I0217 13:27:45.598310 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e826189-c723-4c9e-8ab8-3e60c06b7e42-kube-api-access-8x7ld" (OuterVolumeSpecName: "kube-api-access-8x7ld") pod "8e826189-c723-4c9e-8ab8-3e60c06b7e42" (UID: "8e826189-c723-4c9e-8ab8-3e60c06b7e42"). InnerVolumeSpecName "kube-api-access-8x7ld". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:27:45 crc kubenswrapper[4955]: I0217 13:27:45.622723 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e826189-c723-4c9e-8ab8-3e60c06b7e42-inventory" (OuterVolumeSpecName: "inventory") pod "8e826189-c723-4c9e-8ab8-3e60c06b7e42" (UID: "8e826189-c723-4c9e-8ab8-3e60c06b7e42"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:27:45 crc kubenswrapper[4955]: I0217 13:27:45.625380 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e826189-c723-4c9e-8ab8-3e60c06b7e42-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "8e826189-c723-4c9e-8ab8-3e60c06b7e42" (UID: "8e826189-c723-4c9e-8ab8-3e60c06b7e42"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:27:45 crc kubenswrapper[4955]: I0217 13:27:45.692670 4955 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8e826189-c723-4c9e-8ab8-3e60c06b7e42-inventory\") on node \"crc\" DevicePath \"\"" Feb 17 13:27:45 crc kubenswrapper[4955]: I0217 13:27:45.692711 4955 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8e826189-c723-4c9e-8ab8-3e60c06b7e42-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 17 13:27:45 crc kubenswrapper[4955]: I0217 13:27:45.692722 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8x7ld\" (UniqueName: \"kubernetes.io/projected/8e826189-c723-4c9e-8ab8-3e60c06b7e42-kube-api-access-8x7ld\") on node \"crc\" DevicePath \"\"" Feb 17 13:27:45 crc kubenswrapper[4955]: I0217 13:27:45.997437 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-cv8cz" event={"ID":"8e826189-c723-4c9e-8ab8-3e60c06b7e42","Type":"ContainerDied","Data":"dc4bca4bfddbffb763d9f1d95ed0a9fd30746fd57502d2a7c59afa323d36af34"} Feb 17 13:27:45 crc kubenswrapper[4955]: I0217 13:27:45.997806 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc4bca4bfddbffb763d9f1d95ed0a9fd30746fd57502d2a7c59afa323d36af34" Feb 17 13:27:45 crc kubenswrapper[4955]: I0217 13:27:45.997463 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-cv8cz" Feb 17 13:27:46 crc kubenswrapper[4955]: I0217 13:27:46.001487 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dwd4j" event={"ID":"512586cb-2c3c-4332-9f24-875632c4f22e","Type":"ContainerStarted","Data":"62b60d02bbff7bf65ed6020843b3a571666ac5ec53123e7af2f367e7f8cdd1fe"} Feb 17 13:27:46 crc kubenswrapper[4955]: I0217 13:27:46.064583 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dwd4j" podStartSLOduration=2.640432098 podStartE2EDuration="5.064561523s" podCreationTimestamp="2026-02-17 13:27:41 +0000 UTC" firstStartedPulling="2026-02-17 13:27:42.958507722 +0000 UTC m=+1401.481237265" lastFinishedPulling="2026-02-17 13:27:45.382637137 +0000 UTC m=+1403.905366690" observedRunningTime="2026-02-17 13:27:46.035740705 +0000 UTC m=+1404.558470258" watchObservedRunningTime="2026-02-17 13:27:46.064561523 +0000 UTC m=+1404.587291066" Feb 17 13:27:46 crc kubenswrapper[4955]: I0217 13:27:46.065052 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn5l6"] Feb 17 13:27:46 crc kubenswrapper[4955]: E0217 13:27:46.070735 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e826189-c723-4c9e-8ab8-3e60c06b7e42" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Feb 17 13:27:46 crc kubenswrapper[4955]: I0217 13:27:46.070770 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e826189-c723-4c9e-8ab8-3e60c06b7e42" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Feb 17 13:27:46 crc kubenswrapper[4955]: I0217 13:27:46.071074 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e826189-c723-4c9e-8ab8-3e60c06b7e42" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Feb 17 13:27:46 crc kubenswrapper[4955]: I0217 13:27:46.071849 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn5l6" Feb 17 13:27:46 crc kubenswrapper[4955]: I0217 13:27:46.074423 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 17 13:27:46 crc kubenswrapper[4955]: I0217 13:27:46.074434 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 17 13:27:46 crc kubenswrapper[4955]: I0217 13:27:46.074729 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-t9lgc" Feb 17 13:27:46 crc kubenswrapper[4955]: I0217 13:27:46.074927 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 17 13:27:46 crc kubenswrapper[4955]: I0217 13:27:46.095061 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn5l6"] Feb 17 13:27:46 crc kubenswrapper[4955]: I0217 13:27:46.098674 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea41b3c5-2c55-4294-b2dc-a25ab083c780-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zn5l6\" (UID: \"ea41b3c5-2c55-4294-b2dc-a25ab083c780\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn5l6" Feb 17 13:27:46 crc kubenswrapper[4955]: I0217 13:27:46.098728 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ea41b3c5-2c55-4294-b2dc-a25ab083c780-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zn5l6\" (UID: \"ea41b3c5-2c55-4294-b2dc-a25ab083c780\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn5l6" Feb 17 13:27:46 crc kubenswrapper[4955]: I0217 13:27:46.098793 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cx86f\" (UniqueName: \"kubernetes.io/projected/ea41b3c5-2c55-4294-b2dc-a25ab083c780-kube-api-access-cx86f\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zn5l6\" (UID: \"ea41b3c5-2c55-4294-b2dc-a25ab083c780\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn5l6" Feb 17 13:27:46 crc kubenswrapper[4955]: I0217 13:27:46.098850 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ea41b3c5-2c55-4294-b2dc-a25ab083c780-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zn5l6\" (UID: \"ea41b3c5-2c55-4294-b2dc-a25ab083c780\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn5l6" Feb 17 13:27:46 crc kubenswrapper[4955]: I0217 13:27:46.200468 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ea41b3c5-2c55-4294-b2dc-a25ab083c780-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zn5l6\" (UID: \"ea41b3c5-2c55-4294-b2dc-a25ab083c780\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn5l6" Feb 17 13:27:46 crc kubenswrapper[4955]: I0217 13:27:46.200613 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cx86f\" (UniqueName: \"kubernetes.io/projected/ea41b3c5-2c55-4294-b2dc-a25ab083c780-kube-api-access-cx86f\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zn5l6\" (UID: \"ea41b3c5-2c55-4294-b2dc-a25ab083c780\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn5l6" Feb 17 13:27:46 crc kubenswrapper[4955]: I0217 13:27:46.201111 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ea41b3c5-2c55-4294-b2dc-a25ab083c780-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zn5l6\" (UID: \"ea41b3c5-2c55-4294-b2dc-a25ab083c780\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn5l6" Feb 17 13:27:46 crc kubenswrapper[4955]: I0217 13:27:46.201695 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea41b3c5-2c55-4294-b2dc-a25ab083c780-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zn5l6\" (UID: \"ea41b3c5-2c55-4294-b2dc-a25ab083c780\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn5l6" Feb 17 13:27:46 crc kubenswrapper[4955]: I0217 13:27:46.205571 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea41b3c5-2c55-4294-b2dc-a25ab083c780-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zn5l6\" (UID: \"ea41b3c5-2c55-4294-b2dc-a25ab083c780\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn5l6" Feb 17 13:27:46 crc kubenswrapper[4955]: I0217 13:27:46.205569 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ea41b3c5-2c55-4294-b2dc-a25ab083c780-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zn5l6\" (UID: \"ea41b3c5-2c55-4294-b2dc-a25ab083c780\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn5l6" Feb 17 13:27:46 crc kubenswrapper[4955]: I0217 13:27:46.205826 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ea41b3c5-2c55-4294-b2dc-a25ab083c780-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zn5l6\" (UID: \"ea41b3c5-2c55-4294-b2dc-a25ab083c780\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn5l6" Feb 17 13:27:46 crc kubenswrapper[4955]: I0217 13:27:46.220185 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cx86f\" (UniqueName: \"kubernetes.io/projected/ea41b3c5-2c55-4294-b2dc-a25ab083c780-kube-api-access-cx86f\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zn5l6\" (UID: \"ea41b3c5-2c55-4294-b2dc-a25ab083c780\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn5l6" Feb 17 13:27:46 crc kubenswrapper[4955]: I0217 13:27:46.428249 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn5l6" Feb 17 13:27:47 crc kubenswrapper[4955]: I0217 13:27:47.042530 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn5l6"] Feb 17 13:27:47 crc kubenswrapper[4955]: I0217 13:27:47.916091 4955 scope.go:117] "RemoveContainer" containerID="a91687d4b5b34d2b759f39ea553342f9cc3938a66b65c3b62191bbffbeebb176" Feb 17 13:27:47 crc kubenswrapper[4955]: I0217 13:27:47.942856 4955 scope.go:117] "RemoveContainer" containerID="977218e0be442828b357f3cabb51b2efffd7ac393678959c05c0b2f99da8b6aa" Feb 17 13:27:47 crc kubenswrapper[4955]: I0217 13:27:47.985031 4955 scope.go:117] "RemoveContainer" containerID="86273f17c38a0753b27e85668517d95a43d5eda6373742707eac860fe4973011" Feb 17 13:27:48 crc kubenswrapper[4955]: I0217 13:27:48.017964 4955 scope.go:117] "RemoveContainer" containerID="ade39e7ed0dc4c3294ee757df6a6baf431f8c1c351a564d153d5f82cf0896b64" Feb 17 13:27:48 crc kubenswrapper[4955]: I0217 13:27:48.022953 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn5l6" event={"ID":"ea41b3c5-2c55-4294-b2dc-a25ab083c780","Type":"ContainerStarted","Data":"9cc3799076b06fd1ced08eb820730cb665937dc35990eccfe214867e77a09f72"} Feb 17 13:27:48 crc kubenswrapper[4955]: I0217 13:27:48.022997 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn5l6" event={"ID":"ea41b3c5-2c55-4294-b2dc-a25ab083c780","Type":"ContainerStarted","Data":"e9e9d821dc9bde95c915b0c4ed1863824c8bf723119a367f24bdadc829f1d97e"} Feb 17 13:27:48 crc kubenswrapper[4955]: I0217 13:27:48.044664 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn5l6" podStartSLOduration=1.617466589 podStartE2EDuration="2.044647382s" podCreationTimestamp="2026-02-17 13:27:46 +0000 UTC" firstStartedPulling="2026-02-17 13:27:47.040280741 +0000 UTC m=+1405.563010284" lastFinishedPulling="2026-02-17 13:27:47.467461534 +0000 UTC m=+1405.990191077" observedRunningTime="2026-02-17 13:27:48.03619452 +0000 UTC m=+1406.558924083" watchObservedRunningTime="2026-02-17 13:27:48.044647382 +0000 UTC m=+1406.567376925" Feb 17 13:27:48 crc kubenswrapper[4955]: I0217 13:27:48.294738 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jqmrl" Feb 17 13:27:48 crc kubenswrapper[4955]: I0217 13:27:48.366900 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jqmrl" Feb 17 13:27:48 crc kubenswrapper[4955]: I0217 13:27:48.693365 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jqmrl"] Feb 17 13:27:50 crc kubenswrapper[4955]: I0217 13:27:50.057739 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jqmrl" podUID="c485bbf2-8d96-4c64-a2c8-37c76c557d4f" containerName="registry-server" containerID="cri-o://1ca5a8bff6032c823bd77a4179eb8c586e2958762a10dc9dd08e01fd48e6a663" gracePeriod=2 Feb 17 13:27:50 crc kubenswrapper[4955]: I0217 13:27:50.544624 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jqmrl" Feb 17 13:27:50 crc kubenswrapper[4955]: I0217 13:27:50.716727 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2nrsq\" (UniqueName: \"kubernetes.io/projected/c485bbf2-8d96-4c64-a2c8-37c76c557d4f-kube-api-access-2nrsq\") pod \"c485bbf2-8d96-4c64-a2c8-37c76c557d4f\" (UID: \"c485bbf2-8d96-4c64-a2c8-37c76c557d4f\") " Feb 17 13:27:50 crc kubenswrapper[4955]: I0217 13:27:50.717254 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c485bbf2-8d96-4c64-a2c8-37c76c557d4f-catalog-content\") pod \"c485bbf2-8d96-4c64-a2c8-37c76c557d4f\" (UID: \"c485bbf2-8d96-4c64-a2c8-37c76c557d4f\") " Feb 17 13:27:50 crc kubenswrapper[4955]: I0217 13:27:50.717342 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c485bbf2-8d96-4c64-a2c8-37c76c557d4f-utilities\") pod \"c485bbf2-8d96-4c64-a2c8-37c76c557d4f\" (UID: \"c485bbf2-8d96-4c64-a2c8-37c76c557d4f\") " Feb 17 13:27:50 crc kubenswrapper[4955]: I0217 13:27:50.718459 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c485bbf2-8d96-4c64-a2c8-37c76c557d4f-utilities" (OuterVolumeSpecName: "utilities") pod "c485bbf2-8d96-4c64-a2c8-37c76c557d4f" (UID: "c485bbf2-8d96-4c64-a2c8-37c76c557d4f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:27:50 crc kubenswrapper[4955]: I0217 13:27:50.724411 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c485bbf2-8d96-4c64-a2c8-37c76c557d4f-kube-api-access-2nrsq" (OuterVolumeSpecName: "kube-api-access-2nrsq") pod "c485bbf2-8d96-4c64-a2c8-37c76c557d4f" (UID: "c485bbf2-8d96-4c64-a2c8-37c76c557d4f"). InnerVolumeSpecName "kube-api-access-2nrsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:27:50 crc kubenswrapper[4955]: I0217 13:27:50.776870 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c485bbf2-8d96-4c64-a2c8-37c76c557d4f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c485bbf2-8d96-4c64-a2c8-37c76c557d4f" (UID: "c485bbf2-8d96-4c64-a2c8-37c76c557d4f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:27:50 crc kubenswrapper[4955]: I0217 13:27:50.820083 4955 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c485bbf2-8d96-4c64-a2c8-37c76c557d4f-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 13:27:50 crc kubenswrapper[4955]: I0217 13:27:50.820121 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2nrsq\" (UniqueName: \"kubernetes.io/projected/c485bbf2-8d96-4c64-a2c8-37c76c557d4f-kube-api-access-2nrsq\") on node \"crc\" DevicePath \"\"" Feb 17 13:27:50 crc kubenswrapper[4955]: I0217 13:27:50.820140 4955 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c485bbf2-8d96-4c64-a2c8-37c76c557d4f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 13:27:51 crc kubenswrapper[4955]: I0217 13:27:51.067023 4955 generic.go:334] "Generic (PLEG): container finished" podID="c485bbf2-8d96-4c64-a2c8-37c76c557d4f" containerID="1ca5a8bff6032c823bd77a4179eb8c586e2958762a10dc9dd08e01fd48e6a663" exitCode=0 Feb 17 13:27:51 crc kubenswrapper[4955]: I0217 13:27:51.067062 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jqmrl" event={"ID":"c485bbf2-8d96-4c64-a2c8-37c76c557d4f","Type":"ContainerDied","Data":"1ca5a8bff6032c823bd77a4179eb8c586e2958762a10dc9dd08e01fd48e6a663"} Feb 17 13:27:51 crc kubenswrapper[4955]: I0217 13:27:51.067093 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jqmrl" event={"ID":"c485bbf2-8d96-4c64-a2c8-37c76c557d4f","Type":"ContainerDied","Data":"66e37d9c78566a74b52756288f79204aaaa35449e79304f6e468e207b93dd154"} Feb 17 13:27:51 crc kubenswrapper[4955]: I0217 13:27:51.067112 4955 scope.go:117] "RemoveContainer" containerID="1ca5a8bff6032c823bd77a4179eb8c586e2958762a10dc9dd08e01fd48e6a663" Feb 17 13:27:51 crc kubenswrapper[4955]: I0217 13:27:51.067132 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jqmrl" Feb 17 13:27:51 crc kubenswrapper[4955]: I0217 13:27:51.106648 4955 scope.go:117] "RemoveContainer" containerID="b5ea2958904dc0f2d17fad897b4bcc0032985cb7fd2e026fb041a3718bd91384" Feb 17 13:27:51 crc kubenswrapper[4955]: I0217 13:27:51.138473 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jqmrl"] Feb 17 13:27:51 crc kubenswrapper[4955]: I0217 13:27:51.140618 4955 scope.go:117] "RemoveContainer" containerID="6e4966ac31e08553542beabcacb077ab467cbb2f7eb62826dce3f010f81b8c3f" Feb 17 13:27:51 crc kubenswrapper[4955]: I0217 13:27:51.146909 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jqmrl"] Feb 17 13:27:51 crc kubenswrapper[4955]: I0217 13:27:51.178583 4955 scope.go:117] "RemoveContainer" containerID="1ca5a8bff6032c823bd77a4179eb8c586e2958762a10dc9dd08e01fd48e6a663" Feb 17 13:27:51 crc kubenswrapper[4955]: E0217 13:27:51.179097 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ca5a8bff6032c823bd77a4179eb8c586e2958762a10dc9dd08e01fd48e6a663\": container with ID starting with 1ca5a8bff6032c823bd77a4179eb8c586e2958762a10dc9dd08e01fd48e6a663 not found: ID does not exist" containerID="1ca5a8bff6032c823bd77a4179eb8c586e2958762a10dc9dd08e01fd48e6a663" Feb 17 13:27:51 crc kubenswrapper[4955]: I0217 13:27:51.179138 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ca5a8bff6032c823bd77a4179eb8c586e2958762a10dc9dd08e01fd48e6a663"} err="failed to get container status \"1ca5a8bff6032c823bd77a4179eb8c586e2958762a10dc9dd08e01fd48e6a663\": rpc error: code = NotFound desc = could not find container \"1ca5a8bff6032c823bd77a4179eb8c586e2958762a10dc9dd08e01fd48e6a663\": container with ID starting with 1ca5a8bff6032c823bd77a4179eb8c586e2958762a10dc9dd08e01fd48e6a663 not found: ID does not exist" Feb 17 13:27:51 crc kubenswrapper[4955]: I0217 13:27:51.179183 4955 scope.go:117] "RemoveContainer" containerID="b5ea2958904dc0f2d17fad897b4bcc0032985cb7fd2e026fb041a3718bd91384" Feb 17 13:27:51 crc kubenswrapper[4955]: E0217 13:27:51.179524 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5ea2958904dc0f2d17fad897b4bcc0032985cb7fd2e026fb041a3718bd91384\": container with ID starting with b5ea2958904dc0f2d17fad897b4bcc0032985cb7fd2e026fb041a3718bd91384 not found: ID does not exist" containerID="b5ea2958904dc0f2d17fad897b4bcc0032985cb7fd2e026fb041a3718bd91384" Feb 17 13:27:51 crc kubenswrapper[4955]: I0217 13:27:51.179595 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5ea2958904dc0f2d17fad897b4bcc0032985cb7fd2e026fb041a3718bd91384"} err="failed to get container status \"b5ea2958904dc0f2d17fad897b4bcc0032985cb7fd2e026fb041a3718bd91384\": rpc error: code = NotFound desc = could not find container \"b5ea2958904dc0f2d17fad897b4bcc0032985cb7fd2e026fb041a3718bd91384\": container with ID starting with b5ea2958904dc0f2d17fad897b4bcc0032985cb7fd2e026fb041a3718bd91384 not found: ID does not exist" Feb 17 13:27:51 crc kubenswrapper[4955]: I0217 13:27:51.179613 4955 scope.go:117] "RemoveContainer" containerID="6e4966ac31e08553542beabcacb077ab467cbb2f7eb62826dce3f010f81b8c3f" Feb 17 13:27:51 crc kubenswrapper[4955]: E0217 13:27:51.179871 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e4966ac31e08553542beabcacb077ab467cbb2f7eb62826dce3f010f81b8c3f\": container with ID starting with 6e4966ac31e08553542beabcacb077ab467cbb2f7eb62826dce3f010f81b8c3f not found: ID does not exist" containerID="6e4966ac31e08553542beabcacb077ab467cbb2f7eb62826dce3f010f81b8c3f" Feb 17 13:27:51 crc kubenswrapper[4955]: I0217 13:27:51.179898 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e4966ac31e08553542beabcacb077ab467cbb2f7eb62826dce3f010f81b8c3f"} err="failed to get container status \"6e4966ac31e08553542beabcacb077ab467cbb2f7eb62826dce3f010f81b8c3f\": rpc error: code = NotFound desc = could not find container \"6e4966ac31e08553542beabcacb077ab467cbb2f7eb62826dce3f010f81b8c3f\": container with ID starting with 6e4966ac31e08553542beabcacb077ab467cbb2f7eb62826dce3f010f81b8c3f not found: ID does not exist" Feb 17 13:27:51 crc kubenswrapper[4955]: I0217 13:27:51.645095 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dwd4j" Feb 17 13:27:51 crc kubenswrapper[4955]: I0217 13:27:51.645357 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dwd4j" Feb 17 13:27:51 crc kubenswrapper[4955]: I0217 13:27:51.700429 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dwd4j" Feb 17 13:27:52 crc kubenswrapper[4955]: I0217 13:27:52.128532 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dwd4j" Feb 17 13:27:52 crc kubenswrapper[4955]: I0217 13:27:52.234599 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c485bbf2-8d96-4c64-a2c8-37c76c557d4f" path="/var/lib/kubelet/pods/c485bbf2-8d96-4c64-a2c8-37c76c557d4f/volumes" Feb 17 13:27:54 crc kubenswrapper[4955]: I0217 13:27:54.095620 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dwd4j"] Feb 17 13:27:54 crc kubenswrapper[4955]: I0217 13:27:54.098458 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dwd4j" podUID="512586cb-2c3c-4332-9f24-875632c4f22e" containerName="registry-server" containerID="cri-o://62b60d02bbff7bf65ed6020843b3a571666ac5ec53123e7af2f367e7f8cdd1fe" gracePeriod=2 Feb 17 13:27:54 crc kubenswrapper[4955]: I0217 13:27:54.529503 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dwd4j" Feb 17 13:27:54 crc kubenswrapper[4955]: I0217 13:27:54.618351 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/512586cb-2c3c-4332-9f24-875632c4f22e-catalog-content\") pod \"512586cb-2c3c-4332-9f24-875632c4f22e\" (UID: \"512586cb-2c3c-4332-9f24-875632c4f22e\") " Feb 17 13:27:54 crc kubenswrapper[4955]: I0217 13:27:54.619157 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/512586cb-2c3c-4332-9f24-875632c4f22e-utilities\") pod \"512586cb-2c3c-4332-9f24-875632c4f22e\" (UID: \"512586cb-2c3c-4332-9f24-875632c4f22e\") " Feb 17 13:27:54 crc kubenswrapper[4955]: I0217 13:27:54.619277 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h6f8w\" (UniqueName: \"kubernetes.io/projected/512586cb-2c3c-4332-9f24-875632c4f22e-kube-api-access-h6f8w\") pod \"512586cb-2c3c-4332-9f24-875632c4f22e\" (UID: \"512586cb-2c3c-4332-9f24-875632c4f22e\") " Feb 17 13:27:54 crc kubenswrapper[4955]: I0217 13:27:54.620081 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/512586cb-2c3c-4332-9f24-875632c4f22e-utilities" (OuterVolumeSpecName: "utilities") pod "512586cb-2c3c-4332-9f24-875632c4f22e" (UID: "512586cb-2c3c-4332-9f24-875632c4f22e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:27:54 crc kubenswrapper[4955]: I0217 13:27:54.620424 4955 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/512586cb-2c3c-4332-9f24-875632c4f22e-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 13:27:54 crc kubenswrapper[4955]: I0217 13:27:54.635170 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/512586cb-2c3c-4332-9f24-875632c4f22e-kube-api-access-h6f8w" (OuterVolumeSpecName: "kube-api-access-h6f8w") pod "512586cb-2c3c-4332-9f24-875632c4f22e" (UID: "512586cb-2c3c-4332-9f24-875632c4f22e"). InnerVolumeSpecName "kube-api-access-h6f8w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:27:54 crc kubenswrapper[4955]: I0217 13:27:54.641647 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/512586cb-2c3c-4332-9f24-875632c4f22e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "512586cb-2c3c-4332-9f24-875632c4f22e" (UID: "512586cb-2c3c-4332-9f24-875632c4f22e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:27:54 crc kubenswrapper[4955]: I0217 13:27:54.721567 4955 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/512586cb-2c3c-4332-9f24-875632c4f22e-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 13:27:54 crc kubenswrapper[4955]: I0217 13:27:54.721607 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h6f8w\" (UniqueName: \"kubernetes.io/projected/512586cb-2c3c-4332-9f24-875632c4f22e-kube-api-access-h6f8w\") on node \"crc\" DevicePath \"\"" Feb 17 13:27:55 crc kubenswrapper[4955]: I0217 13:27:55.115573 4955 generic.go:334] "Generic (PLEG): container finished" podID="512586cb-2c3c-4332-9f24-875632c4f22e" containerID="62b60d02bbff7bf65ed6020843b3a571666ac5ec53123e7af2f367e7f8cdd1fe" exitCode=0 Feb 17 13:27:55 crc kubenswrapper[4955]: I0217 13:27:55.115947 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dwd4j" event={"ID":"512586cb-2c3c-4332-9f24-875632c4f22e","Type":"ContainerDied","Data":"62b60d02bbff7bf65ed6020843b3a571666ac5ec53123e7af2f367e7f8cdd1fe"} Feb 17 13:27:55 crc kubenswrapper[4955]: I0217 13:27:55.116000 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dwd4j" event={"ID":"512586cb-2c3c-4332-9f24-875632c4f22e","Type":"ContainerDied","Data":"b0b63b33cef043b0b4390f8c8b86faa3b8e7a37f1dd53afc3839b03775de858f"} Feb 17 13:27:55 crc kubenswrapper[4955]: I0217 13:27:55.116023 4955 scope.go:117] "RemoveContainer" containerID="62b60d02bbff7bf65ed6020843b3a571666ac5ec53123e7af2f367e7f8cdd1fe" Feb 17 13:27:55 crc kubenswrapper[4955]: I0217 13:27:55.116400 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dwd4j" Feb 17 13:27:55 crc kubenswrapper[4955]: I0217 13:27:55.147702 4955 scope.go:117] "RemoveContainer" containerID="2a48bb5a87a4190a6c784c8700b0ea717d264f941af9c823cfec070240e05545" Feb 17 13:27:55 crc kubenswrapper[4955]: I0217 13:27:55.170913 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dwd4j"] Feb 17 13:27:55 crc kubenswrapper[4955]: I0217 13:27:55.181628 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dwd4j"] Feb 17 13:27:55 crc kubenswrapper[4955]: I0217 13:27:55.208126 4955 scope.go:117] "RemoveContainer" containerID="e6874c28d872efe0e26677b3f2192516abca9d04f247e129d8e0334acd576089" Feb 17 13:27:55 crc kubenswrapper[4955]: I0217 13:27:55.229995 4955 scope.go:117] "RemoveContainer" containerID="62b60d02bbff7bf65ed6020843b3a571666ac5ec53123e7af2f367e7f8cdd1fe" Feb 17 13:27:55 crc kubenswrapper[4955]: E0217 13:27:55.230321 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62b60d02bbff7bf65ed6020843b3a571666ac5ec53123e7af2f367e7f8cdd1fe\": container with ID starting with 62b60d02bbff7bf65ed6020843b3a571666ac5ec53123e7af2f367e7f8cdd1fe not found: ID does not exist" containerID="62b60d02bbff7bf65ed6020843b3a571666ac5ec53123e7af2f367e7f8cdd1fe" Feb 17 13:27:55 crc kubenswrapper[4955]: I0217 13:27:55.230349 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62b60d02bbff7bf65ed6020843b3a571666ac5ec53123e7af2f367e7f8cdd1fe"} err="failed to get container status \"62b60d02bbff7bf65ed6020843b3a571666ac5ec53123e7af2f367e7f8cdd1fe\": rpc error: code = NotFound desc = could not find container \"62b60d02bbff7bf65ed6020843b3a571666ac5ec53123e7af2f367e7f8cdd1fe\": container with ID starting with 62b60d02bbff7bf65ed6020843b3a571666ac5ec53123e7af2f367e7f8cdd1fe not found: ID does not exist" Feb 17 13:27:55 crc kubenswrapper[4955]: I0217 13:27:55.230368 4955 scope.go:117] "RemoveContainer" containerID="2a48bb5a87a4190a6c784c8700b0ea717d264f941af9c823cfec070240e05545" Feb 17 13:27:55 crc kubenswrapper[4955]: E0217 13:27:55.230733 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a48bb5a87a4190a6c784c8700b0ea717d264f941af9c823cfec070240e05545\": container with ID starting with 2a48bb5a87a4190a6c784c8700b0ea717d264f941af9c823cfec070240e05545 not found: ID does not exist" containerID="2a48bb5a87a4190a6c784c8700b0ea717d264f941af9c823cfec070240e05545" Feb 17 13:27:55 crc kubenswrapper[4955]: I0217 13:27:55.230755 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a48bb5a87a4190a6c784c8700b0ea717d264f941af9c823cfec070240e05545"} err="failed to get container status \"2a48bb5a87a4190a6c784c8700b0ea717d264f941af9c823cfec070240e05545\": rpc error: code = NotFound desc = could not find container \"2a48bb5a87a4190a6c784c8700b0ea717d264f941af9c823cfec070240e05545\": container with ID starting with 2a48bb5a87a4190a6c784c8700b0ea717d264f941af9c823cfec070240e05545 not found: ID does not exist" Feb 17 13:27:55 crc kubenswrapper[4955]: I0217 13:27:55.230770 4955 scope.go:117] "RemoveContainer" containerID="e6874c28d872efe0e26677b3f2192516abca9d04f247e129d8e0334acd576089" Feb 17 13:27:55 crc kubenswrapper[4955]: E0217 13:27:55.231008 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6874c28d872efe0e26677b3f2192516abca9d04f247e129d8e0334acd576089\": container with ID starting with e6874c28d872efe0e26677b3f2192516abca9d04f247e129d8e0334acd576089 not found: ID does not exist" containerID="e6874c28d872efe0e26677b3f2192516abca9d04f247e129d8e0334acd576089" Feb 17 13:27:55 crc kubenswrapper[4955]: I0217 13:27:55.231133 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6874c28d872efe0e26677b3f2192516abca9d04f247e129d8e0334acd576089"} err="failed to get container status \"e6874c28d872efe0e26677b3f2192516abca9d04f247e129d8e0334acd576089\": rpc error: code = NotFound desc = could not find container \"e6874c28d872efe0e26677b3f2192516abca9d04f247e129d8e0334acd576089\": container with ID starting with e6874c28d872efe0e26677b3f2192516abca9d04f247e129d8e0334acd576089 not found: ID does not exist" Feb 17 13:27:56 crc kubenswrapper[4955]: I0217 13:27:56.238878 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="512586cb-2c3c-4332-9f24-875632c4f22e" path="/var/lib/kubelet/pods/512586cb-2c3c-4332-9f24-875632c4f22e/volumes" Feb 17 13:28:46 crc kubenswrapper[4955]: I0217 13:28:46.504504 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qjr6h"] Feb 17 13:28:46 crc kubenswrapper[4955]: E0217 13:28:46.505623 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c485bbf2-8d96-4c64-a2c8-37c76c557d4f" containerName="extract-utilities" Feb 17 13:28:46 crc kubenswrapper[4955]: I0217 13:28:46.505641 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="c485bbf2-8d96-4c64-a2c8-37c76c557d4f" containerName="extract-utilities" Feb 17 13:28:46 crc kubenswrapper[4955]: E0217 13:28:46.505664 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="512586cb-2c3c-4332-9f24-875632c4f22e" containerName="registry-server" Feb 17 13:28:46 crc kubenswrapper[4955]: I0217 13:28:46.505672 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="512586cb-2c3c-4332-9f24-875632c4f22e" containerName="registry-server" Feb 17 13:28:46 crc kubenswrapper[4955]: E0217 13:28:46.505688 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c485bbf2-8d96-4c64-a2c8-37c76c557d4f" containerName="extract-content" Feb 17 13:28:46 crc kubenswrapper[4955]: I0217 13:28:46.505696 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="c485bbf2-8d96-4c64-a2c8-37c76c557d4f" containerName="extract-content" Feb 17 13:28:46 crc kubenswrapper[4955]: E0217 13:28:46.505717 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="512586cb-2c3c-4332-9f24-875632c4f22e" containerName="extract-content" Feb 17 13:28:46 crc kubenswrapper[4955]: I0217 13:28:46.505725 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="512586cb-2c3c-4332-9f24-875632c4f22e" containerName="extract-content" Feb 17 13:28:46 crc kubenswrapper[4955]: E0217 13:28:46.505743 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c485bbf2-8d96-4c64-a2c8-37c76c557d4f" containerName="registry-server" Feb 17 13:28:46 crc kubenswrapper[4955]: I0217 13:28:46.505751 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="c485bbf2-8d96-4c64-a2c8-37c76c557d4f" containerName="registry-server" Feb 17 13:28:46 crc kubenswrapper[4955]: E0217 13:28:46.505769 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="512586cb-2c3c-4332-9f24-875632c4f22e" containerName="extract-utilities" Feb 17 13:28:46 crc kubenswrapper[4955]: I0217 13:28:46.505797 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="512586cb-2c3c-4332-9f24-875632c4f22e" containerName="extract-utilities" Feb 17 13:28:46 crc kubenswrapper[4955]: I0217 13:28:46.506025 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="c485bbf2-8d96-4c64-a2c8-37c76c557d4f" containerName="registry-server" Feb 17 13:28:46 crc kubenswrapper[4955]: I0217 13:28:46.506059 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="512586cb-2c3c-4332-9f24-875632c4f22e" containerName="registry-server" Feb 17 13:28:46 crc kubenswrapper[4955]: I0217 13:28:46.507921 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qjr6h" Feb 17 13:28:46 crc kubenswrapper[4955]: I0217 13:28:46.522145 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qjr6h"] Feb 17 13:28:46 crc kubenswrapper[4955]: I0217 13:28:46.650446 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcw64\" (UniqueName: \"kubernetes.io/projected/0c87ece9-73fc-4f21-983c-9aac705b9d5b-kube-api-access-rcw64\") pod \"community-operators-qjr6h\" (UID: \"0c87ece9-73fc-4f21-983c-9aac705b9d5b\") " pod="openshift-marketplace/community-operators-qjr6h" Feb 17 13:28:46 crc kubenswrapper[4955]: I0217 13:28:46.650513 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c87ece9-73fc-4f21-983c-9aac705b9d5b-catalog-content\") pod \"community-operators-qjr6h\" (UID: \"0c87ece9-73fc-4f21-983c-9aac705b9d5b\") " pod="openshift-marketplace/community-operators-qjr6h" Feb 17 13:28:46 crc kubenswrapper[4955]: I0217 13:28:46.650620 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c87ece9-73fc-4f21-983c-9aac705b9d5b-utilities\") pod \"community-operators-qjr6h\" (UID: \"0c87ece9-73fc-4f21-983c-9aac705b9d5b\") " pod="openshift-marketplace/community-operators-qjr6h" Feb 17 13:28:46 crc kubenswrapper[4955]: I0217 13:28:46.752398 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcw64\" (UniqueName: \"kubernetes.io/projected/0c87ece9-73fc-4f21-983c-9aac705b9d5b-kube-api-access-rcw64\") pod \"community-operators-qjr6h\" (UID: \"0c87ece9-73fc-4f21-983c-9aac705b9d5b\") " pod="openshift-marketplace/community-operators-qjr6h" Feb 17 13:28:46 crc kubenswrapper[4955]: I0217 13:28:46.752479 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c87ece9-73fc-4f21-983c-9aac705b9d5b-catalog-content\") pod \"community-operators-qjr6h\" (UID: \"0c87ece9-73fc-4f21-983c-9aac705b9d5b\") " pod="openshift-marketplace/community-operators-qjr6h" Feb 17 13:28:46 crc kubenswrapper[4955]: I0217 13:28:46.752607 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c87ece9-73fc-4f21-983c-9aac705b9d5b-utilities\") pod \"community-operators-qjr6h\" (UID: \"0c87ece9-73fc-4f21-983c-9aac705b9d5b\") " pod="openshift-marketplace/community-operators-qjr6h" Feb 17 13:28:46 crc kubenswrapper[4955]: I0217 13:28:46.753283 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c87ece9-73fc-4f21-983c-9aac705b9d5b-catalog-content\") pod \"community-operators-qjr6h\" (UID: \"0c87ece9-73fc-4f21-983c-9aac705b9d5b\") " pod="openshift-marketplace/community-operators-qjr6h" Feb 17 13:28:46 crc kubenswrapper[4955]: I0217 13:28:46.753309 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c87ece9-73fc-4f21-983c-9aac705b9d5b-utilities\") pod \"community-operators-qjr6h\" (UID: \"0c87ece9-73fc-4f21-983c-9aac705b9d5b\") " pod="openshift-marketplace/community-operators-qjr6h" Feb 17 13:28:46 crc kubenswrapper[4955]: I0217 13:28:46.778373 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcw64\" (UniqueName: \"kubernetes.io/projected/0c87ece9-73fc-4f21-983c-9aac705b9d5b-kube-api-access-rcw64\") pod \"community-operators-qjr6h\" (UID: \"0c87ece9-73fc-4f21-983c-9aac705b9d5b\") " pod="openshift-marketplace/community-operators-qjr6h" Feb 17 13:28:46 crc kubenswrapper[4955]: I0217 13:28:46.830738 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qjr6h" Feb 17 13:28:47 crc kubenswrapper[4955]: I0217 13:28:47.371169 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qjr6h"] Feb 17 13:28:47 crc kubenswrapper[4955]: I0217 13:28:47.653235 4955 generic.go:334] "Generic (PLEG): container finished" podID="0c87ece9-73fc-4f21-983c-9aac705b9d5b" containerID="e895e2398263494586ece2ab72b065c36ee76dbafeb17f337ec5300fc869de18" exitCode=0 Feb 17 13:28:47 crc kubenswrapper[4955]: I0217 13:28:47.653347 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qjr6h" event={"ID":"0c87ece9-73fc-4f21-983c-9aac705b9d5b","Type":"ContainerDied","Data":"e895e2398263494586ece2ab72b065c36ee76dbafeb17f337ec5300fc869de18"} Feb 17 13:28:47 crc kubenswrapper[4955]: I0217 13:28:47.653588 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qjr6h" event={"ID":"0c87ece9-73fc-4f21-983c-9aac705b9d5b","Type":"ContainerStarted","Data":"d8139417eb0ee799be7b19a5447a817f143d524d8e6b6da53f2c2d03f06a4627"} Feb 17 13:28:48 crc kubenswrapper[4955]: I0217 13:28:48.395615 4955 scope.go:117] "RemoveContainer" containerID="902464c41c5c22295b964e133a26a55d00752d1e0938fe4d135fcb53ce02936a" Feb 17 13:28:48 crc kubenswrapper[4955]: I0217 13:28:48.671608 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qjr6h" event={"ID":"0c87ece9-73fc-4f21-983c-9aac705b9d5b","Type":"ContainerStarted","Data":"309be69d884794acb1ef74c52f176e3928da48498a769931a2d0846df98028e6"} Feb 17 13:28:49 crc kubenswrapper[4955]: I0217 13:28:49.683417 4955 generic.go:334] "Generic (PLEG): container finished" podID="0c87ece9-73fc-4f21-983c-9aac705b9d5b" containerID="309be69d884794acb1ef74c52f176e3928da48498a769931a2d0846df98028e6" exitCode=0 Feb 17 13:28:49 crc kubenswrapper[4955]: I0217 13:28:49.683518 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qjr6h" event={"ID":"0c87ece9-73fc-4f21-983c-9aac705b9d5b","Type":"ContainerDied","Data":"309be69d884794acb1ef74c52f176e3928da48498a769931a2d0846df98028e6"} Feb 17 13:28:50 crc kubenswrapper[4955]: I0217 13:28:50.700223 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qjr6h" event={"ID":"0c87ece9-73fc-4f21-983c-9aac705b9d5b","Type":"ContainerStarted","Data":"8be5efa81953b9b37bcf5abffee677dbb5fb83e4774929c53dbabab74984cb64"} Feb 17 13:28:50 crc kubenswrapper[4955]: I0217 13:28:50.721631 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-qjr6h" podStartSLOduration=2.319706741 podStartE2EDuration="4.721609669s" podCreationTimestamp="2026-02-17 13:28:46 +0000 UTC" firstStartedPulling="2026-02-17 13:28:47.656049371 +0000 UTC m=+1466.178778914" lastFinishedPulling="2026-02-17 13:28:50.057952299 +0000 UTC m=+1468.580681842" observedRunningTime="2026-02-17 13:28:50.717378788 +0000 UTC m=+1469.240108331" watchObservedRunningTime="2026-02-17 13:28:50.721609669 +0000 UTC m=+1469.244339212" Feb 17 13:28:56 crc kubenswrapper[4955]: I0217 13:28:56.831101 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-qjr6h" Feb 17 13:28:56 crc kubenswrapper[4955]: I0217 13:28:56.831712 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-qjr6h" Feb 17 13:28:56 crc kubenswrapper[4955]: I0217 13:28:56.880569 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-qjr6h" Feb 17 13:28:57 crc kubenswrapper[4955]: I0217 13:28:57.817613 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-qjr6h" Feb 17 13:28:57 crc kubenswrapper[4955]: I0217 13:28:57.871755 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qjr6h"] Feb 17 13:28:59 crc kubenswrapper[4955]: I0217 13:28:59.795312 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-qjr6h" podUID="0c87ece9-73fc-4f21-983c-9aac705b9d5b" containerName="registry-server" containerID="cri-o://8be5efa81953b9b37bcf5abffee677dbb5fb83e4774929c53dbabab74984cb64" gracePeriod=2 Feb 17 13:29:00 crc kubenswrapper[4955]: I0217 13:29:00.428792 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qjr6h" Feb 17 13:29:00 crc kubenswrapper[4955]: I0217 13:29:00.537143 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c87ece9-73fc-4f21-983c-9aac705b9d5b-utilities\") pod \"0c87ece9-73fc-4f21-983c-9aac705b9d5b\" (UID: \"0c87ece9-73fc-4f21-983c-9aac705b9d5b\") " Feb 17 13:29:00 crc kubenswrapper[4955]: I0217 13:29:00.537329 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rcw64\" (UniqueName: \"kubernetes.io/projected/0c87ece9-73fc-4f21-983c-9aac705b9d5b-kube-api-access-rcw64\") pod \"0c87ece9-73fc-4f21-983c-9aac705b9d5b\" (UID: \"0c87ece9-73fc-4f21-983c-9aac705b9d5b\") " Feb 17 13:29:00 crc kubenswrapper[4955]: I0217 13:29:00.537481 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c87ece9-73fc-4f21-983c-9aac705b9d5b-catalog-content\") pod \"0c87ece9-73fc-4f21-983c-9aac705b9d5b\" (UID: \"0c87ece9-73fc-4f21-983c-9aac705b9d5b\") " Feb 17 13:29:00 crc kubenswrapper[4955]: I0217 13:29:00.538496 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c87ece9-73fc-4f21-983c-9aac705b9d5b-utilities" (OuterVolumeSpecName: "utilities") pod "0c87ece9-73fc-4f21-983c-9aac705b9d5b" (UID: "0c87ece9-73fc-4f21-983c-9aac705b9d5b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:29:00 crc kubenswrapper[4955]: I0217 13:29:00.544831 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c87ece9-73fc-4f21-983c-9aac705b9d5b-kube-api-access-rcw64" (OuterVolumeSpecName: "kube-api-access-rcw64") pod "0c87ece9-73fc-4f21-983c-9aac705b9d5b" (UID: "0c87ece9-73fc-4f21-983c-9aac705b9d5b"). InnerVolumeSpecName "kube-api-access-rcw64". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:29:00 crc kubenswrapper[4955]: I0217 13:29:00.598218 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c87ece9-73fc-4f21-983c-9aac705b9d5b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0c87ece9-73fc-4f21-983c-9aac705b9d5b" (UID: "0c87ece9-73fc-4f21-983c-9aac705b9d5b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:29:00 crc kubenswrapper[4955]: I0217 13:29:00.639966 4955 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c87ece9-73fc-4f21-983c-9aac705b9d5b-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 13:29:00 crc kubenswrapper[4955]: I0217 13:29:00.640009 4955 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c87ece9-73fc-4f21-983c-9aac705b9d5b-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 13:29:00 crc kubenswrapper[4955]: I0217 13:29:00.640022 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rcw64\" (UniqueName: \"kubernetes.io/projected/0c87ece9-73fc-4f21-983c-9aac705b9d5b-kube-api-access-rcw64\") on node \"crc\" DevicePath \"\"" Feb 17 13:29:00 crc kubenswrapper[4955]: I0217 13:29:00.811232 4955 generic.go:334] "Generic (PLEG): container finished" podID="0c87ece9-73fc-4f21-983c-9aac705b9d5b" containerID="8be5efa81953b9b37bcf5abffee677dbb5fb83e4774929c53dbabab74984cb64" exitCode=0 Feb 17 13:29:00 crc kubenswrapper[4955]: I0217 13:29:00.811319 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qjr6h" event={"ID":"0c87ece9-73fc-4f21-983c-9aac705b9d5b","Type":"ContainerDied","Data":"8be5efa81953b9b37bcf5abffee677dbb5fb83e4774929c53dbabab74984cb64"} Feb 17 13:29:00 crc kubenswrapper[4955]: I0217 13:29:00.811369 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qjr6h" Feb 17 13:29:00 crc kubenswrapper[4955]: I0217 13:29:00.811413 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qjr6h" event={"ID":"0c87ece9-73fc-4f21-983c-9aac705b9d5b","Type":"ContainerDied","Data":"d8139417eb0ee799be7b19a5447a817f143d524d8e6b6da53f2c2d03f06a4627"} Feb 17 13:29:00 crc kubenswrapper[4955]: I0217 13:29:00.811442 4955 scope.go:117] "RemoveContainer" containerID="8be5efa81953b9b37bcf5abffee677dbb5fb83e4774929c53dbabab74984cb64" Feb 17 13:29:00 crc kubenswrapper[4955]: I0217 13:29:00.850852 4955 scope.go:117] "RemoveContainer" containerID="309be69d884794acb1ef74c52f176e3928da48498a769931a2d0846df98028e6" Feb 17 13:29:00 crc kubenswrapper[4955]: I0217 13:29:00.858386 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qjr6h"] Feb 17 13:29:00 crc kubenswrapper[4955]: I0217 13:29:00.867655 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-qjr6h"] Feb 17 13:29:00 crc kubenswrapper[4955]: I0217 13:29:00.879332 4955 scope.go:117] "RemoveContainer" containerID="e895e2398263494586ece2ab72b065c36ee76dbafeb17f337ec5300fc869de18" Feb 17 13:29:00 crc kubenswrapper[4955]: I0217 13:29:00.979096 4955 scope.go:117] "RemoveContainer" containerID="8be5efa81953b9b37bcf5abffee677dbb5fb83e4774929c53dbabab74984cb64" Feb 17 13:29:00 crc kubenswrapper[4955]: E0217 13:29:00.982456 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8be5efa81953b9b37bcf5abffee677dbb5fb83e4774929c53dbabab74984cb64\": container with ID starting with 8be5efa81953b9b37bcf5abffee677dbb5fb83e4774929c53dbabab74984cb64 not found: ID does not exist" containerID="8be5efa81953b9b37bcf5abffee677dbb5fb83e4774929c53dbabab74984cb64" Feb 17 13:29:00 crc kubenswrapper[4955]: I0217 13:29:00.982486 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8be5efa81953b9b37bcf5abffee677dbb5fb83e4774929c53dbabab74984cb64"} err="failed to get container status \"8be5efa81953b9b37bcf5abffee677dbb5fb83e4774929c53dbabab74984cb64\": rpc error: code = NotFound desc = could not find container \"8be5efa81953b9b37bcf5abffee677dbb5fb83e4774929c53dbabab74984cb64\": container with ID starting with 8be5efa81953b9b37bcf5abffee677dbb5fb83e4774929c53dbabab74984cb64 not found: ID does not exist" Feb 17 13:29:00 crc kubenswrapper[4955]: I0217 13:29:00.982510 4955 scope.go:117] "RemoveContainer" containerID="309be69d884794acb1ef74c52f176e3928da48498a769931a2d0846df98028e6" Feb 17 13:29:00 crc kubenswrapper[4955]: E0217 13:29:00.982931 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"309be69d884794acb1ef74c52f176e3928da48498a769931a2d0846df98028e6\": container with ID starting with 309be69d884794acb1ef74c52f176e3928da48498a769931a2d0846df98028e6 not found: ID does not exist" containerID="309be69d884794acb1ef74c52f176e3928da48498a769931a2d0846df98028e6" Feb 17 13:29:00 crc kubenswrapper[4955]: I0217 13:29:00.982983 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"309be69d884794acb1ef74c52f176e3928da48498a769931a2d0846df98028e6"} err="failed to get container status \"309be69d884794acb1ef74c52f176e3928da48498a769931a2d0846df98028e6\": rpc error: code = NotFound desc = could not find container \"309be69d884794acb1ef74c52f176e3928da48498a769931a2d0846df98028e6\": container with ID starting with 309be69d884794acb1ef74c52f176e3928da48498a769931a2d0846df98028e6 not found: ID does not exist" Feb 17 13:29:00 crc kubenswrapper[4955]: I0217 13:29:00.983009 4955 scope.go:117] "RemoveContainer" containerID="e895e2398263494586ece2ab72b065c36ee76dbafeb17f337ec5300fc869de18" Feb 17 13:29:00 crc kubenswrapper[4955]: E0217 13:29:00.983390 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e895e2398263494586ece2ab72b065c36ee76dbafeb17f337ec5300fc869de18\": container with ID starting with e895e2398263494586ece2ab72b065c36ee76dbafeb17f337ec5300fc869de18 not found: ID does not exist" containerID="e895e2398263494586ece2ab72b065c36ee76dbafeb17f337ec5300fc869de18" Feb 17 13:29:00 crc kubenswrapper[4955]: I0217 13:29:00.983436 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e895e2398263494586ece2ab72b065c36ee76dbafeb17f337ec5300fc869de18"} err="failed to get container status \"e895e2398263494586ece2ab72b065c36ee76dbafeb17f337ec5300fc869de18\": rpc error: code = NotFound desc = could not find container \"e895e2398263494586ece2ab72b065c36ee76dbafeb17f337ec5300fc869de18\": container with ID starting with e895e2398263494586ece2ab72b065c36ee76dbafeb17f337ec5300fc869de18 not found: ID does not exist" Feb 17 13:29:02 crc kubenswrapper[4955]: I0217 13:29:02.232821 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c87ece9-73fc-4f21-983c-9aac705b9d5b" path="/var/lib/kubelet/pods/0c87ece9-73fc-4f21-983c-9aac705b9d5b/volumes" Feb 17 13:29:04 crc kubenswrapper[4955]: I0217 13:29:04.575132 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 13:29:04 crc kubenswrapper[4955]: I0217 13:29:04.575495 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 13:29:34 crc kubenswrapper[4955]: I0217 13:29:34.574391 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 13:29:34 crc kubenswrapper[4955]: I0217 13:29:34.575944 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 13:29:48 crc kubenswrapper[4955]: I0217 13:29:48.496619 4955 scope.go:117] "RemoveContainer" containerID="800728be94187d7cbb16633c7f612b8d818340e1f95cd868d2a7088601c9ea38" Feb 17 13:29:48 crc kubenswrapper[4955]: I0217 13:29:48.577084 4955 scope.go:117] "RemoveContainer" containerID="ddb90ae2171d4c56825cc492fcade33f9f4837d699bc2112954ddeddd9d99958" Feb 17 13:29:48 crc kubenswrapper[4955]: I0217 13:29:48.603355 4955 scope.go:117] "RemoveContainer" containerID="ed9ae269161b452111acb8fe7e3cac0476b8ea07308d6f3e4c8d0072aacdca7e" Feb 17 13:29:48 crc kubenswrapper[4955]: I0217 13:29:48.628922 4955 scope.go:117] "RemoveContainer" containerID="76e742fa32650a687849fb64674d1d2f47974c7b99026b751786a56009f0ba57" Feb 17 13:30:00 crc kubenswrapper[4955]: I0217 13:30:00.150862 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29522250-m57xx"] Feb 17 13:30:00 crc kubenswrapper[4955]: E0217 13:30:00.151884 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c87ece9-73fc-4f21-983c-9aac705b9d5b" containerName="registry-server" Feb 17 13:30:00 crc kubenswrapper[4955]: I0217 13:30:00.151901 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c87ece9-73fc-4f21-983c-9aac705b9d5b" containerName="registry-server" Feb 17 13:30:00 crc kubenswrapper[4955]: E0217 13:30:00.151936 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c87ece9-73fc-4f21-983c-9aac705b9d5b" containerName="extract-utilities" Feb 17 13:30:00 crc kubenswrapper[4955]: I0217 13:30:00.151943 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c87ece9-73fc-4f21-983c-9aac705b9d5b" containerName="extract-utilities" Feb 17 13:30:00 crc kubenswrapper[4955]: E0217 13:30:00.151961 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c87ece9-73fc-4f21-983c-9aac705b9d5b" containerName="extract-content" Feb 17 13:30:00 crc kubenswrapper[4955]: I0217 13:30:00.151968 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c87ece9-73fc-4f21-983c-9aac705b9d5b" containerName="extract-content" Feb 17 13:30:00 crc kubenswrapper[4955]: I0217 13:30:00.152171 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c87ece9-73fc-4f21-983c-9aac705b9d5b" containerName="registry-server" Feb 17 13:30:00 crc kubenswrapper[4955]: I0217 13:30:00.152899 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29522250-m57xx" Feb 17 13:30:00 crc kubenswrapper[4955]: I0217 13:30:00.155879 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 17 13:30:00 crc kubenswrapper[4955]: I0217 13:30:00.157468 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 17 13:30:00 crc kubenswrapper[4955]: I0217 13:30:00.165940 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29522250-m57xx"] Feb 17 13:30:00 crc kubenswrapper[4955]: I0217 13:30:00.224789 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1fabad5b-2fa4-47a1-93bc-70d5b6299a41-config-volume\") pod \"collect-profiles-29522250-m57xx\" (UID: \"1fabad5b-2fa4-47a1-93bc-70d5b6299a41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522250-m57xx" Feb 17 13:30:00 crc kubenswrapper[4955]: I0217 13:30:00.225160 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdkdn\" (UniqueName: \"kubernetes.io/projected/1fabad5b-2fa4-47a1-93bc-70d5b6299a41-kube-api-access-bdkdn\") pod \"collect-profiles-29522250-m57xx\" (UID: \"1fabad5b-2fa4-47a1-93bc-70d5b6299a41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522250-m57xx" Feb 17 13:30:00 crc kubenswrapper[4955]: I0217 13:30:00.225291 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1fabad5b-2fa4-47a1-93bc-70d5b6299a41-secret-volume\") pod \"collect-profiles-29522250-m57xx\" (UID: \"1fabad5b-2fa4-47a1-93bc-70d5b6299a41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522250-m57xx" Feb 17 13:30:00 crc kubenswrapper[4955]: I0217 13:30:00.326688 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1fabad5b-2fa4-47a1-93bc-70d5b6299a41-config-volume\") pod \"collect-profiles-29522250-m57xx\" (UID: \"1fabad5b-2fa4-47a1-93bc-70d5b6299a41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522250-m57xx" Feb 17 13:30:00 crc kubenswrapper[4955]: I0217 13:30:00.326828 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdkdn\" (UniqueName: \"kubernetes.io/projected/1fabad5b-2fa4-47a1-93bc-70d5b6299a41-kube-api-access-bdkdn\") pod \"collect-profiles-29522250-m57xx\" (UID: \"1fabad5b-2fa4-47a1-93bc-70d5b6299a41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522250-m57xx" Feb 17 13:30:00 crc kubenswrapper[4955]: I0217 13:30:00.326856 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1fabad5b-2fa4-47a1-93bc-70d5b6299a41-secret-volume\") pod \"collect-profiles-29522250-m57xx\" (UID: \"1fabad5b-2fa4-47a1-93bc-70d5b6299a41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522250-m57xx" Feb 17 13:30:00 crc kubenswrapper[4955]: I0217 13:30:00.328032 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1fabad5b-2fa4-47a1-93bc-70d5b6299a41-config-volume\") pod \"collect-profiles-29522250-m57xx\" (UID: \"1fabad5b-2fa4-47a1-93bc-70d5b6299a41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522250-m57xx" Feb 17 13:30:00 crc kubenswrapper[4955]: I0217 13:30:00.335996 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1fabad5b-2fa4-47a1-93bc-70d5b6299a41-secret-volume\") pod \"collect-profiles-29522250-m57xx\" (UID: \"1fabad5b-2fa4-47a1-93bc-70d5b6299a41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522250-m57xx" Feb 17 13:30:00 crc kubenswrapper[4955]: I0217 13:30:00.348761 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdkdn\" (UniqueName: \"kubernetes.io/projected/1fabad5b-2fa4-47a1-93bc-70d5b6299a41-kube-api-access-bdkdn\") pod \"collect-profiles-29522250-m57xx\" (UID: \"1fabad5b-2fa4-47a1-93bc-70d5b6299a41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522250-m57xx" Feb 17 13:30:00 crc kubenswrapper[4955]: I0217 13:30:00.481273 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29522250-m57xx" Feb 17 13:30:00 crc kubenswrapper[4955]: I0217 13:30:00.974098 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29522250-m57xx"] Feb 17 13:30:01 crc kubenswrapper[4955]: I0217 13:30:01.419396 4955 generic.go:334] "Generic (PLEG): container finished" podID="1fabad5b-2fa4-47a1-93bc-70d5b6299a41" containerID="15fa28350c7f25ce62e418796a1b0df9b9d93076a4bf3a9597ac4c87f2caaea8" exitCode=0 Feb 17 13:30:01 crc kubenswrapper[4955]: I0217 13:30:01.419468 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29522250-m57xx" event={"ID":"1fabad5b-2fa4-47a1-93bc-70d5b6299a41","Type":"ContainerDied","Data":"15fa28350c7f25ce62e418796a1b0df9b9d93076a4bf3a9597ac4c87f2caaea8"} Feb 17 13:30:01 crc kubenswrapper[4955]: I0217 13:30:01.419695 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29522250-m57xx" event={"ID":"1fabad5b-2fa4-47a1-93bc-70d5b6299a41","Type":"ContainerStarted","Data":"d406d8af3ce91e659695a800b598790b65edea67302da7d598dc586c9a600622"} Feb 17 13:30:02 crc kubenswrapper[4955]: I0217 13:30:02.762052 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29522250-m57xx" Feb 17 13:30:02 crc kubenswrapper[4955]: I0217 13:30:02.891015 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1fabad5b-2fa4-47a1-93bc-70d5b6299a41-secret-volume\") pod \"1fabad5b-2fa4-47a1-93bc-70d5b6299a41\" (UID: \"1fabad5b-2fa4-47a1-93bc-70d5b6299a41\") " Feb 17 13:30:02 crc kubenswrapper[4955]: I0217 13:30:02.891172 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdkdn\" (UniqueName: \"kubernetes.io/projected/1fabad5b-2fa4-47a1-93bc-70d5b6299a41-kube-api-access-bdkdn\") pod \"1fabad5b-2fa4-47a1-93bc-70d5b6299a41\" (UID: \"1fabad5b-2fa4-47a1-93bc-70d5b6299a41\") " Feb 17 13:30:02 crc kubenswrapper[4955]: I0217 13:30:02.891249 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1fabad5b-2fa4-47a1-93bc-70d5b6299a41-config-volume\") pod \"1fabad5b-2fa4-47a1-93bc-70d5b6299a41\" (UID: \"1fabad5b-2fa4-47a1-93bc-70d5b6299a41\") " Feb 17 13:30:02 crc kubenswrapper[4955]: I0217 13:30:02.892727 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1fabad5b-2fa4-47a1-93bc-70d5b6299a41-config-volume" (OuterVolumeSpecName: "config-volume") pod "1fabad5b-2fa4-47a1-93bc-70d5b6299a41" (UID: "1fabad5b-2fa4-47a1-93bc-70d5b6299a41"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:30:02 crc kubenswrapper[4955]: I0217 13:30:02.904126 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fabad5b-2fa4-47a1-93bc-70d5b6299a41-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1fabad5b-2fa4-47a1-93bc-70d5b6299a41" (UID: "1fabad5b-2fa4-47a1-93bc-70d5b6299a41"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:30:02 crc kubenswrapper[4955]: I0217 13:30:02.904493 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fabad5b-2fa4-47a1-93bc-70d5b6299a41-kube-api-access-bdkdn" (OuterVolumeSpecName: "kube-api-access-bdkdn") pod "1fabad5b-2fa4-47a1-93bc-70d5b6299a41" (UID: "1fabad5b-2fa4-47a1-93bc-70d5b6299a41"). InnerVolumeSpecName "kube-api-access-bdkdn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:30:02 crc kubenswrapper[4955]: I0217 13:30:02.994047 4955 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1fabad5b-2fa4-47a1-93bc-70d5b6299a41-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 17 13:30:02 crc kubenswrapper[4955]: I0217 13:30:02.994111 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdkdn\" (UniqueName: \"kubernetes.io/projected/1fabad5b-2fa4-47a1-93bc-70d5b6299a41-kube-api-access-bdkdn\") on node \"crc\" DevicePath \"\"" Feb 17 13:30:02 crc kubenswrapper[4955]: I0217 13:30:02.994125 4955 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1fabad5b-2fa4-47a1-93bc-70d5b6299a41-config-volume\") on node \"crc\" DevicePath \"\"" Feb 17 13:30:03 crc kubenswrapper[4955]: I0217 13:30:03.441268 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29522250-m57xx" Feb 17 13:30:03 crc kubenswrapper[4955]: I0217 13:30:03.441258 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29522250-m57xx" event={"ID":"1fabad5b-2fa4-47a1-93bc-70d5b6299a41","Type":"ContainerDied","Data":"d406d8af3ce91e659695a800b598790b65edea67302da7d598dc586c9a600622"} Feb 17 13:30:03 crc kubenswrapper[4955]: I0217 13:30:03.441553 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d406d8af3ce91e659695a800b598790b65edea67302da7d598dc586c9a600622" Feb 17 13:30:04 crc kubenswrapper[4955]: I0217 13:30:04.575657 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 13:30:04 crc kubenswrapper[4955]: I0217 13:30:04.576000 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 13:30:04 crc kubenswrapper[4955]: I0217 13:30:04.576046 4955 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" Feb 17 13:30:04 crc kubenswrapper[4955]: I0217 13:30:04.576755 4955 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"969807d9ca63cdcc1aa7cc07aa21a249802167bf58b0b6150149361ce19b14ca"} pod="openshift-machine-config-operator/machine-config-daemon-29qxq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 17 13:30:04 crc kubenswrapper[4955]: I0217 13:30:04.576862 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" containerID="cri-o://969807d9ca63cdcc1aa7cc07aa21a249802167bf58b0b6150149361ce19b14ca" gracePeriod=600 Feb 17 13:30:04 crc kubenswrapper[4955]: E0217 13:30:04.703953 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:30:05 crc kubenswrapper[4955]: I0217 13:30:05.461645 4955 generic.go:334] "Generic (PLEG): container finished" podID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerID="969807d9ca63cdcc1aa7cc07aa21a249802167bf58b0b6150149361ce19b14ca" exitCode=0 Feb 17 13:30:05 crc kubenswrapper[4955]: I0217 13:30:05.462009 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" event={"ID":"5bf36887-2a2b-4193-abdb-a15beaf46300","Type":"ContainerDied","Data":"969807d9ca63cdcc1aa7cc07aa21a249802167bf58b0b6150149361ce19b14ca"} Feb 17 13:30:05 crc kubenswrapper[4955]: I0217 13:30:05.462115 4955 scope.go:117] "RemoveContainer" containerID="2a2497139f3e8754b75d83788b3f1744445e509419949dab14b01c6f81278a2c" Feb 17 13:30:05 crc kubenswrapper[4955]: I0217 13:30:05.462986 4955 scope.go:117] "RemoveContainer" containerID="969807d9ca63cdcc1aa7cc07aa21a249802167bf58b0b6150149361ce19b14ca" Feb 17 13:30:05 crc kubenswrapper[4955]: E0217 13:30:05.463434 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:30:20 crc kubenswrapper[4955]: I0217 13:30:20.230195 4955 scope.go:117] "RemoveContainer" containerID="969807d9ca63cdcc1aa7cc07aa21a249802167bf58b0b6150149361ce19b14ca" Feb 17 13:30:20 crc kubenswrapper[4955]: E0217 13:30:20.231138 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:30:34 crc kubenswrapper[4955]: I0217 13:30:34.223557 4955 scope.go:117] "RemoveContainer" containerID="969807d9ca63cdcc1aa7cc07aa21a249802167bf58b0b6150149361ce19b14ca" Feb 17 13:30:34 crc kubenswrapper[4955]: E0217 13:30:34.224375 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:30:48 crc kubenswrapper[4955]: I0217 13:30:48.748632 4955 scope.go:117] "RemoveContainer" containerID="9b207aca56cd7a2e24f4a4822726c62323aa73716a862d511996c9f70ad4fe05" Feb 17 13:30:49 crc kubenswrapper[4955]: I0217 13:30:49.223619 4955 scope.go:117] "RemoveContainer" containerID="969807d9ca63cdcc1aa7cc07aa21a249802167bf58b0b6150149361ce19b14ca" Feb 17 13:30:49 crc kubenswrapper[4955]: E0217 13:30:49.223906 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:31:04 crc kubenswrapper[4955]: I0217 13:31:04.223584 4955 scope.go:117] "RemoveContainer" containerID="969807d9ca63cdcc1aa7cc07aa21a249802167bf58b0b6150149361ce19b14ca" Feb 17 13:31:04 crc kubenswrapper[4955]: E0217 13:31:04.224962 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:31:07 crc kubenswrapper[4955]: I0217 13:31:07.119528 4955 generic.go:334] "Generic (PLEG): container finished" podID="ea41b3c5-2c55-4294-b2dc-a25ab083c780" containerID="9cc3799076b06fd1ced08eb820730cb665937dc35990eccfe214867e77a09f72" exitCode=0 Feb 17 13:31:07 crc kubenswrapper[4955]: I0217 13:31:07.119604 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn5l6" event={"ID":"ea41b3c5-2c55-4294-b2dc-a25ab083c780","Type":"ContainerDied","Data":"9cc3799076b06fd1ced08eb820730cb665937dc35990eccfe214867e77a09f72"} Feb 17 13:31:08 crc kubenswrapper[4955]: I0217 13:31:08.544663 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn5l6" Feb 17 13:31:08 crc kubenswrapper[4955]: I0217 13:31:08.647145 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ea41b3c5-2c55-4294-b2dc-a25ab083c780-ssh-key-openstack-edpm-ipam\") pod \"ea41b3c5-2c55-4294-b2dc-a25ab083c780\" (UID: \"ea41b3c5-2c55-4294-b2dc-a25ab083c780\") " Feb 17 13:31:08 crc kubenswrapper[4955]: I0217 13:31:08.648096 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ea41b3c5-2c55-4294-b2dc-a25ab083c780-inventory\") pod \"ea41b3c5-2c55-4294-b2dc-a25ab083c780\" (UID: \"ea41b3c5-2c55-4294-b2dc-a25ab083c780\") " Feb 17 13:31:08 crc kubenswrapper[4955]: I0217 13:31:08.648153 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea41b3c5-2c55-4294-b2dc-a25ab083c780-bootstrap-combined-ca-bundle\") pod \"ea41b3c5-2c55-4294-b2dc-a25ab083c780\" (UID: \"ea41b3c5-2c55-4294-b2dc-a25ab083c780\") " Feb 17 13:31:08 crc kubenswrapper[4955]: I0217 13:31:08.648275 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cx86f\" (UniqueName: \"kubernetes.io/projected/ea41b3c5-2c55-4294-b2dc-a25ab083c780-kube-api-access-cx86f\") pod \"ea41b3c5-2c55-4294-b2dc-a25ab083c780\" (UID: \"ea41b3c5-2c55-4294-b2dc-a25ab083c780\") " Feb 17 13:31:08 crc kubenswrapper[4955]: I0217 13:31:08.654279 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea41b3c5-2c55-4294-b2dc-a25ab083c780-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "ea41b3c5-2c55-4294-b2dc-a25ab083c780" (UID: "ea41b3c5-2c55-4294-b2dc-a25ab083c780"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:31:08 crc kubenswrapper[4955]: I0217 13:31:08.656280 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea41b3c5-2c55-4294-b2dc-a25ab083c780-kube-api-access-cx86f" (OuterVolumeSpecName: "kube-api-access-cx86f") pod "ea41b3c5-2c55-4294-b2dc-a25ab083c780" (UID: "ea41b3c5-2c55-4294-b2dc-a25ab083c780"). InnerVolumeSpecName "kube-api-access-cx86f". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:31:08 crc kubenswrapper[4955]: I0217 13:31:08.678465 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea41b3c5-2c55-4294-b2dc-a25ab083c780-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "ea41b3c5-2c55-4294-b2dc-a25ab083c780" (UID: "ea41b3c5-2c55-4294-b2dc-a25ab083c780"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:31:08 crc kubenswrapper[4955]: I0217 13:31:08.678877 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea41b3c5-2c55-4294-b2dc-a25ab083c780-inventory" (OuterVolumeSpecName: "inventory") pod "ea41b3c5-2c55-4294-b2dc-a25ab083c780" (UID: "ea41b3c5-2c55-4294-b2dc-a25ab083c780"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:31:08 crc kubenswrapper[4955]: I0217 13:31:08.750395 4955 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ea41b3c5-2c55-4294-b2dc-a25ab083c780-inventory\") on node \"crc\" DevicePath \"\"" Feb 17 13:31:08 crc kubenswrapper[4955]: I0217 13:31:08.750427 4955 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea41b3c5-2c55-4294-b2dc-a25ab083c780-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:31:08 crc kubenswrapper[4955]: I0217 13:31:08.750437 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cx86f\" (UniqueName: \"kubernetes.io/projected/ea41b3c5-2c55-4294-b2dc-a25ab083c780-kube-api-access-cx86f\") on node \"crc\" DevicePath \"\"" Feb 17 13:31:08 crc kubenswrapper[4955]: I0217 13:31:08.750446 4955 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ea41b3c5-2c55-4294-b2dc-a25ab083c780-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 17 13:31:09 crc kubenswrapper[4955]: I0217 13:31:09.144418 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn5l6" event={"ID":"ea41b3c5-2c55-4294-b2dc-a25ab083c780","Type":"ContainerDied","Data":"e9e9d821dc9bde95c915b0c4ed1863824c8bf723119a367f24bdadc829f1d97e"} Feb 17 13:31:09 crc kubenswrapper[4955]: I0217 13:31:09.144467 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn5l6" Feb 17 13:31:09 crc kubenswrapper[4955]: I0217 13:31:09.144472 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9e9d821dc9bde95c915b0c4ed1863824c8bf723119a367f24bdadc829f1d97e" Feb 17 13:31:09 crc kubenswrapper[4955]: I0217 13:31:09.229312 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bpbzk"] Feb 17 13:31:09 crc kubenswrapper[4955]: E0217 13:31:09.229819 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fabad5b-2fa4-47a1-93bc-70d5b6299a41" containerName="collect-profiles" Feb 17 13:31:09 crc kubenswrapper[4955]: I0217 13:31:09.229838 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fabad5b-2fa4-47a1-93bc-70d5b6299a41" containerName="collect-profiles" Feb 17 13:31:09 crc kubenswrapper[4955]: E0217 13:31:09.229856 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea41b3c5-2c55-4294-b2dc-a25ab083c780" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 17 13:31:09 crc kubenswrapper[4955]: I0217 13:31:09.229895 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea41b3c5-2c55-4294-b2dc-a25ab083c780" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 17 13:31:09 crc kubenswrapper[4955]: I0217 13:31:09.230146 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fabad5b-2fa4-47a1-93bc-70d5b6299a41" containerName="collect-profiles" Feb 17 13:31:09 crc kubenswrapper[4955]: I0217 13:31:09.230183 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea41b3c5-2c55-4294-b2dc-a25ab083c780" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 17 13:31:09 crc kubenswrapper[4955]: I0217 13:31:09.230850 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bpbzk" Feb 17 13:31:09 crc kubenswrapper[4955]: I0217 13:31:09.234303 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-t9lgc" Feb 17 13:31:09 crc kubenswrapper[4955]: I0217 13:31:09.234304 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 17 13:31:09 crc kubenswrapper[4955]: I0217 13:31:09.234319 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 17 13:31:09 crc kubenswrapper[4955]: I0217 13:31:09.234319 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 17 13:31:09 crc kubenswrapper[4955]: I0217 13:31:09.244832 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bpbzk"] Feb 17 13:31:09 crc kubenswrapper[4955]: I0217 13:31:09.363363 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qdsk\" (UniqueName: \"kubernetes.io/projected/682aa1ae-e0dc-476d-862d-45071ee0609a-kube-api-access-4qdsk\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bpbzk\" (UID: \"682aa1ae-e0dc-476d-862d-45071ee0609a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bpbzk" Feb 17 13:31:09 crc kubenswrapper[4955]: I0217 13:31:09.363476 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/682aa1ae-e0dc-476d-862d-45071ee0609a-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bpbzk\" (UID: \"682aa1ae-e0dc-476d-862d-45071ee0609a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bpbzk" Feb 17 13:31:09 crc kubenswrapper[4955]: I0217 13:31:09.363700 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/682aa1ae-e0dc-476d-862d-45071ee0609a-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bpbzk\" (UID: \"682aa1ae-e0dc-476d-862d-45071ee0609a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bpbzk" Feb 17 13:31:09 crc kubenswrapper[4955]: I0217 13:31:09.466336 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/682aa1ae-e0dc-476d-862d-45071ee0609a-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bpbzk\" (UID: \"682aa1ae-e0dc-476d-862d-45071ee0609a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bpbzk" Feb 17 13:31:09 crc kubenswrapper[4955]: I0217 13:31:09.466433 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qdsk\" (UniqueName: \"kubernetes.io/projected/682aa1ae-e0dc-476d-862d-45071ee0609a-kube-api-access-4qdsk\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bpbzk\" (UID: \"682aa1ae-e0dc-476d-862d-45071ee0609a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bpbzk" Feb 17 13:31:09 crc kubenswrapper[4955]: I0217 13:31:09.466482 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/682aa1ae-e0dc-476d-862d-45071ee0609a-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bpbzk\" (UID: \"682aa1ae-e0dc-476d-862d-45071ee0609a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bpbzk" Feb 17 13:31:09 crc kubenswrapper[4955]: I0217 13:31:09.474549 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/682aa1ae-e0dc-476d-862d-45071ee0609a-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bpbzk\" (UID: \"682aa1ae-e0dc-476d-862d-45071ee0609a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bpbzk" Feb 17 13:31:09 crc kubenswrapper[4955]: I0217 13:31:09.475124 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/682aa1ae-e0dc-476d-862d-45071ee0609a-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bpbzk\" (UID: \"682aa1ae-e0dc-476d-862d-45071ee0609a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bpbzk" Feb 17 13:31:09 crc kubenswrapper[4955]: I0217 13:31:09.486665 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qdsk\" (UniqueName: \"kubernetes.io/projected/682aa1ae-e0dc-476d-862d-45071ee0609a-kube-api-access-4qdsk\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bpbzk\" (UID: \"682aa1ae-e0dc-476d-862d-45071ee0609a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bpbzk" Feb 17 13:31:09 crc kubenswrapper[4955]: I0217 13:31:09.548264 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bpbzk" Feb 17 13:31:10 crc kubenswrapper[4955]: I0217 13:31:10.128875 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bpbzk"] Feb 17 13:31:10 crc kubenswrapper[4955]: I0217 13:31:10.145106 4955 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 17 13:31:10 crc kubenswrapper[4955]: I0217 13:31:10.155634 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bpbzk" event={"ID":"682aa1ae-e0dc-476d-862d-45071ee0609a","Type":"ContainerStarted","Data":"023fe86d0d3bdbd4c5f6b52cc660b52ab33ac965cbe489f70bbf04db6792dff9"} Feb 17 13:31:11 crc kubenswrapper[4955]: I0217 13:31:11.164796 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bpbzk" event={"ID":"682aa1ae-e0dc-476d-862d-45071ee0609a","Type":"ContainerStarted","Data":"d642d4aeb0e8ba4d91d44b70920f73126cd4bdb7ce53260a0861123908ae5924"} Feb 17 13:31:11 crc kubenswrapper[4955]: I0217 13:31:11.181804 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bpbzk" podStartSLOduration=1.751537285 podStartE2EDuration="2.181768716s" podCreationTimestamp="2026-02-17 13:31:09 +0000 UTC" firstStartedPulling="2026-02-17 13:31:10.144867712 +0000 UTC m=+1608.667597255" lastFinishedPulling="2026-02-17 13:31:10.575099143 +0000 UTC m=+1609.097828686" observedRunningTime="2026-02-17 13:31:11.179863882 +0000 UTC m=+1609.702593435" watchObservedRunningTime="2026-02-17 13:31:11.181768716 +0000 UTC m=+1609.704498259" Feb 17 13:31:19 crc kubenswrapper[4955]: I0217 13:31:19.223302 4955 scope.go:117] "RemoveContainer" containerID="969807d9ca63cdcc1aa7cc07aa21a249802167bf58b0b6150149361ce19b14ca" Feb 17 13:31:19 crc kubenswrapper[4955]: E0217 13:31:19.223991 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:31:26 crc kubenswrapper[4955]: I0217 13:31:26.041015 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-509d-account-create-update-pfzj7"] Feb 17 13:31:26 crc kubenswrapper[4955]: I0217 13:31:26.048675 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-9f71-account-create-update-fhns5"] Feb 17 13:31:26 crc kubenswrapper[4955]: I0217 13:31:26.056008 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-nbjcj"] Feb 17 13:31:26 crc kubenswrapper[4955]: I0217 13:31:26.065538 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-509d-account-create-update-pfzj7"] Feb 17 13:31:26 crc kubenswrapper[4955]: I0217 13:31:26.073991 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-9f71-account-create-update-fhns5"] Feb 17 13:31:26 crc kubenswrapper[4955]: I0217 13:31:26.082136 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-nbjcj"] Feb 17 13:31:26 crc kubenswrapper[4955]: I0217 13:31:26.237918 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19e2a114-5254-4281-9d4d-209c26dcb1d6" path="/var/lib/kubelet/pods/19e2a114-5254-4281-9d4d-209c26dcb1d6/volumes" Feb 17 13:31:26 crc kubenswrapper[4955]: I0217 13:31:26.242799 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96770935-7ba8-440a-8947-28a74cd86f08" path="/var/lib/kubelet/pods/96770935-7ba8-440a-8947-28a74cd86f08/volumes" Feb 17 13:31:26 crc kubenswrapper[4955]: I0217 13:31:26.244786 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f54e2d0a-6914-42a1-a9d0-f8c6c5e207b0" path="/var/lib/kubelet/pods/f54e2d0a-6914-42a1-a9d0-f8c6c5e207b0/volumes" Feb 17 13:31:27 crc kubenswrapper[4955]: I0217 13:31:27.027821 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-7gmhj"] Feb 17 13:31:27 crc kubenswrapper[4955]: I0217 13:31:27.037692 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-hvwtj"] Feb 17 13:31:27 crc kubenswrapper[4955]: I0217 13:31:27.046340 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-hvwtj"] Feb 17 13:31:27 crc kubenswrapper[4955]: I0217 13:31:27.055623 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-be28-account-create-update-xx2j6"] Feb 17 13:31:27 crc kubenswrapper[4955]: I0217 13:31:27.063921 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-7gmhj"] Feb 17 13:31:27 crc kubenswrapper[4955]: I0217 13:31:27.072450 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-be28-account-create-update-xx2j6"] Feb 17 13:31:28 crc kubenswrapper[4955]: I0217 13:31:28.235346 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d8307b9-b38d-44ed-830f-c0cb10538496" path="/var/lib/kubelet/pods/2d8307b9-b38d-44ed-830f-c0cb10538496/volumes" Feb 17 13:31:28 crc kubenswrapper[4955]: I0217 13:31:28.237225 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a2f748d-5c00-4975-8888-cdea65d3ab36" path="/var/lib/kubelet/pods/7a2f748d-5c00-4975-8888-cdea65d3ab36/volumes" Feb 17 13:31:28 crc kubenswrapper[4955]: I0217 13:31:28.238734 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="966dab54-42a6-452a-8459-e76ec889c9ee" path="/var/lib/kubelet/pods/966dab54-42a6-452a-8459-e76ec889c9ee/volumes" Feb 17 13:31:30 crc kubenswrapper[4955]: I0217 13:31:30.223721 4955 scope.go:117] "RemoveContainer" containerID="969807d9ca63cdcc1aa7cc07aa21a249802167bf58b0b6150149361ce19b14ca" Feb 17 13:31:30 crc kubenswrapper[4955]: E0217 13:31:30.224203 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:31:31 crc kubenswrapper[4955]: I0217 13:31:31.026520 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-hwwtj"] Feb 17 13:31:31 crc kubenswrapper[4955]: I0217 13:31:31.034706 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-hwwtj"] Feb 17 13:31:32 crc kubenswrapper[4955]: I0217 13:31:32.236877 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7293150-5c3f-4348-8f9e-2f46a2373d1d" path="/var/lib/kubelet/pods/b7293150-5c3f-4348-8f9e-2f46a2373d1d/volumes" Feb 17 13:31:44 crc kubenswrapper[4955]: I0217 13:31:44.223695 4955 scope.go:117] "RemoveContainer" containerID="969807d9ca63cdcc1aa7cc07aa21a249802167bf58b0b6150149361ce19b14ca" Feb 17 13:31:44 crc kubenswrapper[4955]: E0217 13:31:44.224531 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:31:48 crc kubenswrapper[4955]: I0217 13:31:48.823636 4955 scope.go:117] "RemoveContainer" containerID="4a2cd1a1e878c8b153deec8babc4bfeb329265cf49676349093f4f2c18eb192d" Feb 17 13:31:48 crc kubenswrapper[4955]: I0217 13:31:48.846178 4955 scope.go:117] "RemoveContainer" containerID="52440ed4e221c2435cc3b96478a7d6c2a4fa2acac333904752ec3fc71ae352ff" Feb 17 13:31:48 crc kubenswrapper[4955]: I0217 13:31:48.891950 4955 scope.go:117] "RemoveContainer" containerID="05c87d393167e736f7b09dc84dbdd5af5657783de253f846a36d99cbf6457afa" Feb 17 13:31:48 crc kubenswrapper[4955]: I0217 13:31:48.915372 4955 scope.go:117] "RemoveContainer" containerID="4a08ccf38d2598943d136d7704d1b698199baa699a28452a09e8febea77a5314" Feb 17 13:31:48 crc kubenswrapper[4955]: I0217 13:31:48.963722 4955 scope.go:117] "RemoveContainer" containerID="9e4eee87ea3f68861eadb36091cc1bcbece301c8473a28c4c868c5197488131a" Feb 17 13:31:48 crc kubenswrapper[4955]: I0217 13:31:48.984279 4955 scope.go:117] "RemoveContainer" containerID="f5994e3d18d3af598e883e39ab79069f680f8acf585814862c87251b0d202062" Feb 17 13:31:49 crc kubenswrapper[4955]: I0217 13:31:49.028485 4955 scope.go:117] "RemoveContainer" containerID="bb5673a151e591e332e71278606624a994c4a495305921eb69fc3e8fe412fcb5" Feb 17 13:31:49 crc kubenswrapper[4955]: I0217 13:31:49.084080 4955 scope.go:117] "RemoveContainer" containerID="db6ad3bcc9953fcc4a859c6bd4fe41a71d2e8ddd5a03484b89669ef344b6c13b" Feb 17 13:31:49 crc kubenswrapper[4955]: I0217 13:31:49.118141 4955 scope.go:117] "RemoveContainer" containerID="04c8cf2fb0cdb10eb8ec8fc8920721e50490e64dc925d5eec0b2cad40553a688" Feb 17 13:31:51 crc kubenswrapper[4955]: I0217 13:31:51.033716 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-jwz6b"] Feb 17 13:31:51 crc kubenswrapper[4955]: I0217 13:31:51.044153 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-jwz6b"] Feb 17 13:31:52 crc kubenswrapper[4955]: I0217 13:31:52.234620 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aed5c07b-3d7a-43c1-a11f-4a4220289bef" path="/var/lib/kubelet/pods/aed5c07b-3d7a-43c1-a11f-4a4220289bef/volumes" Feb 17 13:31:56 crc kubenswrapper[4955]: I0217 13:31:56.223259 4955 scope.go:117] "RemoveContainer" containerID="969807d9ca63cdcc1aa7cc07aa21a249802167bf58b0b6150149361ce19b14ca" Feb 17 13:31:56 crc kubenswrapper[4955]: E0217 13:31:56.223874 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:32:02 crc kubenswrapper[4955]: I0217 13:32:02.050563 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-3d6f-account-create-update-zsh9p"] Feb 17 13:32:02 crc kubenswrapper[4955]: I0217 13:32:02.072224 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-3d6f-account-create-update-zsh9p"] Feb 17 13:32:02 crc kubenswrapper[4955]: I0217 13:32:02.233982 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cb66d5d-60a5-486d-b9ff-35e2ebeb4d8b" path="/var/lib/kubelet/pods/7cb66d5d-60a5-486d-b9ff-35e2ebeb4d8b/volumes" Feb 17 13:32:04 crc kubenswrapper[4955]: I0217 13:32:04.031282 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-tqmd4"] Feb 17 13:32:04 crc kubenswrapper[4955]: I0217 13:32:04.041694 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-tqmd4"] Feb 17 13:32:04 crc kubenswrapper[4955]: I0217 13:32:04.052263 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-ab8c-account-create-update-49rjx"] Feb 17 13:32:04 crc kubenswrapper[4955]: I0217 13:32:04.061261 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-4mktx"] Feb 17 13:32:04 crc kubenswrapper[4955]: I0217 13:32:04.069065 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-a0dc-account-create-update-rbbn6"] Feb 17 13:32:04 crc kubenswrapper[4955]: I0217 13:32:04.076967 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-ab8c-account-create-update-49rjx"] Feb 17 13:32:04 crc kubenswrapper[4955]: I0217 13:32:04.084600 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-6ss66"] Feb 17 13:32:04 crc kubenswrapper[4955]: I0217 13:32:04.092614 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-a0dc-account-create-update-rbbn6"] Feb 17 13:32:04 crc kubenswrapper[4955]: I0217 13:32:04.101247 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-6ss66"] Feb 17 13:32:04 crc kubenswrapper[4955]: I0217 13:32:04.109664 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-4mktx"] Feb 17 13:32:04 crc kubenswrapper[4955]: I0217 13:32:04.232696 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fa427d9-5f2e-4cb9-934b-edd1086e0f42" path="/var/lib/kubelet/pods/2fa427d9-5f2e-4cb9-934b-edd1086e0f42/volumes" Feb 17 13:32:04 crc kubenswrapper[4955]: I0217 13:32:04.234143 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4da596f3-7aff-4cfd-984e-0fd58d82a25c" path="/var/lib/kubelet/pods/4da596f3-7aff-4cfd-984e-0fd58d82a25c/volumes" Feb 17 13:32:04 crc kubenswrapper[4955]: I0217 13:32:04.236415 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4fe8b1c0-db08-4b58-b1cb-218631c424ab" path="/var/lib/kubelet/pods/4fe8b1c0-db08-4b58-b1cb-218631c424ab/volumes" Feb 17 13:32:04 crc kubenswrapper[4955]: I0217 13:32:04.238383 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a61deb1d-ae88-49f8-8954-2059c469e84f" path="/var/lib/kubelet/pods/a61deb1d-ae88-49f8-8954-2059c469e84f/volumes" Feb 17 13:32:04 crc kubenswrapper[4955]: I0217 13:32:04.240083 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd1f70a3-aaee-464a-aa89-621279758849" path="/var/lib/kubelet/pods/dd1f70a3-aaee-464a-aa89-621279758849/volumes" Feb 17 13:32:10 crc kubenswrapper[4955]: I0217 13:32:10.223272 4955 scope.go:117] "RemoveContainer" containerID="969807d9ca63cdcc1aa7cc07aa21a249802167bf58b0b6150149361ce19b14ca" Feb 17 13:32:10 crc kubenswrapper[4955]: E0217 13:32:10.223817 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:32:13 crc kubenswrapper[4955]: I0217 13:32:13.033154 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-hcggw"] Feb 17 13:32:13 crc kubenswrapper[4955]: I0217 13:32:13.041275 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-hcggw"] Feb 17 13:32:14 crc kubenswrapper[4955]: I0217 13:32:14.233051 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a" path="/var/lib/kubelet/pods/25f90ea4-b3cf-4dcf-9449-c2adf8f7f60a/volumes" Feb 17 13:32:25 crc kubenswrapper[4955]: I0217 13:32:25.222884 4955 scope.go:117] "RemoveContainer" containerID="969807d9ca63cdcc1aa7cc07aa21a249802167bf58b0b6150149361ce19b14ca" Feb 17 13:32:25 crc kubenswrapper[4955]: E0217 13:32:25.224074 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:32:33 crc kubenswrapper[4955]: I0217 13:32:33.896859 4955 generic.go:334] "Generic (PLEG): container finished" podID="682aa1ae-e0dc-476d-862d-45071ee0609a" containerID="d642d4aeb0e8ba4d91d44b70920f73126cd4bdb7ce53260a0861123908ae5924" exitCode=0 Feb 17 13:32:33 crc kubenswrapper[4955]: I0217 13:32:33.896922 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bpbzk" event={"ID":"682aa1ae-e0dc-476d-862d-45071ee0609a","Type":"ContainerDied","Data":"d642d4aeb0e8ba4d91d44b70920f73126cd4bdb7ce53260a0861123908ae5924"} Feb 17 13:32:35 crc kubenswrapper[4955]: I0217 13:32:35.359449 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bpbzk" Feb 17 13:32:35 crc kubenswrapper[4955]: I0217 13:32:35.453028 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/682aa1ae-e0dc-476d-862d-45071ee0609a-inventory\") pod \"682aa1ae-e0dc-476d-862d-45071ee0609a\" (UID: \"682aa1ae-e0dc-476d-862d-45071ee0609a\") " Feb 17 13:32:35 crc kubenswrapper[4955]: I0217 13:32:35.453125 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qdsk\" (UniqueName: \"kubernetes.io/projected/682aa1ae-e0dc-476d-862d-45071ee0609a-kube-api-access-4qdsk\") pod \"682aa1ae-e0dc-476d-862d-45071ee0609a\" (UID: \"682aa1ae-e0dc-476d-862d-45071ee0609a\") " Feb 17 13:32:35 crc kubenswrapper[4955]: I0217 13:32:35.453256 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/682aa1ae-e0dc-476d-862d-45071ee0609a-ssh-key-openstack-edpm-ipam\") pod \"682aa1ae-e0dc-476d-862d-45071ee0609a\" (UID: \"682aa1ae-e0dc-476d-862d-45071ee0609a\") " Feb 17 13:32:35 crc kubenswrapper[4955]: I0217 13:32:35.464189 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/682aa1ae-e0dc-476d-862d-45071ee0609a-kube-api-access-4qdsk" (OuterVolumeSpecName: "kube-api-access-4qdsk") pod "682aa1ae-e0dc-476d-862d-45071ee0609a" (UID: "682aa1ae-e0dc-476d-862d-45071ee0609a"). InnerVolumeSpecName "kube-api-access-4qdsk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:32:35 crc kubenswrapper[4955]: E0217 13:32:35.477809 4955 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/682aa1ae-e0dc-476d-862d-45071ee0609a-inventory podName:682aa1ae-e0dc-476d-862d-45071ee0609a nodeName:}" failed. No retries permitted until 2026-02-17 13:32:35.97775554 +0000 UTC m=+1694.500485093 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "inventory" (UniqueName: "kubernetes.io/secret/682aa1ae-e0dc-476d-862d-45071ee0609a-inventory") pod "682aa1ae-e0dc-476d-862d-45071ee0609a" (UID: "682aa1ae-e0dc-476d-862d-45071ee0609a") : error deleting /var/lib/kubelet/pods/682aa1ae-e0dc-476d-862d-45071ee0609a/volume-subpaths: remove /var/lib/kubelet/pods/682aa1ae-e0dc-476d-862d-45071ee0609a/volume-subpaths: no such file or directory Feb 17 13:32:35 crc kubenswrapper[4955]: I0217 13:32:35.480475 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/682aa1ae-e0dc-476d-862d-45071ee0609a-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "682aa1ae-e0dc-476d-862d-45071ee0609a" (UID: "682aa1ae-e0dc-476d-862d-45071ee0609a"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:32:35 crc kubenswrapper[4955]: I0217 13:32:35.556766 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qdsk\" (UniqueName: \"kubernetes.io/projected/682aa1ae-e0dc-476d-862d-45071ee0609a-kube-api-access-4qdsk\") on node \"crc\" DevicePath \"\"" Feb 17 13:32:35 crc kubenswrapper[4955]: I0217 13:32:35.557104 4955 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/682aa1ae-e0dc-476d-862d-45071ee0609a-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 17 13:32:35 crc kubenswrapper[4955]: I0217 13:32:35.914126 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bpbzk" event={"ID":"682aa1ae-e0dc-476d-862d-45071ee0609a","Type":"ContainerDied","Data":"023fe86d0d3bdbd4c5f6b52cc660b52ab33ac965cbe489f70bbf04db6792dff9"} Feb 17 13:32:35 crc kubenswrapper[4955]: I0217 13:32:35.914169 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="023fe86d0d3bdbd4c5f6b52cc660b52ab33ac965cbe489f70bbf04db6792dff9" Feb 17 13:32:35 crc kubenswrapper[4955]: I0217 13:32:35.914221 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bpbzk" Feb 17 13:32:35 crc kubenswrapper[4955]: I0217 13:32:35.993486 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-ncz59"] Feb 17 13:32:35 crc kubenswrapper[4955]: E0217 13:32:35.993951 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="682aa1ae-e0dc-476d-862d-45071ee0609a" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Feb 17 13:32:35 crc kubenswrapper[4955]: I0217 13:32:35.993975 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="682aa1ae-e0dc-476d-862d-45071ee0609a" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Feb 17 13:32:35 crc kubenswrapper[4955]: I0217 13:32:35.994258 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="682aa1ae-e0dc-476d-862d-45071ee0609a" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Feb 17 13:32:35 crc kubenswrapper[4955]: I0217 13:32:35.994983 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-ncz59" Feb 17 13:32:36 crc kubenswrapper[4955]: I0217 13:32:36.012752 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-ncz59"] Feb 17 13:32:36 crc kubenswrapper[4955]: I0217 13:32:36.067483 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/682aa1ae-e0dc-476d-862d-45071ee0609a-inventory\") pod \"682aa1ae-e0dc-476d-862d-45071ee0609a\" (UID: \"682aa1ae-e0dc-476d-862d-45071ee0609a\") " Feb 17 13:32:36 crc kubenswrapper[4955]: I0217 13:32:36.072675 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/682aa1ae-e0dc-476d-862d-45071ee0609a-inventory" (OuterVolumeSpecName: "inventory") pod "682aa1ae-e0dc-476d-862d-45071ee0609a" (UID: "682aa1ae-e0dc-476d-862d-45071ee0609a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:32:36 crc kubenswrapper[4955]: I0217 13:32:36.171341 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qt67\" (UniqueName: \"kubernetes.io/projected/c74ecb0c-c07e-4196-a0a8-7c107404b60f-kube-api-access-8qt67\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-ncz59\" (UID: \"c74ecb0c-c07e-4196-a0a8-7c107404b60f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-ncz59" Feb 17 13:32:36 crc kubenswrapper[4955]: I0217 13:32:36.171458 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c74ecb0c-c07e-4196-a0a8-7c107404b60f-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-ncz59\" (UID: \"c74ecb0c-c07e-4196-a0a8-7c107404b60f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-ncz59" Feb 17 13:32:36 crc kubenswrapper[4955]: I0217 13:32:36.171516 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c74ecb0c-c07e-4196-a0a8-7c107404b60f-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-ncz59\" (UID: \"c74ecb0c-c07e-4196-a0a8-7c107404b60f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-ncz59" Feb 17 13:32:36 crc kubenswrapper[4955]: I0217 13:32:36.171777 4955 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/682aa1ae-e0dc-476d-862d-45071ee0609a-inventory\") on node \"crc\" DevicePath \"\"" Feb 17 13:32:36 crc kubenswrapper[4955]: I0217 13:32:36.274310 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c74ecb0c-c07e-4196-a0a8-7c107404b60f-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-ncz59\" (UID: \"c74ecb0c-c07e-4196-a0a8-7c107404b60f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-ncz59" Feb 17 13:32:36 crc kubenswrapper[4955]: I0217 13:32:36.274585 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qt67\" (UniqueName: \"kubernetes.io/projected/c74ecb0c-c07e-4196-a0a8-7c107404b60f-kube-api-access-8qt67\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-ncz59\" (UID: \"c74ecb0c-c07e-4196-a0a8-7c107404b60f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-ncz59" Feb 17 13:32:36 crc kubenswrapper[4955]: I0217 13:32:36.274751 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c74ecb0c-c07e-4196-a0a8-7c107404b60f-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-ncz59\" (UID: \"c74ecb0c-c07e-4196-a0a8-7c107404b60f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-ncz59" Feb 17 13:32:36 crc kubenswrapper[4955]: I0217 13:32:36.278619 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c74ecb0c-c07e-4196-a0a8-7c107404b60f-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-ncz59\" (UID: \"c74ecb0c-c07e-4196-a0a8-7c107404b60f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-ncz59" Feb 17 13:32:36 crc kubenswrapper[4955]: I0217 13:32:36.278773 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c74ecb0c-c07e-4196-a0a8-7c107404b60f-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-ncz59\" (UID: \"c74ecb0c-c07e-4196-a0a8-7c107404b60f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-ncz59" Feb 17 13:32:36 crc kubenswrapper[4955]: I0217 13:32:36.293333 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qt67\" (UniqueName: \"kubernetes.io/projected/c74ecb0c-c07e-4196-a0a8-7c107404b60f-kube-api-access-8qt67\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-ncz59\" (UID: \"c74ecb0c-c07e-4196-a0a8-7c107404b60f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-ncz59" Feb 17 13:32:36 crc kubenswrapper[4955]: I0217 13:32:36.315593 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-ncz59" Feb 17 13:32:36 crc kubenswrapper[4955]: I0217 13:32:36.861361 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-ncz59"] Feb 17 13:32:36 crc kubenswrapper[4955]: I0217 13:32:36.925008 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-ncz59" event={"ID":"c74ecb0c-c07e-4196-a0a8-7c107404b60f","Type":"ContainerStarted","Data":"47d4d51976aa71a6ed39d7eb0c1b7bd07210e74580fb7f7cfac3a79e7f979022"} Feb 17 13:32:37 crc kubenswrapper[4955]: I0217 13:32:37.223547 4955 scope.go:117] "RemoveContainer" containerID="969807d9ca63cdcc1aa7cc07aa21a249802167bf58b0b6150149361ce19b14ca" Feb 17 13:32:37 crc kubenswrapper[4955]: E0217 13:32:37.223852 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:32:37 crc kubenswrapper[4955]: I0217 13:32:37.940317 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-ncz59" event={"ID":"c74ecb0c-c07e-4196-a0a8-7c107404b60f","Type":"ContainerStarted","Data":"387acad9fbd30abba4b5f5e9a0f642c9ca2e5093b8adffbbeb792842369b41c9"} Feb 17 13:32:37 crc kubenswrapper[4955]: I0217 13:32:37.968072 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-ncz59" podStartSLOduration=2.518225142 podStartE2EDuration="2.968049226s" podCreationTimestamp="2026-02-17 13:32:35 +0000 UTC" firstStartedPulling="2026-02-17 13:32:36.873718697 +0000 UTC m=+1695.396448240" lastFinishedPulling="2026-02-17 13:32:37.323542781 +0000 UTC m=+1695.846272324" observedRunningTime="2026-02-17 13:32:37.964504442 +0000 UTC m=+1696.487233995" watchObservedRunningTime="2026-02-17 13:32:37.968049226 +0000 UTC m=+1696.490778779" Feb 17 13:32:49 crc kubenswrapper[4955]: I0217 13:32:49.049453 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-48hsq"] Feb 17 13:32:49 crc kubenswrapper[4955]: I0217 13:32:49.058595 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-48hsq"] Feb 17 13:32:49 crc kubenswrapper[4955]: I0217 13:32:49.223184 4955 scope.go:117] "RemoveContainer" containerID="969807d9ca63cdcc1aa7cc07aa21a249802167bf58b0b6150149361ce19b14ca" Feb 17 13:32:49 crc kubenswrapper[4955]: E0217 13:32:49.223468 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:32:49 crc kubenswrapper[4955]: I0217 13:32:49.310081 4955 scope.go:117] "RemoveContainer" containerID="498a728e83f06ac0535fc1f340791bb545f474bf48353cf58ae5b6346f916cb8" Feb 17 13:32:49 crc kubenswrapper[4955]: I0217 13:32:49.360735 4955 scope.go:117] "RemoveContainer" containerID="b171b99fad766be9125f2dc416c5ba69a27c46624d56bc5d39d482c4c9459ea1" Feb 17 13:32:49 crc kubenswrapper[4955]: I0217 13:32:49.408504 4955 scope.go:117] "RemoveContainer" containerID="9b4356ed08e500a83ad864d58ae6b2c7ec3a5ce1ae99e89074f50655811b46dd" Feb 17 13:32:49 crc kubenswrapper[4955]: I0217 13:32:49.443672 4955 scope.go:117] "RemoveContainer" containerID="06a07877376e2fb02c7296215cfcdc234914b8b090638fa038a611fbdddd137f" Feb 17 13:32:49 crc kubenswrapper[4955]: I0217 13:32:49.514259 4955 scope.go:117] "RemoveContainer" containerID="f4e8cff5f113bfbc77a2f69000f3d32f76c5c84fd69217f12e593f37eb3f93f1" Feb 17 13:32:49 crc kubenswrapper[4955]: I0217 13:32:49.574129 4955 scope.go:117] "RemoveContainer" containerID="ea54a9d6751ef4e8c620fe8828d31a2851d55b7afa7e29803b3bd5f40fcd3ddb" Feb 17 13:32:49 crc kubenswrapper[4955]: I0217 13:32:49.613043 4955 scope.go:117] "RemoveContainer" containerID="bb55ac7b0c2ff9916def101839e6a50422bafb9dc82d108b0de43548d2d99d62" Feb 17 13:32:49 crc kubenswrapper[4955]: I0217 13:32:49.635969 4955 scope.go:117] "RemoveContainer" containerID="c0f874e3765ae3cf7397baf8468592643150f43238602ea03aaaffc680bb61f5" Feb 17 13:32:49 crc kubenswrapper[4955]: I0217 13:32:49.657419 4955 scope.go:117] "RemoveContainer" containerID="95d036ca02fb8de95169f573c8c35bc0d7f7a2b7c41b806464700bc89b2cd855" Feb 17 13:32:50 crc kubenswrapper[4955]: I0217 13:32:50.237531 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be85ffb8-5128-4903-9f7d-cdd462a44b95" path="/var/lib/kubelet/pods/be85ffb8-5128-4903-9f7d-cdd462a44b95/volumes" Feb 17 13:32:53 crc kubenswrapper[4955]: I0217 13:32:53.029064 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-nfdxn"] Feb 17 13:32:53 crc kubenswrapper[4955]: I0217 13:32:53.037797 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-nfdxn"] Feb 17 13:32:54 crc kubenswrapper[4955]: I0217 13:32:54.236562 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="700f73a1-f56d-4d4e-9822-438472980124" path="/var/lib/kubelet/pods/700f73a1-f56d-4d4e-9822-438472980124/volumes" Feb 17 13:33:00 crc kubenswrapper[4955]: I0217 13:33:00.223919 4955 scope.go:117] "RemoveContainer" containerID="969807d9ca63cdcc1aa7cc07aa21a249802167bf58b0b6150149361ce19b14ca" Feb 17 13:33:00 crc kubenswrapper[4955]: E0217 13:33:00.225298 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:33:01 crc kubenswrapper[4955]: I0217 13:33:01.060279 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-sstln"] Feb 17 13:33:01 crc kubenswrapper[4955]: I0217 13:33:01.069563 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-sstln"] Feb 17 13:33:02 crc kubenswrapper[4955]: I0217 13:33:02.233194 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50465113-9c90-4a32-8485-0b976374107a" path="/var/lib/kubelet/pods/50465113-9c90-4a32-8485-0b976374107a/volumes" Feb 17 13:33:12 crc kubenswrapper[4955]: I0217 13:33:12.031828 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-tr2q9"] Feb 17 13:33:12 crc kubenswrapper[4955]: I0217 13:33:12.040946 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-tr2q9"] Feb 17 13:33:12 crc kubenswrapper[4955]: I0217 13:33:12.236398 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d695519-3143-4f53-ab46-5823bc04dc0d" path="/var/lib/kubelet/pods/1d695519-3143-4f53-ab46-5823bc04dc0d/volumes" Feb 17 13:33:13 crc kubenswrapper[4955]: I0217 13:33:13.225461 4955 scope.go:117] "RemoveContainer" containerID="969807d9ca63cdcc1aa7cc07aa21a249802167bf58b0b6150149361ce19b14ca" Feb 17 13:33:13 crc kubenswrapper[4955]: E0217 13:33:13.225937 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:33:14 crc kubenswrapper[4955]: I0217 13:33:14.033521 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-w6s4s"] Feb 17 13:33:14 crc kubenswrapper[4955]: I0217 13:33:14.042264 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-w6s4s"] Feb 17 13:33:14 crc kubenswrapper[4955]: I0217 13:33:14.235160 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1c2d0c5-7647-447f-a1a6-b37d8c09310c" path="/var/lib/kubelet/pods/e1c2d0c5-7647-447f-a1a6-b37d8c09310c/volumes" Feb 17 13:33:27 crc kubenswrapper[4955]: I0217 13:33:27.223326 4955 scope.go:117] "RemoveContainer" containerID="969807d9ca63cdcc1aa7cc07aa21a249802167bf58b0b6150149361ce19b14ca" Feb 17 13:33:27 crc kubenswrapper[4955]: E0217 13:33:27.224222 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:33:41 crc kubenswrapper[4955]: I0217 13:33:41.223716 4955 scope.go:117] "RemoveContainer" containerID="969807d9ca63cdcc1aa7cc07aa21a249802167bf58b0b6150149361ce19b14ca" Feb 17 13:33:41 crc kubenswrapper[4955]: E0217 13:33:41.226114 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:33:44 crc kubenswrapper[4955]: I0217 13:33:44.567250 4955 generic.go:334] "Generic (PLEG): container finished" podID="c74ecb0c-c07e-4196-a0a8-7c107404b60f" containerID="387acad9fbd30abba4b5f5e9a0f642c9ca2e5093b8adffbbeb792842369b41c9" exitCode=0 Feb 17 13:33:44 crc kubenswrapper[4955]: I0217 13:33:44.567399 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-ncz59" event={"ID":"c74ecb0c-c07e-4196-a0a8-7c107404b60f","Type":"ContainerDied","Data":"387acad9fbd30abba4b5f5e9a0f642c9ca2e5093b8adffbbeb792842369b41c9"} Feb 17 13:33:45 crc kubenswrapper[4955]: I0217 13:33:45.990137 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-ncz59" Feb 17 13:33:46 crc kubenswrapper[4955]: I0217 13:33:46.069558 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c74ecb0c-c07e-4196-a0a8-7c107404b60f-ssh-key-openstack-edpm-ipam\") pod \"c74ecb0c-c07e-4196-a0a8-7c107404b60f\" (UID: \"c74ecb0c-c07e-4196-a0a8-7c107404b60f\") " Feb 17 13:33:46 crc kubenswrapper[4955]: I0217 13:33:46.070338 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c74ecb0c-c07e-4196-a0a8-7c107404b60f-inventory\") pod \"c74ecb0c-c07e-4196-a0a8-7c107404b60f\" (UID: \"c74ecb0c-c07e-4196-a0a8-7c107404b60f\") " Feb 17 13:33:46 crc kubenswrapper[4955]: I0217 13:33:46.070373 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8qt67\" (UniqueName: \"kubernetes.io/projected/c74ecb0c-c07e-4196-a0a8-7c107404b60f-kube-api-access-8qt67\") pod \"c74ecb0c-c07e-4196-a0a8-7c107404b60f\" (UID: \"c74ecb0c-c07e-4196-a0a8-7c107404b60f\") " Feb 17 13:33:46 crc kubenswrapper[4955]: I0217 13:33:46.086139 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c74ecb0c-c07e-4196-a0a8-7c107404b60f-kube-api-access-8qt67" (OuterVolumeSpecName: "kube-api-access-8qt67") pod "c74ecb0c-c07e-4196-a0a8-7c107404b60f" (UID: "c74ecb0c-c07e-4196-a0a8-7c107404b60f"). InnerVolumeSpecName "kube-api-access-8qt67". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:33:46 crc kubenswrapper[4955]: I0217 13:33:46.103681 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c74ecb0c-c07e-4196-a0a8-7c107404b60f-inventory" (OuterVolumeSpecName: "inventory") pod "c74ecb0c-c07e-4196-a0a8-7c107404b60f" (UID: "c74ecb0c-c07e-4196-a0a8-7c107404b60f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:33:46 crc kubenswrapper[4955]: I0217 13:33:46.109624 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c74ecb0c-c07e-4196-a0a8-7c107404b60f-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "c74ecb0c-c07e-4196-a0a8-7c107404b60f" (UID: "c74ecb0c-c07e-4196-a0a8-7c107404b60f"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:33:46 crc kubenswrapper[4955]: I0217 13:33:46.174152 4955 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c74ecb0c-c07e-4196-a0a8-7c107404b60f-inventory\") on node \"crc\" DevicePath \"\"" Feb 17 13:33:46 crc kubenswrapper[4955]: I0217 13:33:46.174221 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8qt67\" (UniqueName: \"kubernetes.io/projected/c74ecb0c-c07e-4196-a0a8-7c107404b60f-kube-api-access-8qt67\") on node \"crc\" DevicePath \"\"" Feb 17 13:33:46 crc kubenswrapper[4955]: I0217 13:33:46.174238 4955 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c74ecb0c-c07e-4196-a0a8-7c107404b60f-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 17 13:33:46 crc kubenswrapper[4955]: I0217 13:33:46.587456 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-ncz59" event={"ID":"c74ecb0c-c07e-4196-a0a8-7c107404b60f","Type":"ContainerDied","Data":"47d4d51976aa71a6ed39d7eb0c1b7bd07210e74580fb7f7cfac3a79e7f979022"} Feb 17 13:33:46 crc kubenswrapper[4955]: I0217 13:33:46.587855 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47d4d51976aa71a6ed39d7eb0c1b7bd07210e74580fb7f7cfac3a79e7f979022" Feb 17 13:33:46 crc kubenswrapper[4955]: I0217 13:33:46.587497 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-ncz59" Feb 17 13:33:46 crc kubenswrapper[4955]: I0217 13:33:46.670990 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jr9rx"] Feb 17 13:33:46 crc kubenswrapper[4955]: E0217 13:33:46.671415 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c74ecb0c-c07e-4196-a0a8-7c107404b60f" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 17 13:33:46 crc kubenswrapper[4955]: I0217 13:33:46.671433 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="c74ecb0c-c07e-4196-a0a8-7c107404b60f" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 17 13:33:46 crc kubenswrapper[4955]: I0217 13:33:46.671610 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="c74ecb0c-c07e-4196-a0a8-7c107404b60f" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 17 13:33:46 crc kubenswrapper[4955]: I0217 13:33:46.672306 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jr9rx" Feb 17 13:33:46 crc kubenswrapper[4955]: I0217 13:33:46.675455 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-t9lgc" Feb 17 13:33:46 crc kubenswrapper[4955]: I0217 13:33:46.675901 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 17 13:33:46 crc kubenswrapper[4955]: I0217 13:33:46.676045 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 17 13:33:46 crc kubenswrapper[4955]: I0217 13:33:46.676241 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 17 13:33:46 crc kubenswrapper[4955]: I0217 13:33:46.686669 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b743858f-111b-4d79-aeaf-6b9608fecbac-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-jr9rx\" (UID: \"b743858f-111b-4d79-aeaf-6b9608fecbac\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jr9rx" Feb 17 13:33:46 crc kubenswrapper[4955]: I0217 13:33:46.686733 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bn97g\" (UniqueName: \"kubernetes.io/projected/b743858f-111b-4d79-aeaf-6b9608fecbac-kube-api-access-bn97g\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-jr9rx\" (UID: \"b743858f-111b-4d79-aeaf-6b9608fecbac\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jr9rx" Feb 17 13:33:46 crc kubenswrapper[4955]: I0217 13:33:46.687448 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b743858f-111b-4d79-aeaf-6b9608fecbac-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-jr9rx\" (UID: \"b743858f-111b-4d79-aeaf-6b9608fecbac\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jr9rx" Feb 17 13:33:46 crc kubenswrapper[4955]: I0217 13:33:46.691721 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jr9rx"] Feb 17 13:33:46 crc kubenswrapper[4955]: I0217 13:33:46.789673 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b743858f-111b-4d79-aeaf-6b9608fecbac-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-jr9rx\" (UID: \"b743858f-111b-4d79-aeaf-6b9608fecbac\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jr9rx" Feb 17 13:33:46 crc kubenswrapper[4955]: I0217 13:33:46.789870 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b743858f-111b-4d79-aeaf-6b9608fecbac-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-jr9rx\" (UID: \"b743858f-111b-4d79-aeaf-6b9608fecbac\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jr9rx" Feb 17 13:33:46 crc kubenswrapper[4955]: I0217 13:33:46.789896 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bn97g\" (UniqueName: \"kubernetes.io/projected/b743858f-111b-4d79-aeaf-6b9608fecbac-kube-api-access-bn97g\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-jr9rx\" (UID: \"b743858f-111b-4d79-aeaf-6b9608fecbac\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jr9rx" Feb 17 13:33:46 crc kubenswrapper[4955]: I0217 13:33:46.794144 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b743858f-111b-4d79-aeaf-6b9608fecbac-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-jr9rx\" (UID: \"b743858f-111b-4d79-aeaf-6b9608fecbac\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jr9rx" Feb 17 13:33:46 crc kubenswrapper[4955]: I0217 13:33:46.794722 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b743858f-111b-4d79-aeaf-6b9608fecbac-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-jr9rx\" (UID: \"b743858f-111b-4d79-aeaf-6b9608fecbac\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jr9rx" Feb 17 13:33:46 crc kubenswrapper[4955]: I0217 13:33:46.808252 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bn97g\" (UniqueName: \"kubernetes.io/projected/b743858f-111b-4d79-aeaf-6b9608fecbac-kube-api-access-bn97g\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-jr9rx\" (UID: \"b743858f-111b-4d79-aeaf-6b9608fecbac\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jr9rx" Feb 17 13:33:46 crc kubenswrapper[4955]: I0217 13:33:46.988340 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jr9rx" Feb 17 13:33:47 crc kubenswrapper[4955]: I0217 13:33:47.545382 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jr9rx"] Feb 17 13:33:47 crc kubenswrapper[4955]: I0217 13:33:47.599513 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jr9rx" event={"ID":"b743858f-111b-4d79-aeaf-6b9608fecbac","Type":"ContainerStarted","Data":"d6b577214b1de25632bb721ac3c8171b7e67ffba2d92dfa6607c082bf3804ae0"} Feb 17 13:33:49 crc kubenswrapper[4955]: I0217 13:33:49.852379 4955 scope.go:117] "RemoveContainer" containerID="823a14c06920720c3445935a196f9ef5e1bcf7bda248b2b9fc3cab24a17e709a" Feb 17 13:33:49 crc kubenswrapper[4955]: I0217 13:33:49.906426 4955 scope.go:117] "RemoveContainer" containerID="1bbb37c0a5d7b5d107d25c7e66b2ba1d7ded8216b4ba2eedae3310ba71cc8d6a" Feb 17 13:33:49 crc kubenswrapper[4955]: I0217 13:33:49.992265 4955 scope.go:117] "RemoveContainer" containerID="e80f7012019a470099c47cd6309d9795aca80788cbfc35c46df2b84756b7c03c" Feb 17 13:33:50 crc kubenswrapper[4955]: I0217 13:33:50.070973 4955 scope.go:117] "RemoveContainer" containerID="d3b62bc3199101ee1ff8f5e21dd262a136664fda289119da42be6248400fa0c1" Feb 17 13:33:50 crc kubenswrapper[4955]: I0217 13:33:50.127561 4955 scope.go:117] "RemoveContainer" containerID="75ea6a58f6a0018b69ba7b7aa87832742c54b55ee898a837323941bb6ed92575" Feb 17 13:33:50 crc kubenswrapper[4955]: I0217 13:33:50.156437 4955 scope.go:117] "RemoveContainer" containerID="cd6b303add2963b8be95890e398356f20f61c8a63f52dd9ccbe7e3a9586c3c6d" Feb 17 13:33:50 crc kubenswrapper[4955]: I0217 13:33:50.236722 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jr9rx" event={"ID":"b743858f-111b-4d79-aeaf-6b9608fecbac","Type":"ContainerStarted","Data":"4e9491f59629d7fd1d5eeb5a0bd2be1c783a2cf37003e47b5df5077e4d3cffca"} Feb 17 13:33:50 crc kubenswrapper[4955]: I0217 13:33:50.259607 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jr9rx" podStartSLOduration=2.498520703 podStartE2EDuration="4.259585488s" podCreationTimestamp="2026-02-17 13:33:46 +0000 UTC" firstStartedPulling="2026-02-17 13:33:47.551368613 +0000 UTC m=+1766.074098156" lastFinishedPulling="2026-02-17 13:33:49.312433398 +0000 UTC m=+1767.835162941" observedRunningTime="2026-02-17 13:33:50.257246861 +0000 UTC m=+1768.779976414" watchObservedRunningTime="2026-02-17 13:33:50.259585488 +0000 UTC m=+1768.782315031" Feb 17 13:33:51 crc kubenswrapper[4955]: I0217 13:33:51.061325 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-9ms22"] Feb 17 13:33:51 crc kubenswrapper[4955]: I0217 13:33:51.070900 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-9ms22"] Feb 17 13:33:52 crc kubenswrapper[4955]: I0217 13:33:52.039490 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-rc66h"] Feb 17 13:33:52 crc kubenswrapper[4955]: I0217 13:33:52.052943 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-301f-account-create-update-wc7xh"] Feb 17 13:33:52 crc kubenswrapper[4955]: I0217 13:33:52.063460 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-6qshk"] Feb 17 13:33:52 crc kubenswrapper[4955]: I0217 13:33:52.075115 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-rc66h"] Feb 17 13:33:52 crc kubenswrapper[4955]: I0217 13:33:52.085310 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-301f-account-create-update-wc7xh"] Feb 17 13:33:52 crc kubenswrapper[4955]: I0217 13:33:52.092747 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-6qshk"] Feb 17 13:33:52 crc kubenswrapper[4955]: I0217 13:33:52.237249 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4031cf36-67c1-425d-98f9-0c4f3a4a6065" path="/var/lib/kubelet/pods/4031cf36-67c1-425d-98f9-0c4f3a4a6065/volumes" Feb 17 13:33:52 crc kubenswrapper[4955]: I0217 13:33:52.239033 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84b8be7f-2891-44a0-beb3-850911ac8b2a" path="/var/lib/kubelet/pods/84b8be7f-2891-44a0-beb3-850911ac8b2a/volumes" Feb 17 13:33:52 crc kubenswrapper[4955]: I0217 13:33:52.239720 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4bbd808-631c-4a9b-81e6-dee199484728" path="/var/lib/kubelet/pods/c4bbd808-631c-4a9b-81e6-dee199484728/volumes" Feb 17 13:33:52 crc kubenswrapper[4955]: I0217 13:33:52.240630 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff0b8526-a0fe-425e-92e8-a4933f3faf61" path="/var/lib/kubelet/pods/ff0b8526-a0fe-425e-92e8-a4933f3faf61/volumes" Feb 17 13:33:53 crc kubenswrapper[4955]: I0217 13:33:53.037669 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-a9af-account-create-update-l42hs"] Feb 17 13:33:53 crc kubenswrapper[4955]: I0217 13:33:53.045508 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-aa00-account-create-update-r9kcs"] Feb 17 13:33:53 crc kubenswrapper[4955]: I0217 13:33:53.053357 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-a9af-account-create-update-l42hs"] Feb 17 13:33:53 crc kubenswrapper[4955]: I0217 13:33:53.062583 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-aa00-account-create-update-r9kcs"] Feb 17 13:33:54 crc kubenswrapper[4955]: I0217 13:33:54.234147 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d0c9416-2de8-4d45-8818-87ae59a2cd6f" path="/var/lib/kubelet/pods/0d0c9416-2de8-4d45-8818-87ae59a2cd6f/volumes" Feb 17 13:33:54 crc kubenswrapper[4955]: I0217 13:33:54.234898 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d747bbd-45e4-4c6c-a6cb-64a259440231" path="/var/lib/kubelet/pods/3d747bbd-45e4-4c6c-a6cb-64a259440231/volumes" Feb 17 13:33:55 crc kubenswrapper[4955]: I0217 13:33:55.222336 4955 scope.go:117] "RemoveContainer" containerID="969807d9ca63cdcc1aa7cc07aa21a249802167bf58b0b6150149361ce19b14ca" Feb 17 13:33:55 crc kubenswrapper[4955]: E0217 13:33:55.222769 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:33:55 crc kubenswrapper[4955]: I0217 13:33:55.287872 4955 generic.go:334] "Generic (PLEG): container finished" podID="b743858f-111b-4d79-aeaf-6b9608fecbac" containerID="4e9491f59629d7fd1d5eeb5a0bd2be1c783a2cf37003e47b5df5077e4d3cffca" exitCode=0 Feb 17 13:33:55 crc kubenswrapper[4955]: I0217 13:33:55.287946 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jr9rx" event={"ID":"b743858f-111b-4d79-aeaf-6b9608fecbac","Type":"ContainerDied","Data":"4e9491f59629d7fd1d5eeb5a0bd2be1c783a2cf37003e47b5df5077e4d3cffca"} Feb 17 13:33:56 crc kubenswrapper[4955]: I0217 13:33:56.750017 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jr9rx" Feb 17 13:33:56 crc kubenswrapper[4955]: I0217 13:33:56.858412 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b743858f-111b-4d79-aeaf-6b9608fecbac-ssh-key-openstack-edpm-ipam\") pod \"b743858f-111b-4d79-aeaf-6b9608fecbac\" (UID: \"b743858f-111b-4d79-aeaf-6b9608fecbac\") " Feb 17 13:33:56 crc kubenswrapper[4955]: I0217 13:33:56.858485 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bn97g\" (UniqueName: \"kubernetes.io/projected/b743858f-111b-4d79-aeaf-6b9608fecbac-kube-api-access-bn97g\") pod \"b743858f-111b-4d79-aeaf-6b9608fecbac\" (UID: \"b743858f-111b-4d79-aeaf-6b9608fecbac\") " Feb 17 13:33:56 crc kubenswrapper[4955]: I0217 13:33:56.858624 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b743858f-111b-4d79-aeaf-6b9608fecbac-inventory\") pod \"b743858f-111b-4d79-aeaf-6b9608fecbac\" (UID: \"b743858f-111b-4d79-aeaf-6b9608fecbac\") " Feb 17 13:33:56 crc kubenswrapper[4955]: I0217 13:33:56.865122 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b743858f-111b-4d79-aeaf-6b9608fecbac-kube-api-access-bn97g" (OuterVolumeSpecName: "kube-api-access-bn97g") pod "b743858f-111b-4d79-aeaf-6b9608fecbac" (UID: "b743858f-111b-4d79-aeaf-6b9608fecbac"). InnerVolumeSpecName "kube-api-access-bn97g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:33:56 crc kubenswrapper[4955]: I0217 13:33:56.888397 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b743858f-111b-4d79-aeaf-6b9608fecbac-inventory" (OuterVolumeSpecName: "inventory") pod "b743858f-111b-4d79-aeaf-6b9608fecbac" (UID: "b743858f-111b-4d79-aeaf-6b9608fecbac"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:33:56 crc kubenswrapper[4955]: I0217 13:33:56.888825 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b743858f-111b-4d79-aeaf-6b9608fecbac-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "b743858f-111b-4d79-aeaf-6b9608fecbac" (UID: "b743858f-111b-4d79-aeaf-6b9608fecbac"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:33:56 crc kubenswrapper[4955]: I0217 13:33:56.960637 4955 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b743858f-111b-4d79-aeaf-6b9608fecbac-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 17 13:33:56 crc kubenswrapper[4955]: I0217 13:33:56.960686 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bn97g\" (UniqueName: \"kubernetes.io/projected/b743858f-111b-4d79-aeaf-6b9608fecbac-kube-api-access-bn97g\") on node \"crc\" DevicePath \"\"" Feb 17 13:33:56 crc kubenswrapper[4955]: I0217 13:33:56.960701 4955 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b743858f-111b-4d79-aeaf-6b9608fecbac-inventory\") on node \"crc\" DevicePath \"\"" Feb 17 13:33:57 crc kubenswrapper[4955]: I0217 13:33:57.312105 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jr9rx" event={"ID":"b743858f-111b-4d79-aeaf-6b9608fecbac","Type":"ContainerDied","Data":"d6b577214b1de25632bb721ac3c8171b7e67ffba2d92dfa6607c082bf3804ae0"} Feb 17 13:33:57 crc kubenswrapper[4955]: I0217 13:33:57.312192 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6b577214b1de25632bb721ac3c8171b7e67ffba2d92dfa6607c082bf3804ae0" Feb 17 13:33:57 crc kubenswrapper[4955]: I0217 13:33:57.312259 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jr9rx" Feb 17 13:33:57 crc kubenswrapper[4955]: I0217 13:33:57.384231 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-5h2ng"] Feb 17 13:33:57 crc kubenswrapper[4955]: E0217 13:33:57.384606 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b743858f-111b-4d79-aeaf-6b9608fecbac" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 17 13:33:57 crc kubenswrapper[4955]: I0217 13:33:57.384629 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="b743858f-111b-4d79-aeaf-6b9608fecbac" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 17 13:33:57 crc kubenswrapper[4955]: I0217 13:33:57.384832 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="b743858f-111b-4d79-aeaf-6b9608fecbac" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 17 13:33:57 crc kubenswrapper[4955]: I0217 13:33:57.385477 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5h2ng" Feb 17 13:33:57 crc kubenswrapper[4955]: I0217 13:33:57.387351 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-t9lgc" Feb 17 13:33:57 crc kubenswrapper[4955]: I0217 13:33:57.395193 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-5h2ng"] Feb 17 13:33:57 crc kubenswrapper[4955]: I0217 13:33:57.402312 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 17 13:33:57 crc kubenswrapper[4955]: I0217 13:33:57.402569 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 17 13:33:57 crc kubenswrapper[4955]: I0217 13:33:57.405830 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 17 13:33:57 crc kubenswrapper[4955]: I0217 13:33:57.573050 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86978b4a-2cb7-4f99-bec3-0afeacdeb7b6-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5h2ng\" (UID: \"86978b4a-2cb7-4f99-bec3-0afeacdeb7b6\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5h2ng" Feb 17 13:33:57 crc kubenswrapper[4955]: I0217 13:33:57.573261 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9st7\" (UniqueName: \"kubernetes.io/projected/86978b4a-2cb7-4f99-bec3-0afeacdeb7b6-kube-api-access-n9st7\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5h2ng\" (UID: \"86978b4a-2cb7-4f99-bec3-0afeacdeb7b6\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5h2ng" Feb 17 13:33:57 crc kubenswrapper[4955]: I0217 13:33:57.573303 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/86978b4a-2cb7-4f99-bec3-0afeacdeb7b6-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5h2ng\" (UID: \"86978b4a-2cb7-4f99-bec3-0afeacdeb7b6\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5h2ng" Feb 17 13:33:57 crc kubenswrapper[4955]: I0217 13:33:57.675357 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86978b4a-2cb7-4f99-bec3-0afeacdeb7b6-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5h2ng\" (UID: \"86978b4a-2cb7-4f99-bec3-0afeacdeb7b6\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5h2ng" Feb 17 13:33:57 crc kubenswrapper[4955]: I0217 13:33:57.675980 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9st7\" (UniqueName: \"kubernetes.io/projected/86978b4a-2cb7-4f99-bec3-0afeacdeb7b6-kube-api-access-n9st7\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5h2ng\" (UID: \"86978b4a-2cb7-4f99-bec3-0afeacdeb7b6\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5h2ng" Feb 17 13:33:57 crc kubenswrapper[4955]: I0217 13:33:57.676037 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/86978b4a-2cb7-4f99-bec3-0afeacdeb7b6-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5h2ng\" (UID: \"86978b4a-2cb7-4f99-bec3-0afeacdeb7b6\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5h2ng" Feb 17 13:33:57 crc kubenswrapper[4955]: I0217 13:33:57.679598 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86978b4a-2cb7-4f99-bec3-0afeacdeb7b6-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5h2ng\" (UID: \"86978b4a-2cb7-4f99-bec3-0afeacdeb7b6\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5h2ng" Feb 17 13:33:57 crc kubenswrapper[4955]: I0217 13:33:57.684714 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/86978b4a-2cb7-4f99-bec3-0afeacdeb7b6-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5h2ng\" (UID: \"86978b4a-2cb7-4f99-bec3-0afeacdeb7b6\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5h2ng" Feb 17 13:33:57 crc kubenswrapper[4955]: I0217 13:33:57.695700 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9st7\" (UniqueName: \"kubernetes.io/projected/86978b4a-2cb7-4f99-bec3-0afeacdeb7b6-kube-api-access-n9st7\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-5h2ng\" (UID: \"86978b4a-2cb7-4f99-bec3-0afeacdeb7b6\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5h2ng" Feb 17 13:33:57 crc kubenswrapper[4955]: I0217 13:33:57.720578 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5h2ng" Feb 17 13:33:58 crc kubenswrapper[4955]: I0217 13:33:58.257115 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-5h2ng"] Feb 17 13:33:58 crc kubenswrapper[4955]: I0217 13:33:58.322359 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5h2ng" event={"ID":"86978b4a-2cb7-4f99-bec3-0afeacdeb7b6","Type":"ContainerStarted","Data":"f47d64b66ce0681c18701e79f8bf042540128eb87c30d38dd9f9971b4ef06b8a"} Feb 17 13:33:59 crc kubenswrapper[4955]: I0217 13:33:59.332581 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5h2ng" event={"ID":"86978b4a-2cb7-4f99-bec3-0afeacdeb7b6","Type":"ContainerStarted","Data":"f9e68079c3e25a31913e28de53a7042199c020856f592673be49b791582749bd"} Feb 17 13:33:59 crc kubenswrapper[4955]: I0217 13:33:59.358097 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5h2ng" podStartSLOduration=1.875054716 podStartE2EDuration="2.358075573s" podCreationTimestamp="2026-02-17 13:33:57 +0000 UTC" firstStartedPulling="2026-02-17 13:33:58.266348718 +0000 UTC m=+1776.789078261" lastFinishedPulling="2026-02-17 13:33:58.749369575 +0000 UTC m=+1777.272099118" observedRunningTime="2026-02-17 13:33:59.354226842 +0000 UTC m=+1777.876956385" watchObservedRunningTime="2026-02-17 13:33:59.358075573 +0000 UTC m=+1777.880805116" Feb 17 13:34:09 crc kubenswrapper[4955]: I0217 13:34:09.223562 4955 scope.go:117] "RemoveContainer" containerID="969807d9ca63cdcc1aa7cc07aa21a249802167bf58b0b6150149361ce19b14ca" Feb 17 13:34:09 crc kubenswrapper[4955]: E0217 13:34:09.224469 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:34:21 crc kubenswrapper[4955]: I0217 13:34:21.222679 4955 scope.go:117] "RemoveContainer" containerID="969807d9ca63cdcc1aa7cc07aa21a249802167bf58b0b6150149361ce19b14ca" Feb 17 13:34:21 crc kubenswrapper[4955]: E0217 13:34:21.223571 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:34:23 crc kubenswrapper[4955]: I0217 13:34:23.039380 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ffnlw"] Feb 17 13:34:23 crc kubenswrapper[4955]: I0217 13:34:23.050178 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ffnlw"] Feb 17 13:34:24 crc kubenswrapper[4955]: I0217 13:34:24.235528 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fbdb6999-85bf-4bd2-b7f4-7b7c8566941f" path="/var/lib/kubelet/pods/fbdb6999-85bf-4bd2-b7f4-7b7c8566941f/volumes" Feb 17 13:34:32 crc kubenswrapper[4955]: I0217 13:34:32.231680 4955 scope.go:117] "RemoveContainer" containerID="969807d9ca63cdcc1aa7cc07aa21a249802167bf58b0b6150149361ce19b14ca" Feb 17 13:34:32 crc kubenswrapper[4955]: E0217 13:34:32.232339 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:34:32 crc kubenswrapper[4955]: I0217 13:34:32.702563 4955 generic.go:334] "Generic (PLEG): container finished" podID="86978b4a-2cb7-4f99-bec3-0afeacdeb7b6" containerID="f9e68079c3e25a31913e28de53a7042199c020856f592673be49b791582749bd" exitCode=0 Feb 17 13:34:32 crc kubenswrapper[4955]: I0217 13:34:32.702690 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5h2ng" event={"ID":"86978b4a-2cb7-4f99-bec3-0afeacdeb7b6","Type":"ContainerDied","Data":"f9e68079c3e25a31913e28de53a7042199c020856f592673be49b791582749bd"} Feb 17 13:34:34 crc kubenswrapper[4955]: I0217 13:34:34.212930 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5h2ng" Feb 17 13:34:34 crc kubenswrapper[4955]: I0217 13:34:34.388484 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/86978b4a-2cb7-4f99-bec3-0afeacdeb7b6-ssh-key-openstack-edpm-ipam\") pod \"86978b4a-2cb7-4f99-bec3-0afeacdeb7b6\" (UID: \"86978b4a-2cb7-4f99-bec3-0afeacdeb7b6\") " Feb 17 13:34:34 crc kubenswrapper[4955]: I0217 13:34:34.388619 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86978b4a-2cb7-4f99-bec3-0afeacdeb7b6-inventory\") pod \"86978b4a-2cb7-4f99-bec3-0afeacdeb7b6\" (UID: \"86978b4a-2cb7-4f99-bec3-0afeacdeb7b6\") " Feb 17 13:34:34 crc kubenswrapper[4955]: I0217 13:34:34.388682 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9st7\" (UniqueName: \"kubernetes.io/projected/86978b4a-2cb7-4f99-bec3-0afeacdeb7b6-kube-api-access-n9st7\") pod \"86978b4a-2cb7-4f99-bec3-0afeacdeb7b6\" (UID: \"86978b4a-2cb7-4f99-bec3-0afeacdeb7b6\") " Feb 17 13:34:34 crc kubenswrapper[4955]: I0217 13:34:34.405057 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86978b4a-2cb7-4f99-bec3-0afeacdeb7b6-kube-api-access-n9st7" (OuterVolumeSpecName: "kube-api-access-n9st7") pod "86978b4a-2cb7-4f99-bec3-0afeacdeb7b6" (UID: "86978b4a-2cb7-4f99-bec3-0afeacdeb7b6"). InnerVolumeSpecName "kube-api-access-n9st7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:34:34 crc kubenswrapper[4955]: I0217 13:34:34.425915 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86978b4a-2cb7-4f99-bec3-0afeacdeb7b6-inventory" (OuterVolumeSpecName: "inventory") pod "86978b4a-2cb7-4f99-bec3-0afeacdeb7b6" (UID: "86978b4a-2cb7-4f99-bec3-0afeacdeb7b6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:34:34 crc kubenswrapper[4955]: I0217 13:34:34.434161 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86978b4a-2cb7-4f99-bec3-0afeacdeb7b6-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "86978b4a-2cb7-4f99-bec3-0afeacdeb7b6" (UID: "86978b4a-2cb7-4f99-bec3-0afeacdeb7b6"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:34:34 crc kubenswrapper[4955]: I0217 13:34:34.491437 4955 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86978b4a-2cb7-4f99-bec3-0afeacdeb7b6-inventory\") on node \"crc\" DevicePath \"\"" Feb 17 13:34:34 crc kubenswrapper[4955]: I0217 13:34:34.491488 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9st7\" (UniqueName: \"kubernetes.io/projected/86978b4a-2cb7-4f99-bec3-0afeacdeb7b6-kube-api-access-n9st7\") on node \"crc\" DevicePath \"\"" Feb 17 13:34:34 crc kubenswrapper[4955]: I0217 13:34:34.491508 4955 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/86978b4a-2cb7-4f99-bec3-0afeacdeb7b6-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 17 13:34:34 crc kubenswrapper[4955]: I0217 13:34:34.724193 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5h2ng" event={"ID":"86978b4a-2cb7-4f99-bec3-0afeacdeb7b6","Type":"ContainerDied","Data":"f47d64b66ce0681c18701e79f8bf042540128eb87c30d38dd9f9971b4ef06b8a"} Feb 17 13:34:34 crc kubenswrapper[4955]: I0217 13:34:34.724229 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f47d64b66ce0681c18701e79f8bf042540128eb87c30d38dd9f9971b4ef06b8a" Feb 17 13:34:34 crc kubenswrapper[4955]: I0217 13:34:34.724269 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-5h2ng" Feb 17 13:34:34 crc kubenswrapper[4955]: I0217 13:34:34.829403 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sj68p"] Feb 17 13:34:34 crc kubenswrapper[4955]: E0217 13:34:34.829911 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86978b4a-2cb7-4f99-bec3-0afeacdeb7b6" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 17 13:34:34 crc kubenswrapper[4955]: I0217 13:34:34.829933 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="86978b4a-2cb7-4f99-bec3-0afeacdeb7b6" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 17 13:34:34 crc kubenswrapper[4955]: I0217 13:34:34.830193 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="86978b4a-2cb7-4f99-bec3-0afeacdeb7b6" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 17 13:34:34 crc kubenswrapper[4955]: I0217 13:34:34.830801 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sj68p" Feb 17 13:34:34 crc kubenswrapper[4955]: I0217 13:34:34.832622 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 17 13:34:34 crc kubenswrapper[4955]: I0217 13:34:34.832903 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-t9lgc" Feb 17 13:34:34 crc kubenswrapper[4955]: I0217 13:34:34.833160 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 17 13:34:34 crc kubenswrapper[4955]: I0217 13:34:34.833513 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 17 13:34:34 crc kubenswrapper[4955]: I0217 13:34:34.839474 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sj68p"] Feb 17 13:34:35 crc kubenswrapper[4955]: I0217 13:34:35.001274 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5e268a8e-d845-4915-b6ff-f10bc386cfc5-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-sj68p\" (UID: \"5e268a8e-d845-4915-b6ff-f10bc386cfc5\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sj68p" Feb 17 13:34:35 crc kubenswrapper[4955]: I0217 13:34:35.001326 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cj96r\" (UniqueName: \"kubernetes.io/projected/5e268a8e-d845-4915-b6ff-f10bc386cfc5-kube-api-access-cj96r\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-sj68p\" (UID: \"5e268a8e-d845-4915-b6ff-f10bc386cfc5\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sj68p" Feb 17 13:34:35 crc kubenswrapper[4955]: I0217 13:34:35.001429 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e268a8e-d845-4915-b6ff-f10bc386cfc5-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-sj68p\" (UID: \"5e268a8e-d845-4915-b6ff-f10bc386cfc5\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sj68p" Feb 17 13:34:35 crc kubenswrapper[4955]: I0217 13:34:35.102771 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5e268a8e-d845-4915-b6ff-f10bc386cfc5-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-sj68p\" (UID: \"5e268a8e-d845-4915-b6ff-f10bc386cfc5\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sj68p" Feb 17 13:34:35 crc kubenswrapper[4955]: I0217 13:34:35.103174 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cj96r\" (UniqueName: \"kubernetes.io/projected/5e268a8e-d845-4915-b6ff-f10bc386cfc5-kube-api-access-cj96r\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-sj68p\" (UID: \"5e268a8e-d845-4915-b6ff-f10bc386cfc5\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sj68p" Feb 17 13:34:35 crc kubenswrapper[4955]: I0217 13:34:35.103269 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e268a8e-d845-4915-b6ff-f10bc386cfc5-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-sj68p\" (UID: \"5e268a8e-d845-4915-b6ff-f10bc386cfc5\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sj68p" Feb 17 13:34:35 crc kubenswrapper[4955]: I0217 13:34:35.108099 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5e268a8e-d845-4915-b6ff-f10bc386cfc5-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-sj68p\" (UID: \"5e268a8e-d845-4915-b6ff-f10bc386cfc5\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sj68p" Feb 17 13:34:35 crc kubenswrapper[4955]: I0217 13:34:35.108858 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e268a8e-d845-4915-b6ff-f10bc386cfc5-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-sj68p\" (UID: \"5e268a8e-d845-4915-b6ff-f10bc386cfc5\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sj68p" Feb 17 13:34:35 crc kubenswrapper[4955]: I0217 13:34:35.122929 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cj96r\" (UniqueName: \"kubernetes.io/projected/5e268a8e-d845-4915-b6ff-f10bc386cfc5-kube-api-access-cj96r\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-sj68p\" (UID: \"5e268a8e-d845-4915-b6ff-f10bc386cfc5\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sj68p" Feb 17 13:34:35 crc kubenswrapper[4955]: I0217 13:34:35.147918 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sj68p" Feb 17 13:34:35 crc kubenswrapper[4955]: I0217 13:34:35.660001 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sj68p"] Feb 17 13:34:35 crc kubenswrapper[4955]: I0217 13:34:35.732372 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sj68p" event={"ID":"5e268a8e-d845-4915-b6ff-f10bc386cfc5","Type":"ContainerStarted","Data":"96e8d3a5e3ec16644af5b55a71c7074eb965dfa46f4c279b98348ce30e68f0d7"} Feb 17 13:34:36 crc kubenswrapper[4955]: I0217 13:34:36.741825 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sj68p" event={"ID":"5e268a8e-d845-4915-b6ff-f10bc386cfc5","Type":"ContainerStarted","Data":"9dfb4740e9bf6e27b6e56efd206d3dd77edbf8893e8bf92ffa09e8939bbaac5c"} Feb 17 13:34:36 crc kubenswrapper[4955]: I0217 13:34:36.766925 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sj68p" podStartSLOduration=2.348304079 podStartE2EDuration="2.766890195s" podCreationTimestamp="2026-02-17 13:34:34 +0000 UTC" firstStartedPulling="2026-02-17 13:34:35.672185396 +0000 UTC m=+1814.194914939" lastFinishedPulling="2026-02-17 13:34:36.090771522 +0000 UTC m=+1814.613501055" observedRunningTime="2026-02-17 13:34:36.757620159 +0000 UTC m=+1815.280349702" watchObservedRunningTime="2026-02-17 13:34:36.766890195 +0000 UTC m=+1815.289619748" Feb 17 13:34:37 crc kubenswrapper[4955]: I0217 13:34:37.751050 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ng8xd"] Feb 17 13:34:37 crc kubenswrapper[4955]: I0217 13:34:37.754051 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ng8xd" Feb 17 13:34:37 crc kubenswrapper[4955]: I0217 13:34:37.767172 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ng8xd"] Feb 17 13:34:37 crc kubenswrapper[4955]: I0217 13:34:37.866085 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9m6p\" (UniqueName: \"kubernetes.io/projected/f261f469-cf03-48b8-8395-aea597e0274b-kube-api-access-n9m6p\") pod \"redhat-operators-ng8xd\" (UID: \"f261f469-cf03-48b8-8395-aea597e0274b\") " pod="openshift-marketplace/redhat-operators-ng8xd" Feb 17 13:34:37 crc kubenswrapper[4955]: I0217 13:34:37.866213 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f261f469-cf03-48b8-8395-aea597e0274b-utilities\") pod \"redhat-operators-ng8xd\" (UID: \"f261f469-cf03-48b8-8395-aea597e0274b\") " pod="openshift-marketplace/redhat-operators-ng8xd" Feb 17 13:34:37 crc kubenswrapper[4955]: I0217 13:34:37.866242 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f261f469-cf03-48b8-8395-aea597e0274b-catalog-content\") pod \"redhat-operators-ng8xd\" (UID: \"f261f469-cf03-48b8-8395-aea597e0274b\") " pod="openshift-marketplace/redhat-operators-ng8xd" Feb 17 13:34:37 crc kubenswrapper[4955]: I0217 13:34:37.968352 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9m6p\" (UniqueName: \"kubernetes.io/projected/f261f469-cf03-48b8-8395-aea597e0274b-kube-api-access-n9m6p\") pod \"redhat-operators-ng8xd\" (UID: \"f261f469-cf03-48b8-8395-aea597e0274b\") " pod="openshift-marketplace/redhat-operators-ng8xd" Feb 17 13:34:37 crc kubenswrapper[4955]: I0217 13:34:37.968467 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f261f469-cf03-48b8-8395-aea597e0274b-utilities\") pod \"redhat-operators-ng8xd\" (UID: \"f261f469-cf03-48b8-8395-aea597e0274b\") " pod="openshift-marketplace/redhat-operators-ng8xd" Feb 17 13:34:37 crc kubenswrapper[4955]: I0217 13:34:37.968495 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f261f469-cf03-48b8-8395-aea597e0274b-catalog-content\") pod \"redhat-operators-ng8xd\" (UID: \"f261f469-cf03-48b8-8395-aea597e0274b\") " pod="openshift-marketplace/redhat-operators-ng8xd" Feb 17 13:34:37 crc kubenswrapper[4955]: I0217 13:34:37.969002 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f261f469-cf03-48b8-8395-aea597e0274b-catalog-content\") pod \"redhat-operators-ng8xd\" (UID: \"f261f469-cf03-48b8-8395-aea597e0274b\") " pod="openshift-marketplace/redhat-operators-ng8xd" Feb 17 13:34:37 crc kubenswrapper[4955]: I0217 13:34:37.969153 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f261f469-cf03-48b8-8395-aea597e0274b-utilities\") pod \"redhat-operators-ng8xd\" (UID: \"f261f469-cf03-48b8-8395-aea597e0274b\") " pod="openshift-marketplace/redhat-operators-ng8xd" Feb 17 13:34:37 crc kubenswrapper[4955]: I0217 13:34:37.986094 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9m6p\" (UniqueName: \"kubernetes.io/projected/f261f469-cf03-48b8-8395-aea597e0274b-kube-api-access-n9m6p\") pod \"redhat-operators-ng8xd\" (UID: \"f261f469-cf03-48b8-8395-aea597e0274b\") " pod="openshift-marketplace/redhat-operators-ng8xd" Feb 17 13:34:38 crc kubenswrapper[4955]: I0217 13:34:38.076779 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ng8xd" Feb 17 13:34:38 crc kubenswrapper[4955]: I0217 13:34:38.520512 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ng8xd"] Feb 17 13:34:38 crc kubenswrapper[4955]: I0217 13:34:38.810461 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ng8xd" event={"ID":"f261f469-cf03-48b8-8395-aea597e0274b","Type":"ContainerStarted","Data":"dae1a0c63c871312eab5eb7277119e94571c7d0158b68d342e88cc6c04e7b3ca"} Feb 17 13:34:39 crc kubenswrapper[4955]: I0217 13:34:39.820897 4955 generic.go:334] "Generic (PLEG): container finished" podID="f261f469-cf03-48b8-8395-aea597e0274b" containerID="e65bad98b63086c63aad27bc330180a3adf72c351725a96329cf26453f3c9ffd" exitCode=0 Feb 17 13:34:39 crc kubenswrapper[4955]: I0217 13:34:39.820942 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ng8xd" event={"ID":"f261f469-cf03-48b8-8395-aea597e0274b","Type":"ContainerDied","Data":"e65bad98b63086c63aad27bc330180a3adf72c351725a96329cf26453f3c9ffd"} Feb 17 13:34:41 crc kubenswrapper[4955]: I0217 13:34:41.848262 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ng8xd" event={"ID":"f261f469-cf03-48b8-8395-aea597e0274b","Type":"ContainerStarted","Data":"d1c7b0484545156ee3d4b0908d13649d32767e2e3a2d76b9fd972da51f5f1976"} Feb 17 13:34:43 crc kubenswrapper[4955]: I0217 13:34:43.865039 4955 generic.go:334] "Generic (PLEG): container finished" podID="f261f469-cf03-48b8-8395-aea597e0274b" containerID="d1c7b0484545156ee3d4b0908d13649d32767e2e3a2d76b9fd972da51f5f1976" exitCode=0 Feb 17 13:34:43 crc kubenswrapper[4955]: I0217 13:34:43.865129 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ng8xd" event={"ID":"f261f469-cf03-48b8-8395-aea597e0274b","Type":"ContainerDied","Data":"d1c7b0484545156ee3d4b0908d13649d32767e2e3a2d76b9fd972da51f5f1976"} Feb 17 13:34:44 crc kubenswrapper[4955]: I0217 13:34:44.878211 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ng8xd" event={"ID":"f261f469-cf03-48b8-8395-aea597e0274b","Type":"ContainerStarted","Data":"76f5b941d824f774d0d3c6f998184e765d940b8875410ef7f427bab9e8524a87"} Feb 17 13:34:44 crc kubenswrapper[4955]: I0217 13:34:44.902875 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ng8xd" podStartSLOduration=3.405062789 podStartE2EDuration="7.902856377s" podCreationTimestamp="2026-02-17 13:34:37 +0000 UTC" firstStartedPulling="2026-02-17 13:34:39.822715298 +0000 UTC m=+1818.345444841" lastFinishedPulling="2026-02-17 13:34:44.320508886 +0000 UTC m=+1822.843238429" observedRunningTime="2026-02-17 13:34:44.900325984 +0000 UTC m=+1823.423055527" watchObservedRunningTime="2026-02-17 13:34:44.902856377 +0000 UTC m=+1823.425585930" Feb 17 13:34:47 crc kubenswrapper[4955]: I0217 13:34:47.223211 4955 scope.go:117] "RemoveContainer" containerID="969807d9ca63cdcc1aa7cc07aa21a249802167bf58b0b6150149361ce19b14ca" Feb 17 13:34:47 crc kubenswrapper[4955]: E0217 13:34:47.223722 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:34:48 crc kubenswrapper[4955]: I0217 13:34:48.077442 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ng8xd" Feb 17 13:34:48 crc kubenswrapper[4955]: I0217 13:34:48.077751 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ng8xd" Feb 17 13:34:49 crc kubenswrapper[4955]: I0217 13:34:49.120101 4955 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-ng8xd" podUID="f261f469-cf03-48b8-8395-aea597e0274b" containerName="registry-server" probeResult="failure" output=< Feb 17 13:34:49 crc kubenswrapper[4955]: timeout: failed to connect service ":50051" within 1s Feb 17 13:34:49 crc kubenswrapper[4955]: > Feb 17 13:34:50 crc kubenswrapper[4955]: I0217 13:34:50.333967 4955 scope.go:117] "RemoveContainer" containerID="ab7acc4aa73173917c4542bceac2bd95c370beb085d9ca33215c40e9a5c13f31" Feb 17 13:34:50 crc kubenswrapper[4955]: I0217 13:34:50.360608 4955 scope.go:117] "RemoveContainer" containerID="73d6d40c5cbbe29be944855ed25368d36864681d3acca1c2904cc68409ee2622" Feb 17 13:34:50 crc kubenswrapper[4955]: I0217 13:34:50.429273 4955 scope.go:117] "RemoveContainer" containerID="bdd7c1b366095bdefcf87a6ccecb349fc25188a4c12fc60e1cbb38b9b6833d28" Feb 17 13:34:50 crc kubenswrapper[4955]: I0217 13:34:50.460103 4955 scope.go:117] "RemoveContainer" containerID="1d4a8311652addf9faf8c3a9147990dbfb37dd838fdc0d55c76b3ff9aa5ebaef" Feb 17 13:34:50 crc kubenswrapper[4955]: I0217 13:34:50.503697 4955 scope.go:117] "RemoveContainer" containerID="6722a7996f5d1b7e07e4582a2b2872753d0693e485f080ab97224b9c557c3cc6" Feb 17 13:34:50 crc kubenswrapper[4955]: I0217 13:34:50.548731 4955 scope.go:117] "RemoveContainer" containerID="c527171461761a90d74648e6aa7ee6ea4bb27ea42c3a36ca0863784547769d13" Feb 17 13:34:50 crc kubenswrapper[4955]: I0217 13:34:50.598101 4955 scope.go:117] "RemoveContainer" containerID="6fe3e468aec84cc15d844b1d42ba7b177b31ca0841242ceea64587d233ef0056" Feb 17 13:34:58 crc kubenswrapper[4955]: I0217 13:34:58.125068 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ng8xd" Feb 17 13:34:58 crc kubenswrapper[4955]: I0217 13:34:58.175992 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ng8xd" Feb 17 13:34:58 crc kubenswrapper[4955]: I0217 13:34:58.359200 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ng8xd"] Feb 17 13:35:00 crc kubenswrapper[4955]: I0217 13:35:00.007712 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ng8xd" podUID="f261f469-cf03-48b8-8395-aea597e0274b" containerName="registry-server" containerID="cri-o://76f5b941d824f774d0d3c6f998184e765d940b8875410ef7f427bab9e8524a87" gracePeriod=2 Feb 17 13:35:00 crc kubenswrapper[4955]: I0217 13:35:00.223433 4955 scope.go:117] "RemoveContainer" containerID="969807d9ca63cdcc1aa7cc07aa21a249802167bf58b0b6150149361ce19b14ca" Feb 17 13:35:00 crc kubenswrapper[4955]: E0217 13:35:00.223731 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:35:00 crc kubenswrapper[4955]: I0217 13:35:00.483729 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ng8xd" Feb 17 13:35:00 crc kubenswrapper[4955]: I0217 13:35:00.548814 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f261f469-cf03-48b8-8395-aea597e0274b-utilities\") pod \"f261f469-cf03-48b8-8395-aea597e0274b\" (UID: \"f261f469-cf03-48b8-8395-aea597e0274b\") " Feb 17 13:35:00 crc kubenswrapper[4955]: I0217 13:35:00.548868 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f261f469-cf03-48b8-8395-aea597e0274b-catalog-content\") pod \"f261f469-cf03-48b8-8395-aea597e0274b\" (UID: \"f261f469-cf03-48b8-8395-aea597e0274b\") " Feb 17 13:35:00 crc kubenswrapper[4955]: I0217 13:35:00.548950 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9m6p\" (UniqueName: \"kubernetes.io/projected/f261f469-cf03-48b8-8395-aea597e0274b-kube-api-access-n9m6p\") pod \"f261f469-cf03-48b8-8395-aea597e0274b\" (UID: \"f261f469-cf03-48b8-8395-aea597e0274b\") " Feb 17 13:35:00 crc kubenswrapper[4955]: I0217 13:35:00.549979 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f261f469-cf03-48b8-8395-aea597e0274b-utilities" (OuterVolumeSpecName: "utilities") pod "f261f469-cf03-48b8-8395-aea597e0274b" (UID: "f261f469-cf03-48b8-8395-aea597e0274b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:35:00 crc kubenswrapper[4955]: I0217 13:35:00.555085 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f261f469-cf03-48b8-8395-aea597e0274b-kube-api-access-n9m6p" (OuterVolumeSpecName: "kube-api-access-n9m6p") pod "f261f469-cf03-48b8-8395-aea597e0274b" (UID: "f261f469-cf03-48b8-8395-aea597e0274b"). InnerVolumeSpecName "kube-api-access-n9m6p". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:35:00 crc kubenswrapper[4955]: I0217 13:35:00.651269 4955 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f261f469-cf03-48b8-8395-aea597e0274b-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 13:35:00 crc kubenswrapper[4955]: I0217 13:35:00.651314 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9m6p\" (UniqueName: \"kubernetes.io/projected/f261f469-cf03-48b8-8395-aea597e0274b-kube-api-access-n9m6p\") on node \"crc\" DevicePath \"\"" Feb 17 13:35:00 crc kubenswrapper[4955]: I0217 13:35:00.682609 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f261f469-cf03-48b8-8395-aea597e0274b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f261f469-cf03-48b8-8395-aea597e0274b" (UID: "f261f469-cf03-48b8-8395-aea597e0274b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:35:00 crc kubenswrapper[4955]: I0217 13:35:00.753133 4955 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f261f469-cf03-48b8-8395-aea597e0274b-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 13:35:01 crc kubenswrapper[4955]: I0217 13:35:01.019065 4955 generic.go:334] "Generic (PLEG): container finished" podID="f261f469-cf03-48b8-8395-aea597e0274b" containerID="76f5b941d824f774d0d3c6f998184e765d940b8875410ef7f427bab9e8524a87" exitCode=0 Feb 17 13:35:01 crc kubenswrapper[4955]: I0217 13:35:01.019130 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ng8xd" event={"ID":"f261f469-cf03-48b8-8395-aea597e0274b","Type":"ContainerDied","Data":"76f5b941d824f774d0d3c6f998184e765d940b8875410ef7f427bab9e8524a87"} Feb 17 13:35:01 crc kubenswrapper[4955]: I0217 13:35:01.019139 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ng8xd" Feb 17 13:35:01 crc kubenswrapper[4955]: I0217 13:35:01.019518 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ng8xd" event={"ID":"f261f469-cf03-48b8-8395-aea597e0274b","Type":"ContainerDied","Data":"dae1a0c63c871312eab5eb7277119e94571c7d0158b68d342e88cc6c04e7b3ca"} Feb 17 13:35:01 crc kubenswrapper[4955]: I0217 13:35:01.019550 4955 scope.go:117] "RemoveContainer" containerID="76f5b941d824f774d0d3c6f998184e765d940b8875410ef7f427bab9e8524a87" Feb 17 13:35:01 crc kubenswrapper[4955]: I0217 13:35:01.041341 4955 scope.go:117] "RemoveContainer" containerID="d1c7b0484545156ee3d4b0908d13649d32767e2e3a2d76b9fd972da51f5f1976" Feb 17 13:35:01 crc kubenswrapper[4955]: I0217 13:35:01.067896 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ng8xd"] Feb 17 13:35:01 crc kubenswrapper[4955]: I0217 13:35:01.074466 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ng8xd"] Feb 17 13:35:01 crc kubenswrapper[4955]: I0217 13:35:01.095840 4955 scope.go:117] "RemoveContainer" containerID="e65bad98b63086c63aad27bc330180a3adf72c351725a96329cf26453f3c9ffd" Feb 17 13:35:01 crc kubenswrapper[4955]: I0217 13:35:01.132112 4955 scope.go:117] "RemoveContainer" containerID="76f5b941d824f774d0d3c6f998184e765d940b8875410ef7f427bab9e8524a87" Feb 17 13:35:01 crc kubenswrapper[4955]: E0217 13:35:01.132665 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76f5b941d824f774d0d3c6f998184e765d940b8875410ef7f427bab9e8524a87\": container with ID starting with 76f5b941d824f774d0d3c6f998184e765d940b8875410ef7f427bab9e8524a87 not found: ID does not exist" containerID="76f5b941d824f774d0d3c6f998184e765d940b8875410ef7f427bab9e8524a87" Feb 17 13:35:01 crc kubenswrapper[4955]: I0217 13:35:01.132706 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76f5b941d824f774d0d3c6f998184e765d940b8875410ef7f427bab9e8524a87"} err="failed to get container status \"76f5b941d824f774d0d3c6f998184e765d940b8875410ef7f427bab9e8524a87\": rpc error: code = NotFound desc = could not find container \"76f5b941d824f774d0d3c6f998184e765d940b8875410ef7f427bab9e8524a87\": container with ID starting with 76f5b941d824f774d0d3c6f998184e765d940b8875410ef7f427bab9e8524a87 not found: ID does not exist" Feb 17 13:35:01 crc kubenswrapper[4955]: I0217 13:35:01.132731 4955 scope.go:117] "RemoveContainer" containerID="d1c7b0484545156ee3d4b0908d13649d32767e2e3a2d76b9fd972da51f5f1976" Feb 17 13:35:01 crc kubenswrapper[4955]: E0217 13:35:01.133247 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1c7b0484545156ee3d4b0908d13649d32767e2e3a2d76b9fd972da51f5f1976\": container with ID starting with d1c7b0484545156ee3d4b0908d13649d32767e2e3a2d76b9fd972da51f5f1976 not found: ID does not exist" containerID="d1c7b0484545156ee3d4b0908d13649d32767e2e3a2d76b9fd972da51f5f1976" Feb 17 13:35:01 crc kubenswrapper[4955]: I0217 13:35:01.133300 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1c7b0484545156ee3d4b0908d13649d32767e2e3a2d76b9fd972da51f5f1976"} err="failed to get container status \"d1c7b0484545156ee3d4b0908d13649d32767e2e3a2d76b9fd972da51f5f1976\": rpc error: code = NotFound desc = could not find container \"d1c7b0484545156ee3d4b0908d13649d32767e2e3a2d76b9fd972da51f5f1976\": container with ID starting with d1c7b0484545156ee3d4b0908d13649d32767e2e3a2d76b9fd972da51f5f1976 not found: ID does not exist" Feb 17 13:35:01 crc kubenswrapper[4955]: I0217 13:35:01.133336 4955 scope.go:117] "RemoveContainer" containerID="e65bad98b63086c63aad27bc330180a3adf72c351725a96329cf26453f3c9ffd" Feb 17 13:35:01 crc kubenswrapper[4955]: E0217 13:35:01.138338 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e65bad98b63086c63aad27bc330180a3adf72c351725a96329cf26453f3c9ffd\": container with ID starting with e65bad98b63086c63aad27bc330180a3adf72c351725a96329cf26453f3c9ffd not found: ID does not exist" containerID="e65bad98b63086c63aad27bc330180a3adf72c351725a96329cf26453f3c9ffd" Feb 17 13:35:01 crc kubenswrapper[4955]: I0217 13:35:01.138389 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e65bad98b63086c63aad27bc330180a3adf72c351725a96329cf26453f3c9ffd"} err="failed to get container status \"e65bad98b63086c63aad27bc330180a3adf72c351725a96329cf26453f3c9ffd\": rpc error: code = NotFound desc = could not find container \"e65bad98b63086c63aad27bc330180a3adf72c351725a96329cf26453f3c9ffd\": container with ID starting with e65bad98b63086c63aad27bc330180a3adf72c351725a96329cf26453f3c9ffd not found: ID does not exist" Feb 17 13:35:02 crc kubenswrapper[4955]: I0217 13:35:02.234912 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f261f469-cf03-48b8-8395-aea597e0274b" path="/var/lib/kubelet/pods/f261f469-cf03-48b8-8395-aea597e0274b/volumes" Feb 17 13:35:13 crc kubenswrapper[4955]: I0217 13:35:13.222746 4955 scope.go:117] "RemoveContainer" containerID="969807d9ca63cdcc1aa7cc07aa21a249802167bf58b0b6150149361ce19b14ca" Feb 17 13:35:14 crc kubenswrapper[4955]: I0217 13:35:14.132146 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" event={"ID":"5bf36887-2a2b-4193-abdb-a15beaf46300","Type":"ContainerStarted","Data":"8cd92c278f1f01b03343af64e98f7995ff57d30e66607734a82317d66752eafc"} Feb 17 13:35:15 crc kubenswrapper[4955]: I0217 13:35:15.043672 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-4kqpw"] Feb 17 13:35:15 crc kubenswrapper[4955]: I0217 13:35:15.051979 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-4kqpw"] Feb 17 13:35:16 crc kubenswrapper[4955]: I0217 13:35:16.028546 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-mzzkq"] Feb 17 13:35:16 crc kubenswrapper[4955]: I0217 13:35:16.037427 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-mzzkq"] Feb 17 13:35:16 crc kubenswrapper[4955]: I0217 13:35:16.236885 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c40c5ab-a929-4850-9a70-b98106008500" path="/var/lib/kubelet/pods/0c40c5ab-a929-4850-9a70-b98106008500/volumes" Feb 17 13:35:16 crc kubenswrapper[4955]: I0217 13:35:16.237966 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41b91660-a2ac-4f9e-8315-5de36243f161" path="/var/lib/kubelet/pods/41b91660-a2ac-4f9e-8315-5de36243f161/volumes" Feb 17 13:35:18 crc kubenswrapper[4955]: I0217 13:35:18.163439 4955 generic.go:334] "Generic (PLEG): container finished" podID="5e268a8e-d845-4915-b6ff-f10bc386cfc5" containerID="9dfb4740e9bf6e27b6e56efd206d3dd77edbf8893e8bf92ffa09e8939bbaac5c" exitCode=0 Feb 17 13:35:18 crc kubenswrapper[4955]: I0217 13:35:18.163548 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sj68p" event={"ID":"5e268a8e-d845-4915-b6ff-f10bc386cfc5","Type":"ContainerDied","Data":"9dfb4740e9bf6e27b6e56efd206d3dd77edbf8893e8bf92ffa09e8939bbaac5c"} Feb 17 13:35:19 crc kubenswrapper[4955]: I0217 13:35:19.574692 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sj68p" Feb 17 13:35:19 crc kubenswrapper[4955]: I0217 13:35:19.738017 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5e268a8e-d845-4915-b6ff-f10bc386cfc5-ssh-key-openstack-edpm-ipam\") pod \"5e268a8e-d845-4915-b6ff-f10bc386cfc5\" (UID: \"5e268a8e-d845-4915-b6ff-f10bc386cfc5\") " Feb 17 13:35:19 crc kubenswrapper[4955]: I0217 13:35:19.738115 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cj96r\" (UniqueName: \"kubernetes.io/projected/5e268a8e-d845-4915-b6ff-f10bc386cfc5-kube-api-access-cj96r\") pod \"5e268a8e-d845-4915-b6ff-f10bc386cfc5\" (UID: \"5e268a8e-d845-4915-b6ff-f10bc386cfc5\") " Feb 17 13:35:19 crc kubenswrapper[4955]: I0217 13:35:19.738175 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e268a8e-d845-4915-b6ff-f10bc386cfc5-inventory\") pod \"5e268a8e-d845-4915-b6ff-f10bc386cfc5\" (UID: \"5e268a8e-d845-4915-b6ff-f10bc386cfc5\") " Feb 17 13:35:19 crc kubenswrapper[4955]: I0217 13:35:19.746253 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e268a8e-d845-4915-b6ff-f10bc386cfc5-kube-api-access-cj96r" (OuterVolumeSpecName: "kube-api-access-cj96r") pod "5e268a8e-d845-4915-b6ff-f10bc386cfc5" (UID: "5e268a8e-d845-4915-b6ff-f10bc386cfc5"). InnerVolumeSpecName "kube-api-access-cj96r". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:35:19 crc kubenswrapper[4955]: I0217 13:35:19.767613 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e268a8e-d845-4915-b6ff-f10bc386cfc5-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "5e268a8e-d845-4915-b6ff-f10bc386cfc5" (UID: "5e268a8e-d845-4915-b6ff-f10bc386cfc5"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:35:19 crc kubenswrapper[4955]: I0217 13:35:19.767910 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e268a8e-d845-4915-b6ff-f10bc386cfc5-inventory" (OuterVolumeSpecName: "inventory") pod "5e268a8e-d845-4915-b6ff-f10bc386cfc5" (UID: "5e268a8e-d845-4915-b6ff-f10bc386cfc5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:35:19 crc kubenswrapper[4955]: I0217 13:35:19.840352 4955 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5e268a8e-d845-4915-b6ff-f10bc386cfc5-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 17 13:35:19 crc kubenswrapper[4955]: I0217 13:35:19.840419 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cj96r\" (UniqueName: \"kubernetes.io/projected/5e268a8e-d845-4915-b6ff-f10bc386cfc5-kube-api-access-cj96r\") on node \"crc\" DevicePath \"\"" Feb 17 13:35:19 crc kubenswrapper[4955]: I0217 13:35:19.840434 4955 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e268a8e-d845-4915-b6ff-f10bc386cfc5-inventory\") on node \"crc\" DevicePath \"\"" Feb 17 13:35:20 crc kubenswrapper[4955]: I0217 13:35:20.182405 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sj68p" event={"ID":"5e268a8e-d845-4915-b6ff-f10bc386cfc5","Type":"ContainerDied","Data":"96e8d3a5e3ec16644af5b55a71c7074eb965dfa46f4c279b98348ce30e68f0d7"} Feb 17 13:35:20 crc kubenswrapper[4955]: I0217 13:35:20.182444 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="96e8d3a5e3ec16644af5b55a71c7074eb965dfa46f4c279b98348ce30e68f0d7" Feb 17 13:35:20 crc kubenswrapper[4955]: I0217 13:35:20.182828 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sj68p" Feb 17 13:35:20 crc kubenswrapper[4955]: I0217 13:35:20.266219 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-t5v8s"] Feb 17 13:35:20 crc kubenswrapper[4955]: E0217 13:35:20.266932 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e268a8e-d845-4915-b6ff-f10bc386cfc5" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 17 13:35:20 crc kubenswrapper[4955]: I0217 13:35:20.266960 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e268a8e-d845-4915-b6ff-f10bc386cfc5" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 17 13:35:20 crc kubenswrapper[4955]: E0217 13:35:20.266978 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f261f469-cf03-48b8-8395-aea597e0274b" containerName="extract-utilities" Feb 17 13:35:20 crc kubenswrapper[4955]: I0217 13:35:20.266985 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="f261f469-cf03-48b8-8395-aea597e0274b" containerName="extract-utilities" Feb 17 13:35:20 crc kubenswrapper[4955]: E0217 13:35:20.266999 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f261f469-cf03-48b8-8395-aea597e0274b" containerName="extract-content" Feb 17 13:35:20 crc kubenswrapper[4955]: I0217 13:35:20.267005 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="f261f469-cf03-48b8-8395-aea597e0274b" containerName="extract-content" Feb 17 13:35:20 crc kubenswrapper[4955]: E0217 13:35:20.267016 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f261f469-cf03-48b8-8395-aea597e0274b" containerName="registry-server" Feb 17 13:35:20 crc kubenswrapper[4955]: I0217 13:35:20.267021 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="f261f469-cf03-48b8-8395-aea597e0274b" containerName="registry-server" Feb 17 13:35:20 crc kubenswrapper[4955]: I0217 13:35:20.267203 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e268a8e-d845-4915-b6ff-f10bc386cfc5" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 17 13:35:20 crc kubenswrapper[4955]: I0217 13:35:20.267232 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="f261f469-cf03-48b8-8395-aea597e0274b" containerName="registry-server" Feb 17 13:35:20 crc kubenswrapper[4955]: I0217 13:35:20.268163 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-t5v8s" Feb 17 13:35:20 crc kubenswrapper[4955]: I0217 13:35:20.272329 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 17 13:35:20 crc kubenswrapper[4955]: I0217 13:35:20.272743 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-t9lgc" Feb 17 13:35:20 crc kubenswrapper[4955]: I0217 13:35:20.277404 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-t5v8s"] Feb 17 13:35:20 crc kubenswrapper[4955]: I0217 13:35:20.279150 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 17 13:35:20 crc kubenswrapper[4955]: I0217 13:35:20.279149 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 17 13:35:20 crc kubenswrapper[4955]: I0217 13:35:20.350245 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e6728d3d-3842-41be-a973-efb3d3fe67b9-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-t5v8s\" (UID: \"e6728d3d-3842-41be-a973-efb3d3fe67b9\") " pod="openstack/ssh-known-hosts-edpm-deployment-t5v8s" Feb 17 13:35:20 crc kubenswrapper[4955]: I0217 13:35:20.350313 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e6728d3d-3842-41be-a973-efb3d3fe67b9-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-t5v8s\" (UID: \"e6728d3d-3842-41be-a973-efb3d3fe67b9\") " pod="openstack/ssh-known-hosts-edpm-deployment-t5v8s" Feb 17 13:35:20 crc kubenswrapper[4955]: I0217 13:35:20.350335 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzpcf\" (UniqueName: \"kubernetes.io/projected/e6728d3d-3842-41be-a973-efb3d3fe67b9-kube-api-access-hzpcf\") pod \"ssh-known-hosts-edpm-deployment-t5v8s\" (UID: \"e6728d3d-3842-41be-a973-efb3d3fe67b9\") " pod="openstack/ssh-known-hosts-edpm-deployment-t5v8s" Feb 17 13:35:20 crc kubenswrapper[4955]: I0217 13:35:20.451819 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e6728d3d-3842-41be-a973-efb3d3fe67b9-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-t5v8s\" (UID: \"e6728d3d-3842-41be-a973-efb3d3fe67b9\") " pod="openstack/ssh-known-hosts-edpm-deployment-t5v8s" Feb 17 13:35:20 crc kubenswrapper[4955]: I0217 13:35:20.451895 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e6728d3d-3842-41be-a973-efb3d3fe67b9-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-t5v8s\" (UID: \"e6728d3d-3842-41be-a973-efb3d3fe67b9\") " pod="openstack/ssh-known-hosts-edpm-deployment-t5v8s" Feb 17 13:35:20 crc kubenswrapper[4955]: I0217 13:35:20.451917 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzpcf\" (UniqueName: \"kubernetes.io/projected/e6728d3d-3842-41be-a973-efb3d3fe67b9-kube-api-access-hzpcf\") pod \"ssh-known-hosts-edpm-deployment-t5v8s\" (UID: \"e6728d3d-3842-41be-a973-efb3d3fe67b9\") " pod="openstack/ssh-known-hosts-edpm-deployment-t5v8s" Feb 17 13:35:20 crc kubenswrapper[4955]: I0217 13:35:20.456319 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e6728d3d-3842-41be-a973-efb3d3fe67b9-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-t5v8s\" (UID: \"e6728d3d-3842-41be-a973-efb3d3fe67b9\") " pod="openstack/ssh-known-hosts-edpm-deployment-t5v8s" Feb 17 13:35:20 crc kubenswrapper[4955]: I0217 13:35:20.456620 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e6728d3d-3842-41be-a973-efb3d3fe67b9-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-t5v8s\" (UID: \"e6728d3d-3842-41be-a973-efb3d3fe67b9\") " pod="openstack/ssh-known-hosts-edpm-deployment-t5v8s" Feb 17 13:35:20 crc kubenswrapper[4955]: I0217 13:35:20.474212 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzpcf\" (UniqueName: \"kubernetes.io/projected/e6728d3d-3842-41be-a973-efb3d3fe67b9-kube-api-access-hzpcf\") pod \"ssh-known-hosts-edpm-deployment-t5v8s\" (UID: \"e6728d3d-3842-41be-a973-efb3d3fe67b9\") " pod="openstack/ssh-known-hosts-edpm-deployment-t5v8s" Feb 17 13:35:20 crc kubenswrapper[4955]: I0217 13:35:20.609437 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-t5v8s" Feb 17 13:35:21 crc kubenswrapper[4955]: I0217 13:35:21.143300 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-t5v8s"] Feb 17 13:35:21 crc kubenswrapper[4955]: I0217 13:35:21.193931 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-t5v8s" event={"ID":"e6728d3d-3842-41be-a973-efb3d3fe67b9","Type":"ContainerStarted","Data":"b304fc41985cc5f4420e9889ba2eaac6d4d5acb6db0276622dfd26a23fee82f1"} Feb 17 13:35:22 crc kubenswrapper[4955]: I0217 13:35:22.238810 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-t5v8s" event={"ID":"e6728d3d-3842-41be-a973-efb3d3fe67b9","Type":"ContainerStarted","Data":"f5ac8689b083eaaa8b5a46d611a9aacef47e7424660427bbfb11eb3688c256cf"} Feb 17 13:35:22 crc kubenswrapper[4955]: I0217 13:35:22.272597 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-t5v8s" podStartSLOduration=1.756912672 podStartE2EDuration="2.272573987s" podCreationTimestamp="2026-02-17 13:35:20 +0000 UTC" firstStartedPulling="2026-02-17 13:35:21.148312907 +0000 UTC m=+1859.671042460" lastFinishedPulling="2026-02-17 13:35:21.663974232 +0000 UTC m=+1860.186703775" observedRunningTime="2026-02-17 13:35:22.269352934 +0000 UTC m=+1860.792082497" watchObservedRunningTime="2026-02-17 13:35:22.272573987 +0000 UTC m=+1860.795303530" Feb 17 13:35:28 crc kubenswrapper[4955]: I0217 13:35:28.286734 4955 generic.go:334] "Generic (PLEG): container finished" podID="e6728d3d-3842-41be-a973-efb3d3fe67b9" containerID="f5ac8689b083eaaa8b5a46d611a9aacef47e7424660427bbfb11eb3688c256cf" exitCode=0 Feb 17 13:35:28 crc kubenswrapper[4955]: I0217 13:35:28.287037 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-t5v8s" event={"ID":"e6728d3d-3842-41be-a973-efb3d3fe67b9","Type":"ContainerDied","Data":"f5ac8689b083eaaa8b5a46d611a9aacef47e7424660427bbfb11eb3688c256cf"} Feb 17 13:35:29 crc kubenswrapper[4955]: I0217 13:35:29.733425 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-t5v8s" Feb 17 13:35:29 crc kubenswrapper[4955]: I0217 13:35:29.865904 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hzpcf\" (UniqueName: \"kubernetes.io/projected/e6728d3d-3842-41be-a973-efb3d3fe67b9-kube-api-access-hzpcf\") pod \"e6728d3d-3842-41be-a973-efb3d3fe67b9\" (UID: \"e6728d3d-3842-41be-a973-efb3d3fe67b9\") " Feb 17 13:35:29 crc kubenswrapper[4955]: I0217 13:35:29.866051 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e6728d3d-3842-41be-a973-efb3d3fe67b9-inventory-0\") pod \"e6728d3d-3842-41be-a973-efb3d3fe67b9\" (UID: \"e6728d3d-3842-41be-a973-efb3d3fe67b9\") " Feb 17 13:35:29 crc kubenswrapper[4955]: I0217 13:35:29.866162 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e6728d3d-3842-41be-a973-efb3d3fe67b9-ssh-key-openstack-edpm-ipam\") pod \"e6728d3d-3842-41be-a973-efb3d3fe67b9\" (UID: \"e6728d3d-3842-41be-a973-efb3d3fe67b9\") " Feb 17 13:35:29 crc kubenswrapper[4955]: I0217 13:35:29.871876 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6728d3d-3842-41be-a973-efb3d3fe67b9-kube-api-access-hzpcf" (OuterVolumeSpecName: "kube-api-access-hzpcf") pod "e6728d3d-3842-41be-a973-efb3d3fe67b9" (UID: "e6728d3d-3842-41be-a973-efb3d3fe67b9"). InnerVolumeSpecName "kube-api-access-hzpcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:35:29 crc kubenswrapper[4955]: I0217 13:35:29.894827 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6728d3d-3842-41be-a973-efb3d3fe67b9-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "e6728d3d-3842-41be-a973-efb3d3fe67b9" (UID: "e6728d3d-3842-41be-a973-efb3d3fe67b9"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:35:29 crc kubenswrapper[4955]: I0217 13:35:29.899036 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6728d3d-3842-41be-a973-efb3d3fe67b9-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "e6728d3d-3842-41be-a973-efb3d3fe67b9" (UID: "e6728d3d-3842-41be-a973-efb3d3fe67b9"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:35:29 crc kubenswrapper[4955]: I0217 13:35:29.968483 4955 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e6728d3d-3842-41be-a973-efb3d3fe67b9-inventory-0\") on node \"crc\" DevicePath \"\"" Feb 17 13:35:29 crc kubenswrapper[4955]: I0217 13:35:29.968523 4955 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e6728d3d-3842-41be-a973-efb3d3fe67b9-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 17 13:35:29 crc kubenswrapper[4955]: I0217 13:35:29.968534 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hzpcf\" (UniqueName: \"kubernetes.io/projected/e6728d3d-3842-41be-a973-efb3d3fe67b9-kube-api-access-hzpcf\") on node \"crc\" DevicePath \"\"" Feb 17 13:35:30 crc kubenswrapper[4955]: I0217 13:35:30.309881 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-t5v8s" event={"ID":"e6728d3d-3842-41be-a973-efb3d3fe67b9","Type":"ContainerDied","Data":"b304fc41985cc5f4420e9889ba2eaac6d4d5acb6db0276622dfd26a23fee82f1"} Feb 17 13:35:30 crc kubenswrapper[4955]: I0217 13:35:30.309932 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b304fc41985cc5f4420e9889ba2eaac6d4d5acb6db0276622dfd26a23fee82f1" Feb 17 13:35:30 crc kubenswrapper[4955]: I0217 13:35:30.309972 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-t5v8s" Feb 17 13:35:30 crc kubenswrapper[4955]: I0217 13:35:30.379862 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-qplrv"] Feb 17 13:35:30 crc kubenswrapper[4955]: E0217 13:35:30.380349 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6728d3d-3842-41be-a973-efb3d3fe67b9" containerName="ssh-known-hosts-edpm-deployment" Feb 17 13:35:30 crc kubenswrapper[4955]: I0217 13:35:30.380373 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6728d3d-3842-41be-a973-efb3d3fe67b9" containerName="ssh-known-hosts-edpm-deployment" Feb 17 13:35:30 crc kubenswrapper[4955]: I0217 13:35:30.380597 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6728d3d-3842-41be-a973-efb3d3fe67b9" containerName="ssh-known-hosts-edpm-deployment" Feb 17 13:35:30 crc kubenswrapper[4955]: I0217 13:35:30.381589 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qplrv" Feb 17 13:35:30 crc kubenswrapper[4955]: I0217 13:35:30.386211 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 17 13:35:30 crc kubenswrapper[4955]: I0217 13:35:30.386302 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 17 13:35:30 crc kubenswrapper[4955]: I0217 13:35:30.386700 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 17 13:35:30 crc kubenswrapper[4955]: I0217 13:35:30.387798 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-t9lgc" Feb 17 13:35:30 crc kubenswrapper[4955]: I0217 13:35:30.388414 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-qplrv"] Feb 17 13:35:30 crc kubenswrapper[4955]: I0217 13:35:30.581165 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvtzb\" (UniqueName: \"kubernetes.io/projected/15be064a-4dca-49b2-aac9-6891c1afb237-kube-api-access-qvtzb\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-qplrv\" (UID: \"15be064a-4dca-49b2-aac9-6891c1afb237\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qplrv" Feb 17 13:35:30 crc kubenswrapper[4955]: I0217 13:35:30.581266 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/15be064a-4dca-49b2-aac9-6891c1afb237-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-qplrv\" (UID: \"15be064a-4dca-49b2-aac9-6891c1afb237\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qplrv" Feb 17 13:35:30 crc kubenswrapper[4955]: I0217 13:35:30.582811 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/15be064a-4dca-49b2-aac9-6891c1afb237-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-qplrv\" (UID: \"15be064a-4dca-49b2-aac9-6891c1afb237\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qplrv" Feb 17 13:35:30 crc kubenswrapper[4955]: I0217 13:35:30.684192 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvtzb\" (UniqueName: \"kubernetes.io/projected/15be064a-4dca-49b2-aac9-6891c1afb237-kube-api-access-qvtzb\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-qplrv\" (UID: \"15be064a-4dca-49b2-aac9-6891c1afb237\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qplrv" Feb 17 13:35:30 crc kubenswrapper[4955]: I0217 13:35:30.684276 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/15be064a-4dca-49b2-aac9-6891c1afb237-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-qplrv\" (UID: \"15be064a-4dca-49b2-aac9-6891c1afb237\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qplrv" Feb 17 13:35:30 crc kubenswrapper[4955]: I0217 13:35:30.684397 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/15be064a-4dca-49b2-aac9-6891c1afb237-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-qplrv\" (UID: \"15be064a-4dca-49b2-aac9-6891c1afb237\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qplrv" Feb 17 13:35:30 crc kubenswrapper[4955]: I0217 13:35:30.689591 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/15be064a-4dca-49b2-aac9-6891c1afb237-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-qplrv\" (UID: \"15be064a-4dca-49b2-aac9-6891c1afb237\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qplrv" Feb 17 13:35:30 crc kubenswrapper[4955]: I0217 13:35:30.692289 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/15be064a-4dca-49b2-aac9-6891c1afb237-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-qplrv\" (UID: \"15be064a-4dca-49b2-aac9-6891c1afb237\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qplrv" Feb 17 13:35:30 crc kubenswrapper[4955]: I0217 13:35:30.707866 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvtzb\" (UniqueName: \"kubernetes.io/projected/15be064a-4dca-49b2-aac9-6891c1afb237-kube-api-access-qvtzb\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-qplrv\" (UID: \"15be064a-4dca-49b2-aac9-6891c1afb237\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qplrv" Feb 17 13:35:31 crc kubenswrapper[4955]: I0217 13:35:31.000279 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qplrv" Feb 17 13:35:31 crc kubenswrapper[4955]: I0217 13:35:31.605447 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-qplrv"] Feb 17 13:35:31 crc kubenswrapper[4955]: W0217 13:35:31.615010 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod15be064a_4dca_49b2_aac9_6891c1afb237.slice/crio-93bc87850356913d4889671c8cf127b1bab296d3ffc571f566be135cdbfc6f29 WatchSource:0}: Error finding container 93bc87850356913d4889671c8cf127b1bab296d3ffc571f566be135cdbfc6f29: Status 404 returned error can't find the container with id 93bc87850356913d4889671c8cf127b1bab296d3ffc571f566be135cdbfc6f29 Feb 17 13:35:32 crc kubenswrapper[4955]: I0217 13:35:32.334178 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qplrv" event={"ID":"15be064a-4dca-49b2-aac9-6891c1afb237","Type":"ContainerStarted","Data":"93bc87850356913d4889671c8cf127b1bab296d3ffc571f566be135cdbfc6f29"} Feb 17 13:35:33 crc kubenswrapper[4955]: I0217 13:35:33.343324 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qplrv" event={"ID":"15be064a-4dca-49b2-aac9-6891c1afb237","Type":"ContainerStarted","Data":"4dcc814320d50745cae0c40123f8f28b4a896b078de34e90bb5841e77d91240b"} Feb 17 13:35:33 crc kubenswrapper[4955]: I0217 13:35:33.369623 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qplrv" podStartSLOduration=2.7778834679999997 podStartE2EDuration="3.369602788s" podCreationTimestamp="2026-02-17 13:35:30 +0000 UTC" firstStartedPulling="2026-02-17 13:35:31.617929263 +0000 UTC m=+1870.140658806" lastFinishedPulling="2026-02-17 13:35:32.209648583 +0000 UTC m=+1870.732378126" observedRunningTime="2026-02-17 13:35:33.362814553 +0000 UTC m=+1871.885544096" watchObservedRunningTime="2026-02-17 13:35:33.369602788 +0000 UTC m=+1871.892332341" Feb 17 13:35:40 crc kubenswrapper[4955]: I0217 13:35:40.403888 4955 generic.go:334] "Generic (PLEG): container finished" podID="15be064a-4dca-49b2-aac9-6891c1afb237" containerID="4dcc814320d50745cae0c40123f8f28b4a896b078de34e90bb5841e77d91240b" exitCode=0 Feb 17 13:35:40 crc kubenswrapper[4955]: I0217 13:35:40.403953 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qplrv" event={"ID":"15be064a-4dca-49b2-aac9-6891c1afb237","Type":"ContainerDied","Data":"4dcc814320d50745cae0c40123f8f28b4a896b078de34e90bb5841e77d91240b"} Feb 17 13:35:41 crc kubenswrapper[4955]: I0217 13:35:41.820587 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qplrv" Feb 17 13:35:41 crc kubenswrapper[4955]: I0217 13:35:41.944764 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/15be064a-4dca-49b2-aac9-6891c1afb237-inventory\") pod \"15be064a-4dca-49b2-aac9-6891c1afb237\" (UID: \"15be064a-4dca-49b2-aac9-6891c1afb237\") " Feb 17 13:35:41 crc kubenswrapper[4955]: I0217 13:35:41.945179 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/15be064a-4dca-49b2-aac9-6891c1afb237-ssh-key-openstack-edpm-ipam\") pod \"15be064a-4dca-49b2-aac9-6891c1afb237\" (UID: \"15be064a-4dca-49b2-aac9-6891c1afb237\") " Feb 17 13:35:41 crc kubenswrapper[4955]: I0217 13:35:41.945345 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvtzb\" (UniqueName: \"kubernetes.io/projected/15be064a-4dca-49b2-aac9-6891c1afb237-kube-api-access-qvtzb\") pod \"15be064a-4dca-49b2-aac9-6891c1afb237\" (UID: \"15be064a-4dca-49b2-aac9-6891c1afb237\") " Feb 17 13:35:41 crc kubenswrapper[4955]: I0217 13:35:41.950720 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15be064a-4dca-49b2-aac9-6891c1afb237-kube-api-access-qvtzb" (OuterVolumeSpecName: "kube-api-access-qvtzb") pod "15be064a-4dca-49b2-aac9-6891c1afb237" (UID: "15be064a-4dca-49b2-aac9-6891c1afb237"). InnerVolumeSpecName "kube-api-access-qvtzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:35:41 crc kubenswrapper[4955]: I0217 13:35:41.990865 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15be064a-4dca-49b2-aac9-6891c1afb237-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "15be064a-4dca-49b2-aac9-6891c1afb237" (UID: "15be064a-4dca-49b2-aac9-6891c1afb237"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:35:41 crc kubenswrapper[4955]: I0217 13:35:41.990937 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15be064a-4dca-49b2-aac9-6891c1afb237-inventory" (OuterVolumeSpecName: "inventory") pod "15be064a-4dca-49b2-aac9-6891c1afb237" (UID: "15be064a-4dca-49b2-aac9-6891c1afb237"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:35:42 crc kubenswrapper[4955]: I0217 13:35:42.047577 4955 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/15be064a-4dca-49b2-aac9-6891c1afb237-inventory\") on node \"crc\" DevicePath \"\"" Feb 17 13:35:42 crc kubenswrapper[4955]: I0217 13:35:42.047610 4955 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/15be064a-4dca-49b2-aac9-6891c1afb237-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 17 13:35:42 crc kubenswrapper[4955]: I0217 13:35:42.047622 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvtzb\" (UniqueName: \"kubernetes.io/projected/15be064a-4dca-49b2-aac9-6891c1afb237-kube-api-access-qvtzb\") on node \"crc\" DevicePath \"\"" Feb 17 13:35:42 crc kubenswrapper[4955]: I0217 13:35:42.422149 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qplrv" event={"ID":"15be064a-4dca-49b2-aac9-6891c1afb237","Type":"ContainerDied","Data":"93bc87850356913d4889671c8cf127b1bab296d3ffc571f566be135cdbfc6f29"} Feb 17 13:35:42 crc kubenswrapper[4955]: I0217 13:35:42.422441 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="93bc87850356913d4889671c8cf127b1bab296d3ffc571f566be135cdbfc6f29" Feb 17 13:35:42 crc kubenswrapper[4955]: I0217 13:35:42.422270 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qplrv" Feb 17 13:35:42 crc kubenswrapper[4955]: I0217 13:35:42.518736 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mqnxx"] Feb 17 13:35:42 crc kubenswrapper[4955]: E0217 13:35:42.519452 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15be064a-4dca-49b2-aac9-6891c1afb237" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 17 13:35:42 crc kubenswrapper[4955]: I0217 13:35:42.519480 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="15be064a-4dca-49b2-aac9-6891c1afb237" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 17 13:35:42 crc kubenswrapper[4955]: I0217 13:35:42.520009 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="15be064a-4dca-49b2-aac9-6891c1afb237" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 17 13:35:42 crc kubenswrapper[4955]: I0217 13:35:42.520991 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mqnxx" Feb 17 13:35:42 crc kubenswrapper[4955]: I0217 13:35:42.529220 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 17 13:35:42 crc kubenswrapper[4955]: I0217 13:35:42.529568 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 17 13:35:42 crc kubenswrapper[4955]: I0217 13:35:42.529838 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-t9lgc" Feb 17 13:35:42 crc kubenswrapper[4955]: I0217 13:35:42.530046 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 17 13:35:42 crc kubenswrapper[4955]: I0217 13:35:42.536767 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mqnxx"] Feb 17 13:35:42 crc kubenswrapper[4955]: I0217 13:35:42.567191 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfsp8\" (UniqueName: \"kubernetes.io/projected/5fe6a78c-e6a0-441f-b374-ca6603778eab-kube-api-access-dfsp8\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mqnxx\" (UID: \"5fe6a78c-e6a0-441f-b374-ca6603778eab\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mqnxx" Feb 17 13:35:42 crc kubenswrapper[4955]: I0217 13:35:42.567328 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fe6a78c-e6a0-441f-b374-ca6603778eab-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mqnxx\" (UID: \"5fe6a78c-e6a0-441f-b374-ca6603778eab\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mqnxx" Feb 17 13:35:42 crc kubenswrapper[4955]: I0217 13:35:42.567372 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5fe6a78c-e6a0-441f-b374-ca6603778eab-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mqnxx\" (UID: \"5fe6a78c-e6a0-441f-b374-ca6603778eab\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mqnxx" Feb 17 13:35:42 crc kubenswrapper[4955]: I0217 13:35:42.669216 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fe6a78c-e6a0-441f-b374-ca6603778eab-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mqnxx\" (UID: \"5fe6a78c-e6a0-441f-b374-ca6603778eab\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mqnxx" Feb 17 13:35:42 crc kubenswrapper[4955]: I0217 13:35:42.669334 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5fe6a78c-e6a0-441f-b374-ca6603778eab-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mqnxx\" (UID: \"5fe6a78c-e6a0-441f-b374-ca6603778eab\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mqnxx" Feb 17 13:35:42 crc kubenswrapper[4955]: I0217 13:35:42.669412 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfsp8\" (UniqueName: \"kubernetes.io/projected/5fe6a78c-e6a0-441f-b374-ca6603778eab-kube-api-access-dfsp8\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mqnxx\" (UID: \"5fe6a78c-e6a0-441f-b374-ca6603778eab\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mqnxx" Feb 17 13:35:42 crc kubenswrapper[4955]: I0217 13:35:42.674454 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5fe6a78c-e6a0-441f-b374-ca6603778eab-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mqnxx\" (UID: \"5fe6a78c-e6a0-441f-b374-ca6603778eab\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mqnxx" Feb 17 13:35:42 crc kubenswrapper[4955]: I0217 13:35:42.675155 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fe6a78c-e6a0-441f-b374-ca6603778eab-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mqnxx\" (UID: \"5fe6a78c-e6a0-441f-b374-ca6603778eab\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mqnxx" Feb 17 13:35:42 crc kubenswrapper[4955]: I0217 13:35:42.685683 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfsp8\" (UniqueName: \"kubernetes.io/projected/5fe6a78c-e6a0-441f-b374-ca6603778eab-kube-api-access-dfsp8\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-mqnxx\" (UID: \"5fe6a78c-e6a0-441f-b374-ca6603778eab\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mqnxx" Feb 17 13:35:42 crc kubenswrapper[4955]: I0217 13:35:42.845710 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mqnxx" Feb 17 13:35:43 crc kubenswrapper[4955]: I0217 13:35:43.349056 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mqnxx"] Feb 17 13:35:43 crc kubenswrapper[4955]: I0217 13:35:43.433275 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mqnxx" event={"ID":"5fe6a78c-e6a0-441f-b374-ca6603778eab","Type":"ContainerStarted","Data":"0bf75ff44b8a0e0a0cd22882f44835767ee015dc9a2c05ff8a5acb1a2c010f31"} Feb 17 13:35:44 crc kubenswrapper[4955]: I0217 13:35:44.442090 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mqnxx" event={"ID":"5fe6a78c-e6a0-441f-b374-ca6603778eab","Type":"ContainerStarted","Data":"726dd53f07b8f49f8223a5ed2c00d4eddefe4c3b5663fc86ad8696d65a2a0742"} Feb 17 13:35:44 crc kubenswrapper[4955]: I0217 13:35:44.461463 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mqnxx" podStartSLOduration=1.852756364 podStartE2EDuration="2.461441779s" podCreationTimestamp="2026-02-17 13:35:42 +0000 UTC" firstStartedPulling="2026-02-17 13:35:43.363347758 +0000 UTC m=+1881.886077301" lastFinishedPulling="2026-02-17 13:35:43.972033173 +0000 UTC m=+1882.494762716" observedRunningTime="2026-02-17 13:35:44.455043126 +0000 UTC m=+1882.977772679" watchObservedRunningTime="2026-02-17 13:35:44.461441779 +0000 UTC m=+1882.984171322" Feb 17 13:35:50 crc kubenswrapper[4955]: I0217 13:35:50.773600 4955 scope.go:117] "RemoveContainer" containerID="7fbd0bf75c3a27db6ce7417c4a97396f46f96609a3d031db30a4adcab8f30fd3" Feb 17 13:35:50 crc kubenswrapper[4955]: I0217 13:35:50.820859 4955 scope.go:117] "RemoveContainer" containerID="53bb653afbe4c0a45289c67de05edbd4e34d900f6ef49832c55c7b0d6452eeeb" Feb 17 13:35:53 crc kubenswrapper[4955]: I0217 13:35:53.736381 4955 generic.go:334] "Generic (PLEG): container finished" podID="5fe6a78c-e6a0-441f-b374-ca6603778eab" containerID="726dd53f07b8f49f8223a5ed2c00d4eddefe4c3b5663fc86ad8696d65a2a0742" exitCode=0 Feb 17 13:35:53 crc kubenswrapper[4955]: I0217 13:35:53.736461 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mqnxx" event={"ID":"5fe6a78c-e6a0-441f-b374-ca6603778eab","Type":"ContainerDied","Data":"726dd53f07b8f49f8223a5ed2c00d4eddefe4c3b5663fc86ad8696d65a2a0742"} Feb 17 13:35:55 crc kubenswrapper[4955]: I0217 13:35:55.173988 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mqnxx" Feb 17 13:35:55 crc kubenswrapper[4955]: I0217 13:35:55.312422 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dfsp8\" (UniqueName: \"kubernetes.io/projected/5fe6a78c-e6a0-441f-b374-ca6603778eab-kube-api-access-dfsp8\") pod \"5fe6a78c-e6a0-441f-b374-ca6603778eab\" (UID: \"5fe6a78c-e6a0-441f-b374-ca6603778eab\") " Feb 17 13:35:55 crc kubenswrapper[4955]: I0217 13:35:55.312555 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fe6a78c-e6a0-441f-b374-ca6603778eab-inventory\") pod \"5fe6a78c-e6a0-441f-b374-ca6603778eab\" (UID: \"5fe6a78c-e6a0-441f-b374-ca6603778eab\") " Feb 17 13:35:55 crc kubenswrapper[4955]: I0217 13:35:55.312613 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5fe6a78c-e6a0-441f-b374-ca6603778eab-ssh-key-openstack-edpm-ipam\") pod \"5fe6a78c-e6a0-441f-b374-ca6603778eab\" (UID: \"5fe6a78c-e6a0-441f-b374-ca6603778eab\") " Feb 17 13:35:55 crc kubenswrapper[4955]: I0217 13:35:55.318162 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe6a78c-e6a0-441f-b374-ca6603778eab-kube-api-access-dfsp8" (OuterVolumeSpecName: "kube-api-access-dfsp8") pod "5fe6a78c-e6a0-441f-b374-ca6603778eab" (UID: "5fe6a78c-e6a0-441f-b374-ca6603778eab"). InnerVolumeSpecName "kube-api-access-dfsp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:35:55 crc kubenswrapper[4955]: I0217 13:35:55.341051 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe6a78c-e6a0-441f-b374-ca6603778eab-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "5fe6a78c-e6a0-441f-b374-ca6603778eab" (UID: "5fe6a78c-e6a0-441f-b374-ca6603778eab"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:35:55 crc kubenswrapper[4955]: I0217 13:35:55.357491 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe6a78c-e6a0-441f-b374-ca6603778eab-inventory" (OuterVolumeSpecName: "inventory") pod "5fe6a78c-e6a0-441f-b374-ca6603778eab" (UID: "5fe6a78c-e6a0-441f-b374-ca6603778eab"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:35:55 crc kubenswrapper[4955]: I0217 13:35:55.414865 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dfsp8\" (UniqueName: \"kubernetes.io/projected/5fe6a78c-e6a0-441f-b374-ca6603778eab-kube-api-access-dfsp8\") on node \"crc\" DevicePath \"\"" Feb 17 13:35:55 crc kubenswrapper[4955]: I0217 13:35:55.414923 4955 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fe6a78c-e6a0-441f-b374-ca6603778eab-inventory\") on node \"crc\" DevicePath \"\"" Feb 17 13:35:55 crc kubenswrapper[4955]: I0217 13:35:55.414943 4955 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5fe6a78c-e6a0-441f-b374-ca6603778eab-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 17 13:35:55 crc kubenswrapper[4955]: I0217 13:35:55.753146 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mqnxx" event={"ID":"5fe6a78c-e6a0-441f-b374-ca6603778eab","Type":"ContainerDied","Data":"0bf75ff44b8a0e0a0cd22882f44835767ee015dc9a2c05ff8a5acb1a2c010f31"} Feb 17 13:35:55 crc kubenswrapper[4955]: I0217 13:35:55.753191 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0bf75ff44b8a0e0a0cd22882f44835767ee015dc9a2c05ff8a5acb1a2c010f31" Feb 17 13:35:55 crc kubenswrapper[4955]: I0217 13:35:55.753248 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-mqnxx" Feb 17 13:35:55 crc kubenswrapper[4955]: I0217 13:35:55.834520 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx"] Feb 17 13:35:55 crc kubenswrapper[4955]: E0217 13:35:55.835192 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fe6a78c-e6a0-441f-b374-ca6603778eab" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 17 13:35:55 crc kubenswrapper[4955]: I0217 13:35:55.835214 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fe6a78c-e6a0-441f-b374-ca6603778eab" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 17 13:35:55 crc kubenswrapper[4955]: I0217 13:35:55.835392 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fe6a78c-e6a0-441f-b374-ca6603778eab" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 17 13:35:55 crc kubenswrapper[4955]: I0217 13:35:55.835977 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:55 crc kubenswrapper[4955]: I0217 13:35:55.839048 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Feb 17 13:35:55 crc kubenswrapper[4955]: I0217 13:35:55.839302 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 17 13:35:55 crc kubenswrapper[4955]: I0217 13:35:55.839465 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 17 13:35:55 crc kubenswrapper[4955]: I0217 13:35:55.839517 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 17 13:35:55 crc kubenswrapper[4955]: I0217 13:35:55.839312 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Feb 17 13:35:55 crc kubenswrapper[4955]: I0217 13:35:55.839708 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Feb 17 13:35:55 crc kubenswrapper[4955]: I0217 13:35:55.839769 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Feb 17 13:35:55 crc kubenswrapper[4955]: I0217 13:35:55.842534 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-t9lgc" Feb 17 13:35:55 crc kubenswrapper[4955]: I0217 13:35:55.847176 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx"] Feb 17 13:35:55 crc kubenswrapper[4955]: I0217 13:35:55.923092 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:55 crc kubenswrapper[4955]: I0217 13:35:55.923147 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:55 crc kubenswrapper[4955]: I0217 13:35:55.923183 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:55 crc kubenswrapper[4955]: I0217 13:35:55.923214 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:55 crc kubenswrapper[4955]: I0217 13:35:55.923246 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:55 crc kubenswrapper[4955]: I0217 13:35:55.923274 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:55 crc kubenswrapper[4955]: I0217 13:35:55.923291 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:55 crc kubenswrapper[4955]: I0217 13:35:55.923316 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:55 crc kubenswrapper[4955]: I0217 13:35:55.923334 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:55 crc kubenswrapper[4955]: I0217 13:35:55.923353 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:55 crc kubenswrapper[4955]: I0217 13:35:55.923380 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:55 crc kubenswrapper[4955]: I0217 13:35:55.923398 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:55 crc kubenswrapper[4955]: I0217 13:35:55.923440 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8nrx\" (UniqueName: \"kubernetes.io/projected/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-kube-api-access-p8nrx\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:55 crc kubenswrapper[4955]: I0217 13:35:55.923479 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:56 crc kubenswrapper[4955]: I0217 13:35:56.024825 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:56 crc kubenswrapper[4955]: I0217 13:35:56.024873 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:56 crc kubenswrapper[4955]: I0217 13:35:56.024911 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:56 crc kubenswrapper[4955]: I0217 13:35:56.024948 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:56 crc kubenswrapper[4955]: I0217 13:35:56.024984 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:56 crc kubenswrapper[4955]: I0217 13:35:56.025013 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:56 crc kubenswrapper[4955]: I0217 13:35:56.025030 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:56 crc kubenswrapper[4955]: I0217 13:35:56.025172 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:56 crc kubenswrapper[4955]: I0217 13:35:56.025203 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:56 crc kubenswrapper[4955]: I0217 13:35:56.025232 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:56 crc kubenswrapper[4955]: I0217 13:35:56.025272 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:56 crc kubenswrapper[4955]: I0217 13:35:56.025297 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:56 crc kubenswrapper[4955]: I0217 13:35:56.025738 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8nrx\" (UniqueName: \"kubernetes.io/projected/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-kube-api-access-p8nrx\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:56 crc kubenswrapper[4955]: I0217 13:35:56.025799 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:56 crc kubenswrapper[4955]: I0217 13:35:56.030621 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:56 crc kubenswrapper[4955]: I0217 13:35:56.030621 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:56 crc kubenswrapper[4955]: I0217 13:35:56.031040 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:56 crc kubenswrapper[4955]: I0217 13:35:56.031171 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:56 crc kubenswrapper[4955]: I0217 13:35:56.031930 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:56 crc kubenswrapper[4955]: I0217 13:35:56.031964 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:56 crc kubenswrapper[4955]: I0217 13:35:56.032256 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:56 crc kubenswrapper[4955]: I0217 13:35:56.032759 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:56 crc kubenswrapper[4955]: I0217 13:35:56.033294 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:56 crc kubenswrapper[4955]: I0217 13:35:56.033650 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:56 crc kubenswrapper[4955]: I0217 13:35:56.035828 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:56 crc kubenswrapper[4955]: I0217 13:35:56.035866 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:56 crc kubenswrapper[4955]: I0217 13:35:56.042014 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8nrx\" (UniqueName: \"kubernetes.io/projected/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-kube-api-access-p8nrx\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:56 crc kubenswrapper[4955]: I0217 13:35:56.046731 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:56 crc kubenswrapper[4955]: I0217 13:35:56.199214 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:35:56 crc kubenswrapper[4955]: I0217 13:35:56.715043 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx"] Feb 17 13:35:56 crc kubenswrapper[4955]: I0217 13:35:56.763073 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" event={"ID":"f3e71e93-00f0-46f0-b87c-df0ddde2af3d","Type":"ContainerStarted","Data":"8239efeb9dece0b990adb8bfb959e5322467c08e66ccbdc6eb5025584b45ca93"} Feb 17 13:35:58 crc kubenswrapper[4955]: I0217 13:35:58.797157 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" event={"ID":"f3e71e93-00f0-46f0-b87c-df0ddde2af3d","Type":"ContainerStarted","Data":"404456f86119cc5045b22cb747ec5e626fb842eacd6facebe58e510c9e7cd2f5"} Feb 17 13:35:58 crc kubenswrapper[4955]: I0217 13:35:58.822771 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" podStartSLOduration=2.962860715 podStartE2EDuration="3.822749709s" podCreationTimestamp="2026-02-17 13:35:55 +0000 UTC" firstStartedPulling="2026-02-17 13:35:56.716849535 +0000 UTC m=+1895.239579078" lastFinishedPulling="2026-02-17 13:35:57.576738539 +0000 UTC m=+1896.099468072" observedRunningTime="2026-02-17 13:35:58.819267899 +0000 UTC m=+1897.341997442" watchObservedRunningTime="2026-02-17 13:35:58.822749709 +0000 UTC m=+1897.345479252" Feb 17 13:35:59 crc kubenswrapper[4955]: I0217 13:35:59.043532 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-4g28c"] Feb 17 13:35:59 crc kubenswrapper[4955]: I0217 13:35:59.051444 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-4g28c"] Feb 17 13:36:00 crc kubenswrapper[4955]: I0217 13:36:00.234005 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3a87515-64e3-4ed0-bc43-39ea2caa11c0" path="/var/lib/kubelet/pods/f3a87515-64e3-4ed0-bc43-39ea2caa11c0/volumes" Feb 17 13:36:30 crc kubenswrapper[4955]: I0217 13:36:30.048969 4955 generic.go:334] "Generic (PLEG): container finished" podID="f3e71e93-00f0-46f0-b87c-df0ddde2af3d" containerID="404456f86119cc5045b22cb747ec5e626fb842eacd6facebe58e510c9e7cd2f5" exitCode=0 Feb 17 13:36:30 crc kubenswrapper[4955]: I0217 13:36:30.049326 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" event={"ID":"f3e71e93-00f0-46f0-b87c-df0ddde2af3d","Type":"ContainerDied","Data":"404456f86119cc5045b22cb747ec5e626fb842eacd6facebe58e510c9e7cd2f5"} Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.453426 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.558732 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-openstack-edpm-ipam-ovn-default-certs-0\") pod \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.558805 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-ssh-key-openstack-edpm-ipam\") pod \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.559705 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.559741 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8nrx\" (UniqueName: \"kubernetes.io/projected/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-kube-api-access-p8nrx\") pod \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.559775 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-repo-setup-combined-ca-bundle\") pod \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.559816 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-libvirt-combined-ca-bundle\") pod \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.559880 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.559918 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-nova-combined-ca-bundle\") pod \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.559952 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-inventory\") pod \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.560037 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.560059 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-bootstrap-combined-ca-bundle\") pod \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.560123 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-telemetry-combined-ca-bundle\") pod \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.560177 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-ovn-combined-ca-bundle\") pod \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.560199 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-neutron-metadata-combined-ca-bundle\") pod \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\" (UID: \"f3e71e93-00f0-46f0-b87c-df0ddde2af3d\") " Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.565091 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "f3e71e93-00f0-46f0-b87c-df0ddde2af3d" (UID: "f3e71e93-00f0-46f0-b87c-df0ddde2af3d"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.566577 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-kube-api-access-p8nrx" (OuterVolumeSpecName: "kube-api-access-p8nrx") pod "f3e71e93-00f0-46f0-b87c-df0ddde2af3d" (UID: "f3e71e93-00f0-46f0-b87c-df0ddde2af3d"). InnerVolumeSpecName "kube-api-access-p8nrx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.567029 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "f3e71e93-00f0-46f0-b87c-df0ddde2af3d" (UID: "f3e71e93-00f0-46f0-b87c-df0ddde2af3d"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.567932 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "f3e71e93-00f0-46f0-b87c-df0ddde2af3d" (UID: "f3e71e93-00f0-46f0-b87c-df0ddde2af3d"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.568105 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "f3e71e93-00f0-46f0-b87c-df0ddde2af3d" (UID: "f3e71e93-00f0-46f0-b87c-df0ddde2af3d"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.568426 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "f3e71e93-00f0-46f0-b87c-df0ddde2af3d" (UID: "f3e71e93-00f0-46f0-b87c-df0ddde2af3d"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.568933 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "f3e71e93-00f0-46f0-b87c-df0ddde2af3d" (UID: "f3e71e93-00f0-46f0-b87c-df0ddde2af3d"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.570934 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "f3e71e93-00f0-46f0-b87c-df0ddde2af3d" (UID: "f3e71e93-00f0-46f0-b87c-df0ddde2af3d"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.571361 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "f3e71e93-00f0-46f0-b87c-df0ddde2af3d" (UID: "f3e71e93-00f0-46f0-b87c-df0ddde2af3d"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.573048 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "f3e71e93-00f0-46f0-b87c-df0ddde2af3d" (UID: "f3e71e93-00f0-46f0-b87c-df0ddde2af3d"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.573798 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "f3e71e93-00f0-46f0-b87c-df0ddde2af3d" (UID: "f3e71e93-00f0-46f0-b87c-df0ddde2af3d"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.574751 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "f3e71e93-00f0-46f0-b87c-df0ddde2af3d" (UID: "f3e71e93-00f0-46f0-b87c-df0ddde2af3d"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.595112 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-inventory" (OuterVolumeSpecName: "inventory") pod "f3e71e93-00f0-46f0-b87c-df0ddde2af3d" (UID: "f3e71e93-00f0-46f0-b87c-df0ddde2af3d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.607575 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "f3e71e93-00f0-46f0-b87c-df0ddde2af3d" (UID: "f3e71e93-00f0-46f0-b87c-df0ddde2af3d"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.662146 4955 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.662194 4955 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.662205 4955 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.662213 4955 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.662222 4955 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.662236 4955 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.662245 4955 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.662253 4955 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.662262 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8nrx\" (UniqueName: \"kubernetes.io/projected/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-kube-api-access-p8nrx\") on node \"crc\" DevicePath \"\"" Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.662272 4955 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.662282 4955 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.662291 4955 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.662300 4955 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:36:31 crc kubenswrapper[4955]: I0217 13:36:31.662309 4955 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f3e71e93-00f0-46f0-b87c-df0ddde2af3d-inventory\") on node \"crc\" DevicePath \"\"" Feb 17 13:36:32 crc kubenswrapper[4955]: I0217 13:36:32.066138 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" event={"ID":"f3e71e93-00f0-46f0-b87c-df0ddde2af3d","Type":"ContainerDied","Data":"8239efeb9dece0b990adb8bfb959e5322467c08e66ccbdc6eb5025584b45ca93"} Feb 17 13:36:32 crc kubenswrapper[4955]: I0217 13:36:32.066186 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8239efeb9dece0b990adb8bfb959e5322467c08e66ccbdc6eb5025584b45ca93" Feb 17 13:36:32 crc kubenswrapper[4955]: I0217 13:36:32.066243 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx" Feb 17 13:36:32 crc kubenswrapper[4955]: I0217 13:36:32.204591 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-5zsq2"] Feb 17 13:36:32 crc kubenswrapper[4955]: E0217 13:36:32.205108 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3e71e93-00f0-46f0-b87c-df0ddde2af3d" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Feb 17 13:36:32 crc kubenswrapper[4955]: I0217 13:36:32.205134 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3e71e93-00f0-46f0-b87c-df0ddde2af3d" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Feb 17 13:36:32 crc kubenswrapper[4955]: I0217 13:36:32.205344 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3e71e93-00f0-46f0-b87c-df0ddde2af3d" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Feb 17 13:36:32 crc kubenswrapper[4955]: I0217 13:36:32.206041 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5zsq2" Feb 17 13:36:32 crc kubenswrapper[4955]: I0217 13:36:32.207802 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-t9lgc" Feb 17 13:36:32 crc kubenswrapper[4955]: I0217 13:36:32.208057 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 17 13:36:32 crc kubenswrapper[4955]: I0217 13:36:32.208179 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Feb 17 13:36:32 crc kubenswrapper[4955]: I0217 13:36:32.208754 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 17 13:36:32 crc kubenswrapper[4955]: I0217 13:36:32.209676 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 17 13:36:32 crc kubenswrapper[4955]: I0217 13:36:32.216919 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-5zsq2"] Feb 17 13:36:32 crc kubenswrapper[4955]: I0217 13:36:32.376413 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27e2834f-9629-4902-b1b9-e613e29164c0-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5zsq2\" (UID: \"27e2834f-9629-4902-b1b9-e613e29164c0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5zsq2" Feb 17 13:36:32 crc kubenswrapper[4955]: I0217 13:36:32.376639 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77mzs\" (UniqueName: \"kubernetes.io/projected/27e2834f-9629-4902-b1b9-e613e29164c0-kube-api-access-77mzs\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5zsq2\" (UID: \"27e2834f-9629-4902-b1b9-e613e29164c0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5zsq2" Feb 17 13:36:32 crc kubenswrapper[4955]: I0217 13:36:32.376709 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e2834f-9629-4902-b1b9-e613e29164c0-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5zsq2\" (UID: \"27e2834f-9629-4902-b1b9-e613e29164c0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5zsq2" Feb 17 13:36:32 crc kubenswrapper[4955]: I0217 13:36:32.376801 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/27e2834f-9629-4902-b1b9-e613e29164c0-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5zsq2\" (UID: \"27e2834f-9629-4902-b1b9-e613e29164c0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5zsq2" Feb 17 13:36:32 crc kubenswrapper[4955]: I0217 13:36:32.376910 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/27e2834f-9629-4902-b1b9-e613e29164c0-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5zsq2\" (UID: \"27e2834f-9629-4902-b1b9-e613e29164c0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5zsq2" Feb 17 13:36:32 crc kubenswrapper[4955]: I0217 13:36:32.479135 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77mzs\" (UniqueName: \"kubernetes.io/projected/27e2834f-9629-4902-b1b9-e613e29164c0-kube-api-access-77mzs\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5zsq2\" (UID: \"27e2834f-9629-4902-b1b9-e613e29164c0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5zsq2" Feb 17 13:36:32 crc kubenswrapper[4955]: I0217 13:36:32.479205 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e2834f-9629-4902-b1b9-e613e29164c0-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5zsq2\" (UID: \"27e2834f-9629-4902-b1b9-e613e29164c0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5zsq2" Feb 17 13:36:32 crc kubenswrapper[4955]: I0217 13:36:32.479239 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/27e2834f-9629-4902-b1b9-e613e29164c0-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5zsq2\" (UID: \"27e2834f-9629-4902-b1b9-e613e29164c0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5zsq2" Feb 17 13:36:32 crc kubenswrapper[4955]: I0217 13:36:32.479292 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/27e2834f-9629-4902-b1b9-e613e29164c0-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5zsq2\" (UID: \"27e2834f-9629-4902-b1b9-e613e29164c0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5zsq2" Feb 17 13:36:32 crc kubenswrapper[4955]: I0217 13:36:32.479421 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27e2834f-9629-4902-b1b9-e613e29164c0-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5zsq2\" (UID: \"27e2834f-9629-4902-b1b9-e613e29164c0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5zsq2" Feb 17 13:36:32 crc kubenswrapper[4955]: I0217 13:36:32.480403 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/27e2834f-9629-4902-b1b9-e613e29164c0-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5zsq2\" (UID: \"27e2834f-9629-4902-b1b9-e613e29164c0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5zsq2" Feb 17 13:36:32 crc kubenswrapper[4955]: I0217 13:36:32.483103 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e2834f-9629-4902-b1b9-e613e29164c0-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5zsq2\" (UID: \"27e2834f-9629-4902-b1b9-e613e29164c0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5zsq2" Feb 17 13:36:32 crc kubenswrapper[4955]: I0217 13:36:32.483371 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/27e2834f-9629-4902-b1b9-e613e29164c0-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5zsq2\" (UID: \"27e2834f-9629-4902-b1b9-e613e29164c0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5zsq2" Feb 17 13:36:32 crc kubenswrapper[4955]: I0217 13:36:32.488493 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27e2834f-9629-4902-b1b9-e613e29164c0-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5zsq2\" (UID: \"27e2834f-9629-4902-b1b9-e613e29164c0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5zsq2" Feb 17 13:36:32 crc kubenswrapper[4955]: I0217 13:36:32.496973 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77mzs\" (UniqueName: \"kubernetes.io/projected/27e2834f-9629-4902-b1b9-e613e29164c0-kube-api-access-77mzs\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5zsq2\" (UID: \"27e2834f-9629-4902-b1b9-e613e29164c0\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5zsq2" Feb 17 13:36:32 crc kubenswrapper[4955]: I0217 13:36:32.569391 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5zsq2" Feb 17 13:36:33 crc kubenswrapper[4955]: I0217 13:36:33.074165 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-5zsq2"] Feb 17 13:36:33 crc kubenswrapper[4955]: I0217 13:36:33.084297 4955 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 17 13:36:34 crc kubenswrapper[4955]: I0217 13:36:34.082236 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5zsq2" event={"ID":"27e2834f-9629-4902-b1b9-e613e29164c0","Type":"ContainerStarted","Data":"ffcd4f99a81261f72d36f0a5cb51a404e2967d0aa84cf0b9f74d5f1abbf8a91d"} Feb 17 13:36:34 crc kubenswrapper[4955]: I0217 13:36:34.082559 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5zsq2" event={"ID":"27e2834f-9629-4902-b1b9-e613e29164c0","Type":"ContainerStarted","Data":"9cc98e131f6a3db0980e3ca3dd02d4049b01e7ec391127ecb2a492dae1fa6de6"} Feb 17 13:36:34 crc kubenswrapper[4955]: I0217 13:36:34.099592 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5zsq2" podStartSLOduration=1.62188109 podStartE2EDuration="2.09957793s" podCreationTimestamp="2026-02-17 13:36:32 +0000 UTC" firstStartedPulling="2026-02-17 13:36:33.084042745 +0000 UTC m=+1931.606772288" lastFinishedPulling="2026-02-17 13:36:33.561739585 +0000 UTC m=+1932.084469128" observedRunningTime="2026-02-17 13:36:34.097193911 +0000 UTC m=+1932.619923454" watchObservedRunningTime="2026-02-17 13:36:34.09957793 +0000 UTC m=+1932.622307473" Feb 17 13:36:50 crc kubenswrapper[4955]: I0217 13:36:50.925517 4955 scope.go:117] "RemoveContainer" containerID="bdbc3d9379e7a7d7f836ef2ed615cfbcc7a7d7f6708259816dc806b94dd3b3cd" Feb 17 13:37:26 crc kubenswrapper[4955]: I0217 13:37:26.521090 4955 generic.go:334] "Generic (PLEG): container finished" podID="27e2834f-9629-4902-b1b9-e613e29164c0" containerID="ffcd4f99a81261f72d36f0a5cb51a404e2967d0aa84cf0b9f74d5f1abbf8a91d" exitCode=0 Feb 17 13:37:26 crc kubenswrapper[4955]: I0217 13:37:26.521283 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5zsq2" event={"ID":"27e2834f-9629-4902-b1b9-e613e29164c0","Type":"ContainerDied","Data":"ffcd4f99a81261f72d36f0a5cb51a404e2967d0aa84cf0b9f74d5f1abbf8a91d"} Feb 17 13:37:27 crc kubenswrapper[4955]: I0217 13:37:27.942587 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5zsq2" Feb 17 13:37:27 crc kubenswrapper[4955]: I0217 13:37:27.991106 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77mzs\" (UniqueName: \"kubernetes.io/projected/27e2834f-9629-4902-b1b9-e613e29164c0-kube-api-access-77mzs\") pod \"27e2834f-9629-4902-b1b9-e613e29164c0\" (UID: \"27e2834f-9629-4902-b1b9-e613e29164c0\") " Feb 17 13:37:27 crc kubenswrapper[4955]: I0217 13:37:27.991477 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/27e2834f-9629-4902-b1b9-e613e29164c0-ssh-key-openstack-edpm-ipam\") pod \"27e2834f-9629-4902-b1b9-e613e29164c0\" (UID: \"27e2834f-9629-4902-b1b9-e613e29164c0\") " Feb 17 13:37:27 crc kubenswrapper[4955]: I0217 13:37:27.991586 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e2834f-9629-4902-b1b9-e613e29164c0-ovn-combined-ca-bundle\") pod \"27e2834f-9629-4902-b1b9-e613e29164c0\" (UID: \"27e2834f-9629-4902-b1b9-e613e29164c0\") " Feb 17 13:37:27 crc kubenswrapper[4955]: I0217 13:37:27.991725 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27e2834f-9629-4902-b1b9-e613e29164c0-inventory\") pod \"27e2834f-9629-4902-b1b9-e613e29164c0\" (UID: \"27e2834f-9629-4902-b1b9-e613e29164c0\") " Feb 17 13:37:27 crc kubenswrapper[4955]: I0217 13:37:27.992305 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/27e2834f-9629-4902-b1b9-e613e29164c0-ovncontroller-config-0\") pod \"27e2834f-9629-4902-b1b9-e613e29164c0\" (UID: \"27e2834f-9629-4902-b1b9-e613e29164c0\") " Feb 17 13:37:27 crc kubenswrapper[4955]: I0217 13:37:27.998018 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27e2834f-9629-4902-b1b9-e613e29164c0-kube-api-access-77mzs" (OuterVolumeSpecName: "kube-api-access-77mzs") pod "27e2834f-9629-4902-b1b9-e613e29164c0" (UID: "27e2834f-9629-4902-b1b9-e613e29164c0"). InnerVolumeSpecName "kube-api-access-77mzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:37:28 crc kubenswrapper[4955]: I0217 13:37:28.009074 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27e2834f-9629-4902-b1b9-e613e29164c0-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "27e2834f-9629-4902-b1b9-e613e29164c0" (UID: "27e2834f-9629-4902-b1b9-e613e29164c0"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:37:28 crc kubenswrapper[4955]: I0217 13:37:28.024598 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27e2834f-9629-4902-b1b9-e613e29164c0-inventory" (OuterVolumeSpecName: "inventory") pod "27e2834f-9629-4902-b1b9-e613e29164c0" (UID: "27e2834f-9629-4902-b1b9-e613e29164c0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:37:28 crc kubenswrapper[4955]: I0217 13:37:28.030213 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27e2834f-9629-4902-b1b9-e613e29164c0-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "27e2834f-9629-4902-b1b9-e613e29164c0" (UID: "27e2834f-9629-4902-b1b9-e613e29164c0"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:37:28 crc kubenswrapper[4955]: I0217 13:37:28.036709 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27e2834f-9629-4902-b1b9-e613e29164c0-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "27e2834f-9629-4902-b1b9-e613e29164c0" (UID: "27e2834f-9629-4902-b1b9-e613e29164c0"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:37:28 crc kubenswrapper[4955]: I0217 13:37:28.095524 4955 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/27e2834f-9629-4902-b1b9-e613e29164c0-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Feb 17 13:37:28 crc kubenswrapper[4955]: I0217 13:37:28.095615 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77mzs\" (UniqueName: \"kubernetes.io/projected/27e2834f-9629-4902-b1b9-e613e29164c0-kube-api-access-77mzs\") on node \"crc\" DevicePath \"\"" Feb 17 13:37:28 crc kubenswrapper[4955]: I0217 13:37:28.095651 4955 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/27e2834f-9629-4902-b1b9-e613e29164c0-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 17 13:37:28 crc kubenswrapper[4955]: I0217 13:37:28.095753 4955 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e2834f-9629-4902-b1b9-e613e29164c0-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:37:28 crc kubenswrapper[4955]: I0217 13:37:28.095769 4955 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27e2834f-9629-4902-b1b9-e613e29164c0-inventory\") on node \"crc\" DevicePath \"\"" Feb 17 13:37:28 crc kubenswrapper[4955]: I0217 13:37:28.541868 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5zsq2" event={"ID":"27e2834f-9629-4902-b1b9-e613e29164c0","Type":"ContainerDied","Data":"9cc98e131f6a3db0980e3ca3dd02d4049b01e7ec391127ecb2a492dae1fa6de6"} Feb 17 13:37:28 crc kubenswrapper[4955]: I0217 13:37:28.541909 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9cc98e131f6a3db0980e3ca3dd02d4049b01e7ec391127ecb2a492dae1fa6de6" Feb 17 13:37:28 crc kubenswrapper[4955]: I0217 13:37:28.541932 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5zsq2" Feb 17 13:37:28 crc kubenswrapper[4955]: I0217 13:37:28.713631 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf"] Feb 17 13:37:28 crc kubenswrapper[4955]: E0217 13:37:28.714071 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27e2834f-9629-4902-b1b9-e613e29164c0" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Feb 17 13:37:28 crc kubenswrapper[4955]: I0217 13:37:28.714090 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="27e2834f-9629-4902-b1b9-e613e29164c0" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Feb 17 13:37:28 crc kubenswrapper[4955]: I0217 13:37:28.714273 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="27e2834f-9629-4902-b1b9-e613e29164c0" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Feb 17 13:37:28 crc kubenswrapper[4955]: I0217 13:37:28.714903 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf" Feb 17 13:37:28 crc kubenswrapper[4955]: I0217 13:37:28.717910 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 17 13:37:28 crc kubenswrapper[4955]: I0217 13:37:28.717939 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 17 13:37:28 crc kubenswrapper[4955]: I0217 13:37:28.718015 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Feb 17 13:37:28 crc kubenswrapper[4955]: I0217 13:37:28.718839 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Feb 17 13:37:28 crc kubenswrapper[4955]: I0217 13:37:28.719262 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 17 13:37:28 crc kubenswrapper[4955]: I0217 13:37:28.719718 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-t9lgc" Feb 17 13:37:28 crc kubenswrapper[4955]: I0217 13:37:28.730838 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf"] Feb 17 13:37:28 crc kubenswrapper[4955]: I0217 13:37:28.810944 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5ff9b191-a542-41b9-a5c4-479d29e04fc3-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf\" (UID: \"5ff9b191-a542-41b9-a5c4-479d29e04fc3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf" Feb 17 13:37:28 crc kubenswrapper[4955]: I0217 13:37:28.811064 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ff9b191-a542-41b9-a5c4-479d29e04fc3-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf\" (UID: \"5ff9b191-a542-41b9-a5c4-479d29e04fc3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf" Feb 17 13:37:28 crc kubenswrapper[4955]: I0217 13:37:28.811106 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5ff9b191-a542-41b9-a5c4-479d29e04fc3-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf\" (UID: \"5ff9b191-a542-41b9-a5c4-479d29e04fc3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf" Feb 17 13:37:28 crc kubenswrapper[4955]: I0217 13:37:28.811173 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5ff9b191-a542-41b9-a5c4-479d29e04fc3-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf\" (UID: \"5ff9b191-a542-41b9-a5c4-479d29e04fc3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf" Feb 17 13:37:28 crc kubenswrapper[4955]: I0217 13:37:28.811296 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5ff9b191-a542-41b9-a5c4-479d29e04fc3-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf\" (UID: \"5ff9b191-a542-41b9-a5c4-479d29e04fc3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf" Feb 17 13:37:28 crc kubenswrapper[4955]: I0217 13:37:28.811349 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwxkk\" (UniqueName: \"kubernetes.io/projected/5ff9b191-a542-41b9-a5c4-479d29e04fc3-kube-api-access-qwxkk\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf\" (UID: \"5ff9b191-a542-41b9-a5c4-479d29e04fc3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf" Feb 17 13:37:28 crc kubenswrapper[4955]: I0217 13:37:28.913689 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5ff9b191-a542-41b9-a5c4-479d29e04fc3-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf\" (UID: \"5ff9b191-a542-41b9-a5c4-479d29e04fc3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf" Feb 17 13:37:28 crc kubenswrapper[4955]: I0217 13:37:28.913815 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ff9b191-a542-41b9-a5c4-479d29e04fc3-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf\" (UID: \"5ff9b191-a542-41b9-a5c4-479d29e04fc3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf" Feb 17 13:37:28 crc kubenswrapper[4955]: I0217 13:37:28.913852 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5ff9b191-a542-41b9-a5c4-479d29e04fc3-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf\" (UID: \"5ff9b191-a542-41b9-a5c4-479d29e04fc3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf" Feb 17 13:37:28 crc kubenswrapper[4955]: I0217 13:37:28.913880 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5ff9b191-a542-41b9-a5c4-479d29e04fc3-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf\" (UID: \"5ff9b191-a542-41b9-a5c4-479d29e04fc3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf" Feb 17 13:37:28 crc kubenswrapper[4955]: I0217 13:37:28.913933 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5ff9b191-a542-41b9-a5c4-479d29e04fc3-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf\" (UID: \"5ff9b191-a542-41b9-a5c4-479d29e04fc3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf" Feb 17 13:37:28 crc kubenswrapper[4955]: I0217 13:37:28.913976 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwxkk\" (UniqueName: \"kubernetes.io/projected/5ff9b191-a542-41b9-a5c4-479d29e04fc3-kube-api-access-qwxkk\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf\" (UID: \"5ff9b191-a542-41b9-a5c4-479d29e04fc3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf" Feb 17 13:37:28 crc kubenswrapper[4955]: I0217 13:37:28.918349 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5ff9b191-a542-41b9-a5c4-479d29e04fc3-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf\" (UID: \"5ff9b191-a542-41b9-a5c4-479d29e04fc3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf" Feb 17 13:37:28 crc kubenswrapper[4955]: I0217 13:37:28.918387 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5ff9b191-a542-41b9-a5c4-479d29e04fc3-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf\" (UID: \"5ff9b191-a542-41b9-a5c4-479d29e04fc3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf" Feb 17 13:37:28 crc kubenswrapper[4955]: I0217 13:37:28.918545 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5ff9b191-a542-41b9-a5c4-479d29e04fc3-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf\" (UID: \"5ff9b191-a542-41b9-a5c4-479d29e04fc3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf" Feb 17 13:37:28 crc kubenswrapper[4955]: I0217 13:37:28.921843 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ff9b191-a542-41b9-a5c4-479d29e04fc3-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf\" (UID: \"5ff9b191-a542-41b9-a5c4-479d29e04fc3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf" Feb 17 13:37:28 crc kubenswrapper[4955]: I0217 13:37:28.925101 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5ff9b191-a542-41b9-a5c4-479d29e04fc3-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf\" (UID: \"5ff9b191-a542-41b9-a5c4-479d29e04fc3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf" Feb 17 13:37:28 crc kubenswrapper[4955]: I0217 13:37:28.934167 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwxkk\" (UniqueName: \"kubernetes.io/projected/5ff9b191-a542-41b9-a5c4-479d29e04fc3-kube-api-access-qwxkk\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf\" (UID: \"5ff9b191-a542-41b9-a5c4-479d29e04fc3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf" Feb 17 13:37:29 crc kubenswrapper[4955]: I0217 13:37:29.040101 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf" Feb 17 13:37:29 crc kubenswrapper[4955]: I0217 13:37:29.562997 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf"] Feb 17 13:37:30 crc kubenswrapper[4955]: I0217 13:37:30.559062 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf" event={"ID":"5ff9b191-a542-41b9-a5c4-479d29e04fc3","Type":"ContainerStarted","Data":"30ed59e52f12cb3e1bb2db750f33e059165cbe3e6f1632b6240b66cbe6dd2200"} Feb 17 13:37:31 crc kubenswrapper[4955]: I0217 13:37:31.576644 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf" event={"ID":"5ff9b191-a542-41b9-a5c4-479d29e04fc3","Type":"ContainerStarted","Data":"761c7139978484a5c3ca81bd00d9c513f2384fa58b63551aa360340467644d49"} Feb 17 13:37:31 crc kubenswrapper[4955]: I0217 13:37:31.613516 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf" podStartSLOduration=2.707584002 podStartE2EDuration="3.613490506s" podCreationTimestamp="2026-02-17 13:37:28 +0000 UTC" firstStartedPulling="2026-02-17 13:37:29.574138077 +0000 UTC m=+1988.096867620" lastFinishedPulling="2026-02-17 13:37:30.480044581 +0000 UTC m=+1989.002774124" observedRunningTime="2026-02-17 13:37:31.606057793 +0000 UTC m=+1990.128787326" watchObservedRunningTime="2026-02-17 13:37:31.613490506 +0000 UTC m=+1990.136220069" Feb 17 13:37:34 crc kubenswrapper[4955]: I0217 13:37:34.575385 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 13:37:34 crc kubenswrapper[4955]: I0217 13:37:34.575446 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 13:38:04 crc kubenswrapper[4955]: I0217 13:38:04.574812 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 13:38:04 crc kubenswrapper[4955]: I0217 13:38:04.575443 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 13:38:11 crc kubenswrapper[4955]: I0217 13:38:11.992695 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-trp7w"] Feb 17 13:38:11 crc kubenswrapper[4955]: I0217 13:38:11.995501 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-trp7w" Feb 17 13:38:12 crc kubenswrapper[4955]: I0217 13:38:12.015226 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-trp7w"] Feb 17 13:38:12 crc kubenswrapper[4955]: I0217 13:38:12.118338 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96477d4d-0277-45a9-98a7-ebde16366f74-utilities\") pod \"redhat-marketplace-trp7w\" (UID: \"96477d4d-0277-45a9-98a7-ebde16366f74\") " pod="openshift-marketplace/redhat-marketplace-trp7w" Feb 17 13:38:12 crc kubenswrapper[4955]: I0217 13:38:12.118424 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dm9kv\" (UniqueName: \"kubernetes.io/projected/96477d4d-0277-45a9-98a7-ebde16366f74-kube-api-access-dm9kv\") pod \"redhat-marketplace-trp7w\" (UID: \"96477d4d-0277-45a9-98a7-ebde16366f74\") " pod="openshift-marketplace/redhat-marketplace-trp7w" Feb 17 13:38:12 crc kubenswrapper[4955]: I0217 13:38:12.118452 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96477d4d-0277-45a9-98a7-ebde16366f74-catalog-content\") pod \"redhat-marketplace-trp7w\" (UID: \"96477d4d-0277-45a9-98a7-ebde16366f74\") " pod="openshift-marketplace/redhat-marketplace-trp7w" Feb 17 13:38:12 crc kubenswrapper[4955]: I0217 13:38:12.220771 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96477d4d-0277-45a9-98a7-ebde16366f74-catalog-content\") pod \"redhat-marketplace-trp7w\" (UID: \"96477d4d-0277-45a9-98a7-ebde16366f74\") " pod="openshift-marketplace/redhat-marketplace-trp7w" Feb 17 13:38:12 crc kubenswrapper[4955]: I0217 13:38:12.221309 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96477d4d-0277-45a9-98a7-ebde16366f74-catalog-content\") pod \"redhat-marketplace-trp7w\" (UID: \"96477d4d-0277-45a9-98a7-ebde16366f74\") " pod="openshift-marketplace/redhat-marketplace-trp7w" Feb 17 13:38:12 crc kubenswrapper[4955]: I0217 13:38:12.221342 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96477d4d-0277-45a9-98a7-ebde16366f74-utilities\") pod \"redhat-marketplace-trp7w\" (UID: \"96477d4d-0277-45a9-98a7-ebde16366f74\") " pod="openshift-marketplace/redhat-marketplace-trp7w" Feb 17 13:38:12 crc kubenswrapper[4955]: I0217 13:38:12.221413 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dm9kv\" (UniqueName: \"kubernetes.io/projected/96477d4d-0277-45a9-98a7-ebde16366f74-kube-api-access-dm9kv\") pod \"redhat-marketplace-trp7w\" (UID: \"96477d4d-0277-45a9-98a7-ebde16366f74\") " pod="openshift-marketplace/redhat-marketplace-trp7w" Feb 17 13:38:12 crc kubenswrapper[4955]: I0217 13:38:12.221765 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96477d4d-0277-45a9-98a7-ebde16366f74-utilities\") pod \"redhat-marketplace-trp7w\" (UID: \"96477d4d-0277-45a9-98a7-ebde16366f74\") " pod="openshift-marketplace/redhat-marketplace-trp7w" Feb 17 13:38:12 crc kubenswrapper[4955]: I0217 13:38:12.243193 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dm9kv\" (UniqueName: \"kubernetes.io/projected/96477d4d-0277-45a9-98a7-ebde16366f74-kube-api-access-dm9kv\") pod \"redhat-marketplace-trp7w\" (UID: \"96477d4d-0277-45a9-98a7-ebde16366f74\") " pod="openshift-marketplace/redhat-marketplace-trp7w" Feb 17 13:38:12 crc kubenswrapper[4955]: I0217 13:38:12.337909 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-trp7w" Feb 17 13:38:12 crc kubenswrapper[4955]: I0217 13:38:12.868705 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-trp7w"] Feb 17 13:38:12 crc kubenswrapper[4955]: I0217 13:38:12.947527 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-trp7w" event={"ID":"96477d4d-0277-45a9-98a7-ebde16366f74","Type":"ContainerStarted","Data":"2f081187439d6218a26fbc4d752b94053a7731605b732aafd6b05ed9e324bdec"} Feb 17 13:38:13 crc kubenswrapper[4955]: I0217 13:38:13.957480 4955 generic.go:334] "Generic (PLEG): container finished" podID="96477d4d-0277-45a9-98a7-ebde16366f74" containerID="a1bc7e7161d1a9f950e9af309e46e4ab26822591ebc1e8c8a165b2213a549d08" exitCode=0 Feb 17 13:38:13 crc kubenswrapper[4955]: I0217 13:38:13.957528 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-trp7w" event={"ID":"96477d4d-0277-45a9-98a7-ebde16366f74","Type":"ContainerDied","Data":"a1bc7e7161d1a9f950e9af309e46e4ab26822591ebc1e8c8a165b2213a549d08"} Feb 17 13:38:14 crc kubenswrapper[4955]: I0217 13:38:14.967071 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-trp7w" event={"ID":"96477d4d-0277-45a9-98a7-ebde16366f74","Type":"ContainerStarted","Data":"61249da7411904ce82eb2bafdb7545ce9a8db4e075523d5d7b8f5743a6b2e910"} Feb 17 13:38:14 crc kubenswrapper[4955]: I0217 13:38:14.968827 4955 generic.go:334] "Generic (PLEG): container finished" podID="5ff9b191-a542-41b9-a5c4-479d29e04fc3" containerID="761c7139978484a5c3ca81bd00d9c513f2384fa58b63551aa360340467644d49" exitCode=0 Feb 17 13:38:14 crc kubenswrapper[4955]: I0217 13:38:14.968991 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf" event={"ID":"5ff9b191-a542-41b9-a5c4-479d29e04fc3","Type":"ContainerDied","Data":"761c7139978484a5c3ca81bd00d9c513f2384fa58b63551aa360340467644d49"} Feb 17 13:38:15 crc kubenswrapper[4955]: I0217 13:38:15.979545 4955 generic.go:334] "Generic (PLEG): container finished" podID="96477d4d-0277-45a9-98a7-ebde16366f74" containerID="61249da7411904ce82eb2bafdb7545ce9a8db4e075523d5d7b8f5743a6b2e910" exitCode=0 Feb 17 13:38:15 crc kubenswrapper[4955]: I0217 13:38:15.979658 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-trp7w" event={"ID":"96477d4d-0277-45a9-98a7-ebde16366f74","Type":"ContainerDied","Data":"61249da7411904ce82eb2bafdb7545ce9a8db4e075523d5d7b8f5743a6b2e910"} Feb 17 13:38:16 crc kubenswrapper[4955]: I0217 13:38:16.393834 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf" Feb 17 13:38:16 crc kubenswrapper[4955]: I0217 13:38:16.499120 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ff9b191-a542-41b9-a5c4-479d29e04fc3-neutron-metadata-combined-ca-bundle\") pod \"5ff9b191-a542-41b9-a5c4-479d29e04fc3\" (UID: \"5ff9b191-a542-41b9-a5c4-479d29e04fc3\") " Feb 17 13:38:16 crc kubenswrapper[4955]: I0217 13:38:16.499184 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5ff9b191-a542-41b9-a5c4-479d29e04fc3-neutron-ovn-metadata-agent-neutron-config-0\") pod \"5ff9b191-a542-41b9-a5c4-479d29e04fc3\" (UID: \"5ff9b191-a542-41b9-a5c4-479d29e04fc3\") " Feb 17 13:38:16 crc kubenswrapper[4955]: I0217 13:38:16.499253 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwxkk\" (UniqueName: \"kubernetes.io/projected/5ff9b191-a542-41b9-a5c4-479d29e04fc3-kube-api-access-qwxkk\") pod \"5ff9b191-a542-41b9-a5c4-479d29e04fc3\" (UID: \"5ff9b191-a542-41b9-a5c4-479d29e04fc3\") " Feb 17 13:38:16 crc kubenswrapper[4955]: I0217 13:38:16.499343 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5ff9b191-a542-41b9-a5c4-479d29e04fc3-inventory\") pod \"5ff9b191-a542-41b9-a5c4-479d29e04fc3\" (UID: \"5ff9b191-a542-41b9-a5c4-479d29e04fc3\") " Feb 17 13:38:16 crc kubenswrapper[4955]: I0217 13:38:16.499386 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5ff9b191-a542-41b9-a5c4-479d29e04fc3-ssh-key-openstack-edpm-ipam\") pod \"5ff9b191-a542-41b9-a5c4-479d29e04fc3\" (UID: \"5ff9b191-a542-41b9-a5c4-479d29e04fc3\") " Feb 17 13:38:16 crc kubenswrapper[4955]: I0217 13:38:16.499452 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5ff9b191-a542-41b9-a5c4-479d29e04fc3-nova-metadata-neutron-config-0\") pod \"5ff9b191-a542-41b9-a5c4-479d29e04fc3\" (UID: \"5ff9b191-a542-41b9-a5c4-479d29e04fc3\") " Feb 17 13:38:16 crc kubenswrapper[4955]: I0217 13:38:16.506489 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ff9b191-a542-41b9-a5c4-479d29e04fc3-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "5ff9b191-a542-41b9-a5c4-479d29e04fc3" (UID: "5ff9b191-a542-41b9-a5c4-479d29e04fc3"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:38:16 crc kubenswrapper[4955]: I0217 13:38:16.507221 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ff9b191-a542-41b9-a5c4-479d29e04fc3-kube-api-access-qwxkk" (OuterVolumeSpecName: "kube-api-access-qwxkk") pod "5ff9b191-a542-41b9-a5c4-479d29e04fc3" (UID: "5ff9b191-a542-41b9-a5c4-479d29e04fc3"). InnerVolumeSpecName "kube-api-access-qwxkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:38:16 crc kubenswrapper[4955]: I0217 13:38:16.530448 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ff9b191-a542-41b9-a5c4-479d29e04fc3-inventory" (OuterVolumeSpecName: "inventory") pod "5ff9b191-a542-41b9-a5c4-479d29e04fc3" (UID: "5ff9b191-a542-41b9-a5c4-479d29e04fc3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:38:16 crc kubenswrapper[4955]: I0217 13:38:16.532079 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ff9b191-a542-41b9-a5c4-479d29e04fc3-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "5ff9b191-a542-41b9-a5c4-479d29e04fc3" (UID: "5ff9b191-a542-41b9-a5c4-479d29e04fc3"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:38:16 crc kubenswrapper[4955]: I0217 13:38:16.536891 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ff9b191-a542-41b9-a5c4-479d29e04fc3-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "5ff9b191-a542-41b9-a5c4-479d29e04fc3" (UID: "5ff9b191-a542-41b9-a5c4-479d29e04fc3"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:38:16 crc kubenswrapper[4955]: I0217 13:38:16.539612 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ff9b191-a542-41b9-a5c4-479d29e04fc3-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "5ff9b191-a542-41b9-a5c4-479d29e04fc3" (UID: "5ff9b191-a542-41b9-a5c4-479d29e04fc3"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:38:16 crc kubenswrapper[4955]: I0217 13:38:16.608060 4955 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5ff9b191-a542-41b9-a5c4-479d29e04fc3-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Feb 17 13:38:16 crc kubenswrapper[4955]: I0217 13:38:16.608096 4955 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ff9b191-a542-41b9-a5c4-479d29e04fc3-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:38:16 crc kubenswrapper[4955]: I0217 13:38:16.608109 4955 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5ff9b191-a542-41b9-a5c4-479d29e04fc3-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Feb 17 13:38:16 crc kubenswrapper[4955]: I0217 13:38:16.608121 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwxkk\" (UniqueName: \"kubernetes.io/projected/5ff9b191-a542-41b9-a5c4-479d29e04fc3-kube-api-access-qwxkk\") on node \"crc\" DevicePath \"\"" Feb 17 13:38:16 crc kubenswrapper[4955]: I0217 13:38:16.608131 4955 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5ff9b191-a542-41b9-a5c4-479d29e04fc3-inventory\") on node \"crc\" DevicePath \"\"" Feb 17 13:38:16 crc kubenswrapper[4955]: I0217 13:38:16.608141 4955 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5ff9b191-a542-41b9-a5c4-479d29e04fc3-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 17 13:38:16 crc kubenswrapper[4955]: I0217 13:38:16.991448 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-trp7w" event={"ID":"96477d4d-0277-45a9-98a7-ebde16366f74","Type":"ContainerStarted","Data":"4136abeb5487e9f8c48e46562cca209137db547fab76dfbaab110322063a83b4"} Feb 17 13:38:16 crc kubenswrapper[4955]: I0217 13:38:16.993644 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf" event={"ID":"5ff9b191-a542-41b9-a5c4-479d29e04fc3","Type":"ContainerDied","Data":"30ed59e52f12cb3e1bb2db750f33e059165cbe3e6f1632b6240b66cbe6dd2200"} Feb 17 13:38:16 crc kubenswrapper[4955]: I0217 13:38:16.993667 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="30ed59e52f12cb3e1bb2db750f33e059165cbe3e6f1632b6240b66cbe6dd2200" Feb 17 13:38:16 crc kubenswrapper[4955]: I0217 13:38:16.993710 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf" Feb 17 13:38:17 crc kubenswrapper[4955]: I0217 13:38:17.028353 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-trp7w" podStartSLOduration=3.515612713 podStartE2EDuration="6.028327376s" podCreationTimestamp="2026-02-17 13:38:11 +0000 UTC" firstStartedPulling="2026-02-17 13:38:13.959110373 +0000 UTC m=+2032.481839916" lastFinishedPulling="2026-02-17 13:38:16.471825036 +0000 UTC m=+2034.994554579" observedRunningTime="2026-02-17 13:38:17.018314179 +0000 UTC m=+2035.541043732" watchObservedRunningTime="2026-02-17 13:38:17.028327376 +0000 UTC m=+2035.551056919" Feb 17 13:38:17 crc kubenswrapper[4955]: I0217 13:38:17.104631 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr"] Feb 17 13:38:17 crc kubenswrapper[4955]: E0217 13:38:17.105339 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ff9b191-a542-41b9-a5c4-479d29e04fc3" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Feb 17 13:38:17 crc kubenswrapper[4955]: I0217 13:38:17.105437 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ff9b191-a542-41b9-a5c4-479d29e04fc3" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Feb 17 13:38:17 crc kubenswrapper[4955]: I0217 13:38:17.105771 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ff9b191-a542-41b9-a5c4-479d29e04fc3" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Feb 17 13:38:17 crc kubenswrapper[4955]: I0217 13:38:17.106633 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr" Feb 17 13:38:17 crc kubenswrapper[4955]: I0217 13:38:17.110676 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Feb 17 13:38:17 crc kubenswrapper[4955]: I0217 13:38:17.110696 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 17 13:38:17 crc kubenswrapper[4955]: I0217 13:38:17.110894 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 17 13:38:17 crc kubenswrapper[4955]: I0217 13:38:17.110974 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 17 13:38:17 crc kubenswrapper[4955]: I0217 13:38:17.111668 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-t9lgc" Feb 17 13:38:17 crc kubenswrapper[4955]: I0217 13:38:17.115741 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr"] Feb 17 13:38:17 crc kubenswrapper[4955]: I0217 13:38:17.223870 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcbhl\" (UniqueName: \"kubernetes.io/projected/e3944196-3712-4506-bde5-758adf20aa63-kube-api-access-mcbhl\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr\" (UID: \"e3944196-3712-4506-bde5-758adf20aa63\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr" Feb 17 13:38:17 crc kubenswrapper[4955]: I0217 13:38:17.224264 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/e3944196-3712-4506-bde5-758adf20aa63-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr\" (UID: \"e3944196-3712-4506-bde5-758adf20aa63\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr" Feb 17 13:38:17 crc kubenswrapper[4955]: I0217 13:38:17.224469 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e3944196-3712-4506-bde5-758adf20aa63-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr\" (UID: \"e3944196-3712-4506-bde5-758adf20aa63\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr" Feb 17 13:38:17 crc kubenswrapper[4955]: I0217 13:38:17.224532 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3944196-3712-4506-bde5-758adf20aa63-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr\" (UID: \"e3944196-3712-4506-bde5-758adf20aa63\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr" Feb 17 13:38:17 crc kubenswrapper[4955]: I0217 13:38:17.224571 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3944196-3712-4506-bde5-758adf20aa63-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr\" (UID: \"e3944196-3712-4506-bde5-758adf20aa63\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr" Feb 17 13:38:17 crc kubenswrapper[4955]: I0217 13:38:17.326703 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3944196-3712-4506-bde5-758adf20aa63-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr\" (UID: \"e3944196-3712-4506-bde5-758adf20aa63\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr" Feb 17 13:38:17 crc kubenswrapper[4955]: I0217 13:38:17.326771 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3944196-3712-4506-bde5-758adf20aa63-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr\" (UID: \"e3944196-3712-4506-bde5-758adf20aa63\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr" Feb 17 13:38:17 crc kubenswrapper[4955]: I0217 13:38:17.326993 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcbhl\" (UniqueName: \"kubernetes.io/projected/e3944196-3712-4506-bde5-758adf20aa63-kube-api-access-mcbhl\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr\" (UID: \"e3944196-3712-4506-bde5-758adf20aa63\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr" Feb 17 13:38:17 crc kubenswrapper[4955]: I0217 13:38:17.327049 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/e3944196-3712-4506-bde5-758adf20aa63-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr\" (UID: \"e3944196-3712-4506-bde5-758adf20aa63\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr" Feb 17 13:38:17 crc kubenswrapper[4955]: I0217 13:38:17.327091 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e3944196-3712-4506-bde5-758adf20aa63-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr\" (UID: \"e3944196-3712-4506-bde5-758adf20aa63\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr" Feb 17 13:38:17 crc kubenswrapper[4955]: I0217 13:38:17.333264 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/e3944196-3712-4506-bde5-758adf20aa63-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr\" (UID: \"e3944196-3712-4506-bde5-758adf20aa63\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr" Feb 17 13:38:17 crc kubenswrapper[4955]: I0217 13:38:17.333343 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3944196-3712-4506-bde5-758adf20aa63-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr\" (UID: \"e3944196-3712-4506-bde5-758adf20aa63\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr" Feb 17 13:38:17 crc kubenswrapper[4955]: I0217 13:38:17.333930 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3944196-3712-4506-bde5-758adf20aa63-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr\" (UID: \"e3944196-3712-4506-bde5-758adf20aa63\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr" Feb 17 13:38:17 crc kubenswrapper[4955]: I0217 13:38:17.339348 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e3944196-3712-4506-bde5-758adf20aa63-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr\" (UID: \"e3944196-3712-4506-bde5-758adf20aa63\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr" Feb 17 13:38:17 crc kubenswrapper[4955]: I0217 13:38:17.348525 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcbhl\" (UniqueName: \"kubernetes.io/projected/e3944196-3712-4506-bde5-758adf20aa63-kube-api-access-mcbhl\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr\" (UID: \"e3944196-3712-4506-bde5-758adf20aa63\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr" Feb 17 13:38:17 crc kubenswrapper[4955]: I0217 13:38:17.423964 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr" Feb 17 13:38:17 crc kubenswrapper[4955]: I0217 13:38:17.994139 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr"] Feb 17 13:38:18 crc kubenswrapper[4955]: I0217 13:38:18.009523 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr" event={"ID":"e3944196-3712-4506-bde5-758adf20aa63","Type":"ContainerStarted","Data":"121e2741a57c0572b32f5d7f2a51af51dbbdaa6d14016a144eb125c2b183dc06"} Feb 17 13:38:19 crc kubenswrapper[4955]: I0217 13:38:19.021237 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr" event={"ID":"e3944196-3712-4506-bde5-758adf20aa63","Type":"ContainerStarted","Data":"6aa324d1361cd57e5a6a9797c1939e6d968539dbad7f337638ba6d0a3647bdd4"} Feb 17 13:38:19 crc kubenswrapper[4955]: I0217 13:38:19.043226 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr" podStartSLOduration=1.577742922 podStartE2EDuration="2.043204741s" podCreationTimestamp="2026-02-17 13:38:17 +0000 UTC" firstStartedPulling="2026-02-17 13:38:18.000612581 +0000 UTC m=+2036.523342134" lastFinishedPulling="2026-02-17 13:38:18.46607442 +0000 UTC m=+2036.988803953" observedRunningTime="2026-02-17 13:38:19.041298676 +0000 UTC m=+2037.564028219" watchObservedRunningTime="2026-02-17 13:38:19.043204741 +0000 UTC m=+2037.565934284" Feb 17 13:38:22 crc kubenswrapper[4955]: I0217 13:38:22.338434 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-trp7w" Feb 17 13:38:22 crc kubenswrapper[4955]: I0217 13:38:22.338986 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-trp7w" Feb 17 13:38:22 crc kubenswrapper[4955]: I0217 13:38:22.411291 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-trp7w" Feb 17 13:38:23 crc kubenswrapper[4955]: I0217 13:38:23.107603 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-trp7w" Feb 17 13:38:23 crc kubenswrapper[4955]: I0217 13:38:23.168530 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-trp7w"] Feb 17 13:38:25 crc kubenswrapper[4955]: I0217 13:38:25.075164 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-trp7w" podUID="96477d4d-0277-45a9-98a7-ebde16366f74" containerName="registry-server" containerID="cri-o://4136abeb5487e9f8c48e46562cca209137db547fab76dfbaab110322063a83b4" gracePeriod=2 Feb 17 13:38:25 crc kubenswrapper[4955]: I0217 13:38:25.542628 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-trp7w" Feb 17 13:38:25 crc kubenswrapper[4955]: I0217 13:38:25.607694 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dm9kv\" (UniqueName: \"kubernetes.io/projected/96477d4d-0277-45a9-98a7-ebde16366f74-kube-api-access-dm9kv\") pod \"96477d4d-0277-45a9-98a7-ebde16366f74\" (UID: \"96477d4d-0277-45a9-98a7-ebde16366f74\") " Feb 17 13:38:25 crc kubenswrapper[4955]: I0217 13:38:25.607876 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96477d4d-0277-45a9-98a7-ebde16366f74-utilities\") pod \"96477d4d-0277-45a9-98a7-ebde16366f74\" (UID: \"96477d4d-0277-45a9-98a7-ebde16366f74\") " Feb 17 13:38:25 crc kubenswrapper[4955]: I0217 13:38:25.607922 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96477d4d-0277-45a9-98a7-ebde16366f74-catalog-content\") pod \"96477d4d-0277-45a9-98a7-ebde16366f74\" (UID: \"96477d4d-0277-45a9-98a7-ebde16366f74\") " Feb 17 13:38:25 crc kubenswrapper[4955]: I0217 13:38:25.608547 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96477d4d-0277-45a9-98a7-ebde16366f74-utilities" (OuterVolumeSpecName: "utilities") pod "96477d4d-0277-45a9-98a7-ebde16366f74" (UID: "96477d4d-0277-45a9-98a7-ebde16366f74"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:38:25 crc kubenswrapper[4955]: I0217 13:38:25.617038 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96477d4d-0277-45a9-98a7-ebde16366f74-kube-api-access-dm9kv" (OuterVolumeSpecName: "kube-api-access-dm9kv") pod "96477d4d-0277-45a9-98a7-ebde16366f74" (UID: "96477d4d-0277-45a9-98a7-ebde16366f74"). InnerVolumeSpecName "kube-api-access-dm9kv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:38:25 crc kubenswrapper[4955]: I0217 13:38:25.647070 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96477d4d-0277-45a9-98a7-ebde16366f74-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "96477d4d-0277-45a9-98a7-ebde16366f74" (UID: "96477d4d-0277-45a9-98a7-ebde16366f74"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:38:25 crc kubenswrapper[4955]: I0217 13:38:25.710855 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dm9kv\" (UniqueName: \"kubernetes.io/projected/96477d4d-0277-45a9-98a7-ebde16366f74-kube-api-access-dm9kv\") on node \"crc\" DevicePath \"\"" Feb 17 13:38:25 crc kubenswrapper[4955]: I0217 13:38:25.710904 4955 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96477d4d-0277-45a9-98a7-ebde16366f74-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 13:38:25 crc kubenswrapper[4955]: I0217 13:38:25.710918 4955 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96477d4d-0277-45a9-98a7-ebde16366f74-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 13:38:26 crc kubenswrapper[4955]: I0217 13:38:26.086406 4955 generic.go:334] "Generic (PLEG): container finished" podID="96477d4d-0277-45a9-98a7-ebde16366f74" containerID="4136abeb5487e9f8c48e46562cca209137db547fab76dfbaab110322063a83b4" exitCode=0 Feb 17 13:38:26 crc kubenswrapper[4955]: I0217 13:38:26.086453 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-trp7w" event={"ID":"96477d4d-0277-45a9-98a7-ebde16366f74","Type":"ContainerDied","Data":"4136abeb5487e9f8c48e46562cca209137db547fab76dfbaab110322063a83b4"} Feb 17 13:38:26 crc kubenswrapper[4955]: I0217 13:38:26.086476 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-trp7w" Feb 17 13:38:26 crc kubenswrapper[4955]: I0217 13:38:26.086495 4955 scope.go:117] "RemoveContainer" containerID="4136abeb5487e9f8c48e46562cca209137db547fab76dfbaab110322063a83b4" Feb 17 13:38:26 crc kubenswrapper[4955]: I0217 13:38:26.086483 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-trp7w" event={"ID":"96477d4d-0277-45a9-98a7-ebde16366f74","Type":"ContainerDied","Data":"2f081187439d6218a26fbc4d752b94053a7731605b732aafd6b05ed9e324bdec"} Feb 17 13:38:26 crc kubenswrapper[4955]: I0217 13:38:26.108010 4955 scope.go:117] "RemoveContainer" containerID="61249da7411904ce82eb2bafdb7545ce9a8db4e075523d5d7b8f5743a6b2e910" Feb 17 13:38:26 crc kubenswrapper[4955]: I0217 13:38:26.127396 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-trp7w"] Feb 17 13:38:26 crc kubenswrapper[4955]: I0217 13:38:26.138269 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-trp7w"] Feb 17 13:38:26 crc kubenswrapper[4955]: I0217 13:38:26.148763 4955 scope.go:117] "RemoveContainer" containerID="a1bc7e7161d1a9f950e9af309e46e4ab26822591ebc1e8c8a165b2213a549d08" Feb 17 13:38:26 crc kubenswrapper[4955]: I0217 13:38:26.183242 4955 scope.go:117] "RemoveContainer" containerID="4136abeb5487e9f8c48e46562cca209137db547fab76dfbaab110322063a83b4" Feb 17 13:38:26 crc kubenswrapper[4955]: E0217 13:38:26.183676 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4136abeb5487e9f8c48e46562cca209137db547fab76dfbaab110322063a83b4\": container with ID starting with 4136abeb5487e9f8c48e46562cca209137db547fab76dfbaab110322063a83b4 not found: ID does not exist" containerID="4136abeb5487e9f8c48e46562cca209137db547fab76dfbaab110322063a83b4" Feb 17 13:38:26 crc kubenswrapper[4955]: I0217 13:38:26.183736 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4136abeb5487e9f8c48e46562cca209137db547fab76dfbaab110322063a83b4"} err="failed to get container status \"4136abeb5487e9f8c48e46562cca209137db547fab76dfbaab110322063a83b4\": rpc error: code = NotFound desc = could not find container \"4136abeb5487e9f8c48e46562cca209137db547fab76dfbaab110322063a83b4\": container with ID starting with 4136abeb5487e9f8c48e46562cca209137db547fab76dfbaab110322063a83b4 not found: ID does not exist" Feb 17 13:38:26 crc kubenswrapper[4955]: I0217 13:38:26.183769 4955 scope.go:117] "RemoveContainer" containerID="61249da7411904ce82eb2bafdb7545ce9a8db4e075523d5d7b8f5743a6b2e910" Feb 17 13:38:26 crc kubenswrapper[4955]: E0217 13:38:26.184063 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61249da7411904ce82eb2bafdb7545ce9a8db4e075523d5d7b8f5743a6b2e910\": container with ID starting with 61249da7411904ce82eb2bafdb7545ce9a8db4e075523d5d7b8f5743a6b2e910 not found: ID does not exist" containerID="61249da7411904ce82eb2bafdb7545ce9a8db4e075523d5d7b8f5743a6b2e910" Feb 17 13:38:26 crc kubenswrapper[4955]: I0217 13:38:26.184096 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61249da7411904ce82eb2bafdb7545ce9a8db4e075523d5d7b8f5743a6b2e910"} err="failed to get container status \"61249da7411904ce82eb2bafdb7545ce9a8db4e075523d5d7b8f5743a6b2e910\": rpc error: code = NotFound desc = could not find container \"61249da7411904ce82eb2bafdb7545ce9a8db4e075523d5d7b8f5743a6b2e910\": container with ID starting with 61249da7411904ce82eb2bafdb7545ce9a8db4e075523d5d7b8f5743a6b2e910 not found: ID does not exist" Feb 17 13:38:26 crc kubenswrapper[4955]: I0217 13:38:26.184118 4955 scope.go:117] "RemoveContainer" containerID="a1bc7e7161d1a9f950e9af309e46e4ab26822591ebc1e8c8a165b2213a549d08" Feb 17 13:38:26 crc kubenswrapper[4955]: E0217 13:38:26.184345 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1bc7e7161d1a9f950e9af309e46e4ab26822591ebc1e8c8a165b2213a549d08\": container with ID starting with a1bc7e7161d1a9f950e9af309e46e4ab26822591ebc1e8c8a165b2213a549d08 not found: ID does not exist" containerID="a1bc7e7161d1a9f950e9af309e46e4ab26822591ebc1e8c8a165b2213a549d08" Feb 17 13:38:26 crc kubenswrapper[4955]: I0217 13:38:26.184378 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1bc7e7161d1a9f950e9af309e46e4ab26822591ebc1e8c8a165b2213a549d08"} err="failed to get container status \"a1bc7e7161d1a9f950e9af309e46e4ab26822591ebc1e8c8a165b2213a549d08\": rpc error: code = NotFound desc = could not find container \"a1bc7e7161d1a9f950e9af309e46e4ab26822591ebc1e8c8a165b2213a549d08\": container with ID starting with a1bc7e7161d1a9f950e9af309e46e4ab26822591ebc1e8c8a165b2213a549d08 not found: ID does not exist" Feb 17 13:38:26 crc kubenswrapper[4955]: I0217 13:38:26.233991 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96477d4d-0277-45a9-98a7-ebde16366f74" path="/var/lib/kubelet/pods/96477d4d-0277-45a9-98a7-ebde16366f74/volumes" Feb 17 13:38:34 crc kubenswrapper[4955]: I0217 13:38:34.574896 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 13:38:34 crc kubenswrapper[4955]: I0217 13:38:34.575488 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 13:38:34 crc kubenswrapper[4955]: I0217 13:38:34.575543 4955 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" Feb 17 13:38:34 crc kubenswrapper[4955]: I0217 13:38:34.576327 4955 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8cd92c278f1f01b03343af64e98f7995ff57d30e66607734a82317d66752eafc"} pod="openshift-machine-config-operator/machine-config-daemon-29qxq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 17 13:38:34 crc kubenswrapper[4955]: I0217 13:38:34.576418 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" containerID="cri-o://8cd92c278f1f01b03343af64e98f7995ff57d30e66607734a82317d66752eafc" gracePeriod=600 Feb 17 13:38:35 crc kubenswrapper[4955]: I0217 13:38:35.163867 4955 generic.go:334] "Generic (PLEG): container finished" podID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerID="8cd92c278f1f01b03343af64e98f7995ff57d30e66607734a82317d66752eafc" exitCode=0 Feb 17 13:38:35 crc kubenswrapper[4955]: I0217 13:38:35.163938 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" event={"ID":"5bf36887-2a2b-4193-abdb-a15beaf46300","Type":"ContainerDied","Data":"8cd92c278f1f01b03343af64e98f7995ff57d30e66607734a82317d66752eafc"} Feb 17 13:38:35 crc kubenswrapper[4955]: I0217 13:38:35.164298 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" event={"ID":"5bf36887-2a2b-4193-abdb-a15beaf46300","Type":"ContainerStarted","Data":"e81d66db8b1e6a694c97af3c245fca0ec09bc5dbf6e24a77ebe044515c97ad4e"} Feb 17 13:38:35 crc kubenswrapper[4955]: I0217 13:38:35.164326 4955 scope.go:117] "RemoveContainer" containerID="969807d9ca63cdcc1aa7cc07aa21a249802167bf58b0b6150149361ce19b14ca" Feb 17 13:38:50 crc kubenswrapper[4955]: I0217 13:38:50.207698 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wbhxt"] Feb 17 13:38:50 crc kubenswrapper[4955]: E0217 13:38:50.208569 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96477d4d-0277-45a9-98a7-ebde16366f74" containerName="registry-server" Feb 17 13:38:50 crc kubenswrapper[4955]: I0217 13:38:50.208581 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="96477d4d-0277-45a9-98a7-ebde16366f74" containerName="registry-server" Feb 17 13:38:50 crc kubenswrapper[4955]: E0217 13:38:50.208592 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96477d4d-0277-45a9-98a7-ebde16366f74" containerName="extract-content" Feb 17 13:38:50 crc kubenswrapper[4955]: I0217 13:38:50.208598 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="96477d4d-0277-45a9-98a7-ebde16366f74" containerName="extract-content" Feb 17 13:38:50 crc kubenswrapper[4955]: E0217 13:38:50.208608 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96477d4d-0277-45a9-98a7-ebde16366f74" containerName="extract-utilities" Feb 17 13:38:50 crc kubenswrapper[4955]: I0217 13:38:50.208614 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="96477d4d-0277-45a9-98a7-ebde16366f74" containerName="extract-utilities" Feb 17 13:38:50 crc kubenswrapper[4955]: I0217 13:38:50.213627 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="96477d4d-0277-45a9-98a7-ebde16366f74" containerName="registry-server" Feb 17 13:38:50 crc kubenswrapper[4955]: I0217 13:38:50.215208 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wbhxt" Feb 17 13:38:50 crc kubenswrapper[4955]: I0217 13:38:50.236526 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wbhxt"] Feb 17 13:38:50 crc kubenswrapper[4955]: I0217 13:38:50.321945 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/172f71ef-1b49-41d9-83a3-0dd423f73cff-catalog-content\") pod \"certified-operators-wbhxt\" (UID: \"172f71ef-1b49-41d9-83a3-0dd423f73cff\") " pod="openshift-marketplace/certified-operators-wbhxt" Feb 17 13:38:50 crc kubenswrapper[4955]: I0217 13:38:50.322354 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvg55\" (UniqueName: \"kubernetes.io/projected/172f71ef-1b49-41d9-83a3-0dd423f73cff-kube-api-access-tvg55\") pod \"certified-operators-wbhxt\" (UID: \"172f71ef-1b49-41d9-83a3-0dd423f73cff\") " pod="openshift-marketplace/certified-operators-wbhxt" Feb 17 13:38:50 crc kubenswrapper[4955]: I0217 13:38:50.322762 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/172f71ef-1b49-41d9-83a3-0dd423f73cff-utilities\") pod \"certified-operators-wbhxt\" (UID: \"172f71ef-1b49-41d9-83a3-0dd423f73cff\") " pod="openshift-marketplace/certified-operators-wbhxt" Feb 17 13:38:50 crc kubenswrapper[4955]: I0217 13:38:50.424495 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/172f71ef-1b49-41d9-83a3-0dd423f73cff-catalog-content\") pod \"certified-operators-wbhxt\" (UID: \"172f71ef-1b49-41d9-83a3-0dd423f73cff\") " pod="openshift-marketplace/certified-operators-wbhxt" Feb 17 13:38:50 crc kubenswrapper[4955]: I0217 13:38:50.424588 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvg55\" (UniqueName: \"kubernetes.io/projected/172f71ef-1b49-41d9-83a3-0dd423f73cff-kube-api-access-tvg55\") pod \"certified-operators-wbhxt\" (UID: \"172f71ef-1b49-41d9-83a3-0dd423f73cff\") " pod="openshift-marketplace/certified-operators-wbhxt" Feb 17 13:38:50 crc kubenswrapper[4955]: I0217 13:38:50.424697 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/172f71ef-1b49-41d9-83a3-0dd423f73cff-utilities\") pod \"certified-operators-wbhxt\" (UID: \"172f71ef-1b49-41d9-83a3-0dd423f73cff\") " pod="openshift-marketplace/certified-operators-wbhxt" Feb 17 13:38:50 crc kubenswrapper[4955]: I0217 13:38:50.425300 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/172f71ef-1b49-41d9-83a3-0dd423f73cff-utilities\") pod \"certified-operators-wbhxt\" (UID: \"172f71ef-1b49-41d9-83a3-0dd423f73cff\") " pod="openshift-marketplace/certified-operators-wbhxt" Feb 17 13:38:50 crc kubenswrapper[4955]: I0217 13:38:50.425327 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/172f71ef-1b49-41d9-83a3-0dd423f73cff-catalog-content\") pod \"certified-operators-wbhxt\" (UID: \"172f71ef-1b49-41d9-83a3-0dd423f73cff\") " pod="openshift-marketplace/certified-operators-wbhxt" Feb 17 13:38:50 crc kubenswrapper[4955]: I0217 13:38:50.452974 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvg55\" (UniqueName: \"kubernetes.io/projected/172f71ef-1b49-41d9-83a3-0dd423f73cff-kube-api-access-tvg55\") pod \"certified-operators-wbhxt\" (UID: \"172f71ef-1b49-41d9-83a3-0dd423f73cff\") " pod="openshift-marketplace/certified-operators-wbhxt" Feb 17 13:38:50 crc kubenswrapper[4955]: I0217 13:38:50.548881 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wbhxt" Feb 17 13:38:51 crc kubenswrapper[4955]: I0217 13:38:51.196463 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wbhxt"] Feb 17 13:38:51 crc kubenswrapper[4955]: I0217 13:38:51.706982 4955 generic.go:334] "Generic (PLEG): container finished" podID="172f71ef-1b49-41d9-83a3-0dd423f73cff" containerID="5ab9f3d31d048abebe969020062976bc928ebdf942876473e4d793e6515e2577" exitCode=0 Feb 17 13:38:51 crc kubenswrapper[4955]: I0217 13:38:51.707028 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wbhxt" event={"ID":"172f71ef-1b49-41d9-83a3-0dd423f73cff","Type":"ContainerDied","Data":"5ab9f3d31d048abebe969020062976bc928ebdf942876473e4d793e6515e2577"} Feb 17 13:38:51 crc kubenswrapper[4955]: I0217 13:38:51.707075 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wbhxt" event={"ID":"172f71ef-1b49-41d9-83a3-0dd423f73cff","Type":"ContainerStarted","Data":"d2e2844b6ba48204c0654ebd9fda78d723a22b08e4385a32161aac83e145da0a"} Feb 17 13:38:54 crc kubenswrapper[4955]: I0217 13:38:54.743132 4955 generic.go:334] "Generic (PLEG): container finished" podID="172f71ef-1b49-41d9-83a3-0dd423f73cff" containerID="e23f062d64b1531a53f7b9516a0a8ade9b8723342f6023cf5ada3ebd867df39b" exitCode=0 Feb 17 13:38:54 crc kubenswrapper[4955]: I0217 13:38:54.743203 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wbhxt" event={"ID":"172f71ef-1b49-41d9-83a3-0dd423f73cff","Type":"ContainerDied","Data":"e23f062d64b1531a53f7b9516a0a8ade9b8723342f6023cf5ada3ebd867df39b"} Feb 17 13:38:55 crc kubenswrapper[4955]: I0217 13:38:55.755770 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wbhxt" event={"ID":"172f71ef-1b49-41d9-83a3-0dd423f73cff","Type":"ContainerStarted","Data":"6fc5d73363fe44838fa287c14fb11e6475f821d616fd9732ebe4c3dc844cbf55"} Feb 17 13:38:55 crc kubenswrapper[4955]: I0217 13:38:55.784059 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wbhxt" podStartSLOduration=2.312271558 podStartE2EDuration="5.784034437s" podCreationTimestamp="2026-02-17 13:38:50 +0000 UTC" firstStartedPulling="2026-02-17 13:38:51.708982527 +0000 UTC m=+2070.231712070" lastFinishedPulling="2026-02-17 13:38:55.180745416 +0000 UTC m=+2073.703474949" observedRunningTime="2026-02-17 13:38:55.780854685 +0000 UTC m=+2074.303584228" watchObservedRunningTime="2026-02-17 13:38:55.784034437 +0000 UTC m=+2074.306763980" Feb 17 13:38:56 crc kubenswrapper[4955]: I0217 13:38:56.605709 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6nhcb"] Feb 17 13:38:56 crc kubenswrapper[4955]: I0217 13:38:56.608518 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6nhcb" Feb 17 13:38:56 crc kubenswrapper[4955]: I0217 13:38:56.620616 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6nhcb"] Feb 17 13:38:56 crc kubenswrapper[4955]: I0217 13:38:56.677875 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7p759\" (UniqueName: \"kubernetes.io/projected/51037160-13be-429b-991a-5b1e2ecdd246-kube-api-access-7p759\") pod \"community-operators-6nhcb\" (UID: \"51037160-13be-429b-991a-5b1e2ecdd246\") " pod="openshift-marketplace/community-operators-6nhcb" Feb 17 13:38:56 crc kubenswrapper[4955]: I0217 13:38:56.678266 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51037160-13be-429b-991a-5b1e2ecdd246-utilities\") pod \"community-operators-6nhcb\" (UID: \"51037160-13be-429b-991a-5b1e2ecdd246\") " pod="openshift-marketplace/community-operators-6nhcb" Feb 17 13:38:56 crc kubenswrapper[4955]: I0217 13:38:56.678587 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51037160-13be-429b-991a-5b1e2ecdd246-catalog-content\") pod \"community-operators-6nhcb\" (UID: \"51037160-13be-429b-991a-5b1e2ecdd246\") " pod="openshift-marketplace/community-operators-6nhcb" Feb 17 13:38:56 crc kubenswrapper[4955]: I0217 13:38:56.779763 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7p759\" (UniqueName: \"kubernetes.io/projected/51037160-13be-429b-991a-5b1e2ecdd246-kube-api-access-7p759\") pod \"community-operators-6nhcb\" (UID: \"51037160-13be-429b-991a-5b1e2ecdd246\") " pod="openshift-marketplace/community-operators-6nhcb" Feb 17 13:38:56 crc kubenswrapper[4955]: I0217 13:38:56.779842 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51037160-13be-429b-991a-5b1e2ecdd246-utilities\") pod \"community-operators-6nhcb\" (UID: \"51037160-13be-429b-991a-5b1e2ecdd246\") " pod="openshift-marketplace/community-operators-6nhcb" Feb 17 13:38:56 crc kubenswrapper[4955]: I0217 13:38:56.779914 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51037160-13be-429b-991a-5b1e2ecdd246-catalog-content\") pod \"community-operators-6nhcb\" (UID: \"51037160-13be-429b-991a-5b1e2ecdd246\") " pod="openshift-marketplace/community-operators-6nhcb" Feb 17 13:38:56 crc kubenswrapper[4955]: I0217 13:38:56.780367 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51037160-13be-429b-991a-5b1e2ecdd246-catalog-content\") pod \"community-operators-6nhcb\" (UID: \"51037160-13be-429b-991a-5b1e2ecdd246\") " pod="openshift-marketplace/community-operators-6nhcb" Feb 17 13:38:56 crc kubenswrapper[4955]: I0217 13:38:56.780489 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51037160-13be-429b-991a-5b1e2ecdd246-utilities\") pod \"community-operators-6nhcb\" (UID: \"51037160-13be-429b-991a-5b1e2ecdd246\") " pod="openshift-marketplace/community-operators-6nhcb" Feb 17 13:38:56 crc kubenswrapper[4955]: I0217 13:38:56.805873 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7p759\" (UniqueName: \"kubernetes.io/projected/51037160-13be-429b-991a-5b1e2ecdd246-kube-api-access-7p759\") pod \"community-operators-6nhcb\" (UID: \"51037160-13be-429b-991a-5b1e2ecdd246\") " pod="openshift-marketplace/community-operators-6nhcb" Feb 17 13:38:56 crc kubenswrapper[4955]: I0217 13:38:56.940437 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6nhcb" Feb 17 13:38:57 crc kubenswrapper[4955]: I0217 13:38:57.483319 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6nhcb"] Feb 17 13:38:57 crc kubenswrapper[4955]: I0217 13:38:57.772905 4955 generic.go:334] "Generic (PLEG): container finished" podID="51037160-13be-429b-991a-5b1e2ecdd246" containerID="8ae5e7810e25711e85c4a291ccc17257b1597ba23de40c8c991f78a458bf00d8" exitCode=0 Feb 17 13:38:57 crc kubenswrapper[4955]: I0217 13:38:57.772972 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nhcb" event={"ID":"51037160-13be-429b-991a-5b1e2ecdd246","Type":"ContainerDied","Data":"8ae5e7810e25711e85c4a291ccc17257b1597ba23de40c8c991f78a458bf00d8"} Feb 17 13:38:57 crc kubenswrapper[4955]: I0217 13:38:57.773175 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nhcb" event={"ID":"51037160-13be-429b-991a-5b1e2ecdd246","Type":"ContainerStarted","Data":"c50be4f58a2910b03ba2cbab5a713a6614ab7060dfef1fa4c7c4d999e2eaa7f8"} Feb 17 13:38:59 crc kubenswrapper[4955]: I0217 13:38:59.792126 4955 generic.go:334] "Generic (PLEG): container finished" podID="51037160-13be-429b-991a-5b1e2ecdd246" containerID="5beb35032f601deda9dce2beee4bfc4df5801bd25dda9951c407d2cdff3e71bc" exitCode=0 Feb 17 13:38:59 crc kubenswrapper[4955]: I0217 13:38:59.792164 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nhcb" event={"ID":"51037160-13be-429b-991a-5b1e2ecdd246","Type":"ContainerDied","Data":"5beb35032f601deda9dce2beee4bfc4df5801bd25dda9951c407d2cdff3e71bc"} Feb 17 13:39:00 crc kubenswrapper[4955]: I0217 13:39:00.550519 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wbhxt" Feb 17 13:39:00 crc kubenswrapper[4955]: I0217 13:39:00.551020 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wbhxt" Feb 17 13:39:00 crc kubenswrapper[4955]: I0217 13:39:00.607220 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wbhxt" Feb 17 13:39:00 crc kubenswrapper[4955]: I0217 13:39:00.804513 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nhcb" event={"ID":"51037160-13be-429b-991a-5b1e2ecdd246","Type":"ContainerStarted","Data":"4d307d8e107cca53b2e09b8f3e0cb48f67aac5ca67356b0462c6b52731f0f880"} Feb 17 13:39:00 crc kubenswrapper[4955]: I0217 13:39:00.823940 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6nhcb" podStartSLOduration=2.326109761 podStartE2EDuration="4.823916208s" podCreationTimestamp="2026-02-17 13:38:56 +0000 UTC" firstStartedPulling="2026-02-17 13:38:57.774547773 +0000 UTC m=+2076.297277316" lastFinishedPulling="2026-02-17 13:39:00.27235422 +0000 UTC m=+2078.795083763" observedRunningTime="2026-02-17 13:39:00.823128495 +0000 UTC m=+2079.345858038" watchObservedRunningTime="2026-02-17 13:39:00.823916208 +0000 UTC m=+2079.346645751" Feb 17 13:39:00 crc kubenswrapper[4955]: I0217 13:39:00.853282 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wbhxt" Feb 17 13:39:02 crc kubenswrapper[4955]: I0217 13:39:02.580278 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wbhxt"] Feb 17 13:39:02 crc kubenswrapper[4955]: I0217 13:39:02.825065 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wbhxt" podUID="172f71ef-1b49-41d9-83a3-0dd423f73cff" containerName="registry-server" containerID="cri-o://6fc5d73363fe44838fa287c14fb11e6475f821d616fd9732ebe4c3dc844cbf55" gracePeriod=2 Feb 17 13:39:03 crc kubenswrapper[4955]: I0217 13:39:03.266438 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wbhxt" Feb 17 13:39:03 crc kubenswrapper[4955]: I0217 13:39:03.331681 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/172f71ef-1b49-41d9-83a3-0dd423f73cff-utilities\") pod \"172f71ef-1b49-41d9-83a3-0dd423f73cff\" (UID: \"172f71ef-1b49-41d9-83a3-0dd423f73cff\") " Feb 17 13:39:03 crc kubenswrapper[4955]: I0217 13:39:03.331828 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tvg55\" (UniqueName: \"kubernetes.io/projected/172f71ef-1b49-41d9-83a3-0dd423f73cff-kube-api-access-tvg55\") pod \"172f71ef-1b49-41d9-83a3-0dd423f73cff\" (UID: \"172f71ef-1b49-41d9-83a3-0dd423f73cff\") " Feb 17 13:39:03 crc kubenswrapper[4955]: I0217 13:39:03.331852 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/172f71ef-1b49-41d9-83a3-0dd423f73cff-catalog-content\") pod \"172f71ef-1b49-41d9-83a3-0dd423f73cff\" (UID: \"172f71ef-1b49-41d9-83a3-0dd423f73cff\") " Feb 17 13:39:03 crc kubenswrapper[4955]: I0217 13:39:03.332711 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/172f71ef-1b49-41d9-83a3-0dd423f73cff-utilities" (OuterVolumeSpecName: "utilities") pod "172f71ef-1b49-41d9-83a3-0dd423f73cff" (UID: "172f71ef-1b49-41d9-83a3-0dd423f73cff"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:39:03 crc kubenswrapper[4955]: I0217 13:39:03.339070 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/172f71ef-1b49-41d9-83a3-0dd423f73cff-kube-api-access-tvg55" (OuterVolumeSpecName: "kube-api-access-tvg55") pod "172f71ef-1b49-41d9-83a3-0dd423f73cff" (UID: "172f71ef-1b49-41d9-83a3-0dd423f73cff"). InnerVolumeSpecName "kube-api-access-tvg55". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:39:03 crc kubenswrapper[4955]: I0217 13:39:03.388710 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/172f71ef-1b49-41d9-83a3-0dd423f73cff-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "172f71ef-1b49-41d9-83a3-0dd423f73cff" (UID: "172f71ef-1b49-41d9-83a3-0dd423f73cff"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:39:03 crc kubenswrapper[4955]: I0217 13:39:03.434421 4955 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/172f71ef-1b49-41d9-83a3-0dd423f73cff-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 13:39:03 crc kubenswrapper[4955]: I0217 13:39:03.434468 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tvg55\" (UniqueName: \"kubernetes.io/projected/172f71ef-1b49-41d9-83a3-0dd423f73cff-kube-api-access-tvg55\") on node \"crc\" DevicePath \"\"" Feb 17 13:39:03 crc kubenswrapper[4955]: I0217 13:39:03.434482 4955 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/172f71ef-1b49-41d9-83a3-0dd423f73cff-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 13:39:03 crc kubenswrapper[4955]: I0217 13:39:03.837584 4955 generic.go:334] "Generic (PLEG): container finished" podID="172f71ef-1b49-41d9-83a3-0dd423f73cff" containerID="6fc5d73363fe44838fa287c14fb11e6475f821d616fd9732ebe4c3dc844cbf55" exitCode=0 Feb 17 13:39:03 crc kubenswrapper[4955]: I0217 13:39:03.837639 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wbhxt" event={"ID":"172f71ef-1b49-41d9-83a3-0dd423f73cff","Type":"ContainerDied","Data":"6fc5d73363fe44838fa287c14fb11e6475f821d616fd9732ebe4c3dc844cbf55"} Feb 17 13:39:03 crc kubenswrapper[4955]: I0217 13:39:03.837673 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wbhxt" event={"ID":"172f71ef-1b49-41d9-83a3-0dd423f73cff","Type":"ContainerDied","Data":"d2e2844b6ba48204c0654ebd9fda78d723a22b08e4385a32161aac83e145da0a"} Feb 17 13:39:03 crc kubenswrapper[4955]: I0217 13:39:03.837700 4955 scope.go:117] "RemoveContainer" containerID="6fc5d73363fe44838fa287c14fb11e6475f821d616fd9732ebe4c3dc844cbf55" Feb 17 13:39:03 crc kubenswrapper[4955]: I0217 13:39:03.837895 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wbhxt" Feb 17 13:39:03 crc kubenswrapper[4955]: I0217 13:39:03.859046 4955 scope.go:117] "RemoveContainer" containerID="e23f062d64b1531a53f7b9516a0a8ade9b8723342f6023cf5ada3ebd867df39b" Feb 17 13:39:03 crc kubenswrapper[4955]: I0217 13:39:03.874897 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wbhxt"] Feb 17 13:39:03 crc kubenswrapper[4955]: I0217 13:39:03.880850 4955 scope.go:117] "RemoveContainer" containerID="5ab9f3d31d048abebe969020062976bc928ebdf942876473e4d793e6515e2577" Feb 17 13:39:03 crc kubenswrapper[4955]: I0217 13:39:03.884294 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wbhxt"] Feb 17 13:39:03 crc kubenswrapper[4955]: I0217 13:39:03.927570 4955 scope.go:117] "RemoveContainer" containerID="6fc5d73363fe44838fa287c14fb11e6475f821d616fd9732ebe4c3dc844cbf55" Feb 17 13:39:03 crc kubenswrapper[4955]: E0217 13:39:03.928022 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6fc5d73363fe44838fa287c14fb11e6475f821d616fd9732ebe4c3dc844cbf55\": container with ID starting with 6fc5d73363fe44838fa287c14fb11e6475f821d616fd9732ebe4c3dc844cbf55 not found: ID does not exist" containerID="6fc5d73363fe44838fa287c14fb11e6475f821d616fd9732ebe4c3dc844cbf55" Feb 17 13:39:03 crc kubenswrapper[4955]: I0217 13:39:03.928084 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fc5d73363fe44838fa287c14fb11e6475f821d616fd9732ebe4c3dc844cbf55"} err="failed to get container status \"6fc5d73363fe44838fa287c14fb11e6475f821d616fd9732ebe4c3dc844cbf55\": rpc error: code = NotFound desc = could not find container \"6fc5d73363fe44838fa287c14fb11e6475f821d616fd9732ebe4c3dc844cbf55\": container with ID starting with 6fc5d73363fe44838fa287c14fb11e6475f821d616fd9732ebe4c3dc844cbf55 not found: ID does not exist" Feb 17 13:39:03 crc kubenswrapper[4955]: I0217 13:39:03.928118 4955 scope.go:117] "RemoveContainer" containerID="e23f062d64b1531a53f7b9516a0a8ade9b8723342f6023cf5ada3ebd867df39b" Feb 17 13:39:03 crc kubenswrapper[4955]: E0217 13:39:03.928416 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e23f062d64b1531a53f7b9516a0a8ade9b8723342f6023cf5ada3ebd867df39b\": container with ID starting with e23f062d64b1531a53f7b9516a0a8ade9b8723342f6023cf5ada3ebd867df39b not found: ID does not exist" containerID="e23f062d64b1531a53f7b9516a0a8ade9b8723342f6023cf5ada3ebd867df39b" Feb 17 13:39:03 crc kubenswrapper[4955]: I0217 13:39:03.928443 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e23f062d64b1531a53f7b9516a0a8ade9b8723342f6023cf5ada3ebd867df39b"} err="failed to get container status \"e23f062d64b1531a53f7b9516a0a8ade9b8723342f6023cf5ada3ebd867df39b\": rpc error: code = NotFound desc = could not find container \"e23f062d64b1531a53f7b9516a0a8ade9b8723342f6023cf5ada3ebd867df39b\": container with ID starting with e23f062d64b1531a53f7b9516a0a8ade9b8723342f6023cf5ada3ebd867df39b not found: ID does not exist" Feb 17 13:39:03 crc kubenswrapper[4955]: I0217 13:39:03.928460 4955 scope.go:117] "RemoveContainer" containerID="5ab9f3d31d048abebe969020062976bc928ebdf942876473e4d793e6515e2577" Feb 17 13:39:03 crc kubenswrapper[4955]: E0217 13:39:03.928733 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ab9f3d31d048abebe969020062976bc928ebdf942876473e4d793e6515e2577\": container with ID starting with 5ab9f3d31d048abebe969020062976bc928ebdf942876473e4d793e6515e2577 not found: ID does not exist" containerID="5ab9f3d31d048abebe969020062976bc928ebdf942876473e4d793e6515e2577" Feb 17 13:39:03 crc kubenswrapper[4955]: I0217 13:39:03.928764 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ab9f3d31d048abebe969020062976bc928ebdf942876473e4d793e6515e2577"} err="failed to get container status \"5ab9f3d31d048abebe969020062976bc928ebdf942876473e4d793e6515e2577\": rpc error: code = NotFound desc = could not find container \"5ab9f3d31d048abebe969020062976bc928ebdf942876473e4d793e6515e2577\": container with ID starting with 5ab9f3d31d048abebe969020062976bc928ebdf942876473e4d793e6515e2577 not found: ID does not exist" Feb 17 13:39:04 crc kubenswrapper[4955]: I0217 13:39:04.234535 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="172f71ef-1b49-41d9-83a3-0dd423f73cff" path="/var/lib/kubelet/pods/172f71ef-1b49-41d9-83a3-0dd423f73cff/volumes" Feb 17 13:39:06 crc kubenswrapper[4955]: I0217 13:39:06.941540 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6nhcb" Feb 17 13:39:06 crc kubenswrapper[4955]: I0217 13:39:06.942159 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6nhcb" Feb 17 13:39:07 crc kubenswrapper[4955]: I0217 13:39:07.000350 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6nhcb" Feb 17 13:39:07 crc kubenswrapper[4955]: I0217 13:39:07.933955 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6nhcb" Feb 17 13:39:07 crc kubenswrapper[4955]: I0217 13:39:07.982603 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6nhcb"] Feb 17 13:39:09 crc kubenswrapper[4955]: I0217 13:39:09.903596 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6nhcb" podUID="51037160-13be-429b-991a-5b1e2ecdd246" containerName="registry-server" containerID="cri-o://4d307d8e107cca53b2e09b8f3e0cb48f67aac5ca67356b0462c6b52731f0f880" gracePeriod=2 Feb 17 13:39:10 crc kubenswrapper[4955]: I0217 13:39:10.367169 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6nhcb" Feb 17 13:39:10 crc kubenswrapper[4955]: I0217 13:39:10.476995 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51037160-13be-429b-991a-5b1e2ecdd246-utilities\") pod \"51037160-13be-429b-991a-5b1e2ecdd246\" (UID: \"51037160-13be-429b-991a-5b1e2ecdd246\") " Feb 17 13:39:10 crc kubenswrapper[4955]: I0217 13:39:10.477142 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7p759\" (UniqueName: \"kubernetes.io/projected/51037160-13be-429b-991a-5b1e2ecdd246-kube-api-access-7p759\") pod \"51037160-13be-429b-991a-5b1e2ecdd246\" (UID: \"51037160-13be-429b-991a-5b1e2ecdd246\") " Feb 17 13:39:10 crc kubenswrapper[4955]: I0217 13:39:10.477185 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51037160-13be-429b-991a-5b1e2ecdd246-catalog-content\") pod \"51037160-13be-429b-991a-5b1e2ecdd246\" (UID: \"51037160-13be-429b-991a-5b1e2ecdd246\") " Feb 17 13:39:10 crc kubenswrapper[4955]: I0217 13:39:10.478213 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51037160-13be-429b-991a-5b1e2ecdd246-utilities" (OuterVolumeSpecName: "utilities") pod "51037160-13be-429b-991a-5b1e2ecdd246" (UID: "51037160-13be-429b-991a-5b1e2ecdd246"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:39:10 crc kubenswrapper[4955]: I0217 13:39:10.485216 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51037160-13be-429b-991a-5b1e2ecdd246-kube-api-access-7p759" (OuterVolumeSpecName: "kube-api-access-7p759") pod "51037160-13be-429b-991a-5b1e2ecdd246" (UID: "51037160-13be-429b-991a-5b1e2ecdd246"). InnerVolumeSpecName "kube-api-access-7p759". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:39:10 crc kubenswrapper[4955]: I0217 13:39:10.531796 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51037160-13be-429b-991a-5b1e2ecdd246-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "51037160-13be-429b-991a-5b1e2ecdd246" (UID: "51037160-13be-429b-991a-5b1e2ecdd246"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:39:10 crc kubenswrapper[4955]: I0217 13:39:10.590072 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7p759\" (UniqueName: \"kubernetes.io/projected/51037160-13be-429b-991a-5b1e2ecdd246-kube-api-access-7p759\") on node \"crc\" DevicePath \"\"" Feb 17 13:39:10 crc kubenswrapper[4955]: I0217 13:39:10.590124 4955 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51037160-13be-429b-991a-5b1e2ecdd246-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 13:39:10 crc kubenswrapper[4955]: I0217 13:39:10.590138 4955 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51037160-13be-429b-991a-5b1e2ecdd246-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 13:39:10 crc kubenswrapper[4955]: I0217 13:39:10.916713 4955 generic.go:334] "Generic (PLEG): container finished" podID="51037160-13be-429b-991a-5b1e2ecdd246" containerID="4d307d8e107cca53b2e09b8f3e0cb48f67aac5ca67356b0462c6b52731f0f880" exitCode=0 Feb 17 13:39:10 crc kubenswrapper[4955]: I0217 13:39:10.916831 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nhcb" event={"ID":"51037160-13be-429b-991a-5b1e2ecdd246","Type":"ContainerDied","Data":"4d307d8e107cca53b2e09b8f3e0cb48f67aac5ca67356b0462c6b52731f0f880"} Feb 17 13:39:10 crc kubenswrapper[4955]: I0217 13:39:10.917295 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nhcb" event={"ID":"51037160-13be-429b-991a-5b1e2ecdd246","Type":"ContainerDied","Data":"c50be4f58a2910b03ba2cbab5a713a6614ab7060dfef1fa4c7c4d999e2eaa7f8"} Feb 17 13:39:10 crc kubenswrapper[4955]: I0217 13:39:10.917331 4955 scope.go:117] "RemoveContainer" containerID="4d307d8e107cca53b2e09b8f3e0cb48f67aac5ca67356b0462c6b52731f0f880" Feb 17 13:39:10 crc kubenswrapper[4955]: I0217 13:39:10.916864 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6nhcb" Feb 17 13:39:10 crc kubenswrapper[4955]: I0217 13:39:10.945543 4955 scope.go:117] "RemoveContainer" containerID="5beb35032f601deda9dce2beee4bfc4df5801bd25dda9951c407d2cdff3e71bc" Feb 17 13:39:10 crc kubenswrapper[4955]: I0217 13:39:10.971771 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6nhcb"] Feb 17 13:39:10 crc kubenswrapper[4955]: I0217 13:39:10.981518 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6nhcb"] Feb 17 13:39:10 crc kubenswrapper[4955]: I0217 13:39:10.984818 4955 scope.go:117] "RemoveContainer" containerID="8ae5e7810e25711e85c4a291ccc17257b1597ba23de40c8c991f78a458bf00d8" Feb 17 13:39:11 crc kubenswrapper[4955]: I0217 13:39:11.032837 4955 scope.go:117] "RemoveContainer" containerID="4d307d8e107cca53b2e09b8f3e0cb48f67aac5ca67356b0462c6b52731f0f880" Feb 17 13:39:11 crc kubenswrapper[4955]: E0217 13:39:11.033380 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d307d8e107cca53b2e09b8f3e0cb48f67aac5ca67356b0462c6b52731f0f880\": container with ID starting with 4d307d8e107cca53b2e09b8f3e0cb48f67aac5ca67356b0462c6b52731f0f880 not found: ID does not exist" containerID="4d307d8e107cca53b2e09b8f3e0cb48f67aac5ca67356b0462c6b52731f0f880" Feb 17 13:39:11 crc kubenswrapper[4955]: I0217 13:39:11.033463 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d307d8e107cca53b2e09b8f3e0cb48f67aac5ca67356b0462c6b52731f0f880"} err="failed to get container status \"4d307d8e107cca53b2e09b8f3e0cb48f67aac5ca67356b0462c6b52731f0f880\": rpc error: code = NotFound desc = could not find container \"4d307d8e107cca53b2e09b8f3e0cb48f67aac5ca67356b0462c6b52731f0f880\": container with ID starting with 4d307d8e107cca53b2e09b8f3e0cb48f67aac5ca67356b0462c6b52731f0f880 not found: ID does not exist" Feb 17 13:39:11 crc kubenswrapper[4955]: I0217 13:39:11.033498 4955 scope.go:117] "RemoveContainer" containerID="5beb35032f601deda9dce2beee4bfc4df5801bd25dda9951c407d2cdff3e71bc" Feb 17 13:39:11 crc kubenswrapper[4955]: E0217 13:39:11.034043 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5beb35032f601deda9dce2beee4bfc4df5801bd25dda9951c407d2cdff3e71bc\": container with ID starting with 5beb35032f601deda9dce2beee4bfc4df5801bd25dda9951c407d2cdff3e71bc not found: ID does not exist" containerID="5beb35032f601deda9dce2beee4bfc4df5801bd25dda9951c407d2cdff3e71bc" Feb 17 13:39:11 crc kubenswrapper[4955]: I0217 13:39:11.034125 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5beb35032f601deda9dce2beee4bfc4df5801bd25dda9951c407d2cdff3e71bc"} err="failed to get container status \"5beb35032f601deda9dce2beee4bfc4df5801bd25dda9951c407d2cdff3e71bc\": rpc error: code = NotFound desc = could not find container \"5beb35032f601deda9dce2beee4bfc4df5801bd25dda9951c407d2cdff3e71bc\": container with ID starting with 5beb35032f601deda9dce2beee4bfc4df5801bd25dda9951c407d2cdff3e71bc not found: ID does not exist" Feb 17 13:39:11 crc kubenswrapper[4955]: I0217 13:39:11.034167 4955 scope.go:117] "RemoveContainer" containerID="8ae5e7810e25711e85c4a291ccc17257b1597ba23de40c8c991f78a458bf00d8" Feb 17 13:39:11 crc kubenswrapper[4955]: E0217 13:39:11.034544 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ae5e7810e25711e85c4a291ccc17257b1597ba23de40c8c991f78a458bf00d8\": container with ID starting with 8ae5e7810e25711e85c4a291ccc17257b1597ba23de40c8c991f78a458bf00d8 not found: ID does not exist" containerID="8ae5e7810e25711e85c4a291ccc17257b1597ba23de40c8c991f78a458bf00d8" Feb 17 13:39:11 crc kubenswrapper[4955]: I0217 13:39:11.034578 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ae5e7810e25711e85c4a291ccc17257b1597ba23de40c8c991f78a458bf00d8"} err="failed to get container status \"8ae5e7810e25711e85c4a291ccc17257b1597ba23de40c8c991f78a458bf00d8\": rpc error: code = NotFound desc = could not find container \"8ae5e7810e25711e85c4a291ccc17257b1597ba23de40c8c991f78a458bf00d8\": container with ID starting with 8ae5e7810e25711e85c4a291ccc17257b1597ba23de40c8c991f78a458bf00d8 not found: ID does not exist" Feb 17 13:39:12 crc kubenswrapper[4955]: I0217 13:39:12.236328 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51037160-13be-429b-991a-5b1e2ecdd246" path="/var/lib/kubelet/pods/51037160-13be-429b-991a-5b1e2ecdd246/volumes" Feb 17 13:40:34 crc kubenswrapper[4955]: I0217 13:40:34.574872 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 13:40:34 crc kubenswrapper[4955]: I0217 13:40:34.575458 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 13:41:04 crc kubenswrapper[4955]: I0217 13:41:04.574339 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 13:41:04 crc kubenswrapper[4955]: I0217 13:41:04.575021 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 13:41:34 crc kubenswrapper[4955]: I0217 13:41:34.574498 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 13:41:34 crc kubenswrapper[4955]: I0217 13:41:34.575088 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 13:41:34 crc kubenswrapper[4955]: I0217 13:41:34.575129 4955 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" Feb 17 13:41:34 crc kubenswrapper[4955]: I0217 13:41:34.575992 4955 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e81d66db8b1e6a694c97af3c245fca0ec09bc5dbf6e24a77ebe044515c97ad4e"} pod="openshift-machine-config-operator/machine-config-daemon-29qxq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 17 13:41:34 crc kubenswrapper[4955]: I0217 13:41:34.576057 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" containerID="cri-o://e81d66db8b1e6a694c97af3c245fca0ec09bc5dbf6e24a77ebe044515c97ad4e" gracePeriod=600 Feb 17 13:41:34 crc kubenswrapper[4955]: E0217 13:41:34.694366 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:41:35 crc kubenswrapper[4955]: I0217 13:41:35.204683 4955 generic.go:334] "Generic (PLEG): container finished" podID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerID="e81d66db8b1e6a694c97af3c245fca0ec09bc5dbf6e24a77ebe044515c97ad4e" exitCode=0 Feb 17 13:41:35 crc kubenswrapper[4955]: I0217 13:41:35.204751 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" event={"ID":"5bf36887-2a2b-4193-abdb-a15beaf46300","Type":"ContainerDied","Data":"e81d66db8b1e6a694c97af3c245fca0ec09bc5dbf6e24a77ebe044515c97ad4e"} Feb 17 13:41:35 crc kubenswrapper[4955]: I0217 13:41:35.204895 4955 scope.go:117] "RemoveContainer" containerID="8cd92c278f1f01b03343af64e98f7995ff57d30e66607734a82317d66752eafc" Feb 17 13:41:35 crc kubenswrapper[4955]: I0217 13:41:35.207528 4955 scope.go:117] "RemoveContainer" containerID="e81d66db8b1e6a694c97af3c245fca0ec09bc5dbf6e24a77ebe044515c97ad4e" Feb 17 13:41:35 crc kubenswrapper[4955]: E0217 13:41:35.208051 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:41:46 crc kubenswrapper[4955]: I0217 13:41:46.227052 4955 scope.go:117] "RemoveContainer" containerID="e81d66db8b1e6a694c97af3c245fca0ec09bc5dbf6e24a77ebe044515c97ad4e" Feb 17 13:41:46 crc kubenswrapper[4955]: E0217 13:41:46.231852 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:41:59 crc kubenswrapper[4955]: I0217 13:41:59.223856 4955 scope.go:117] "RemoveContainer" containerID="e81d66db8b1e6a694c97af3c245fca0ec09bc5dbf6e24a77ebe044515c97ad4e" Feb 17 13:41:59 crc kubenswrapper[4955]: E0217 13:41:59.224774 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:42:05 crc kubenswrapper[4955]: I0217 13:42:05.572659 4955 generic.go:334] "Generic (PLEG): container finished" podID="e3944196-3712-4506-bde5-758adf20aa63" containerID="6aa324d1361cd57e5a6a9797c1939e6d968539dbad7f337638ba6d0a3647bdd4" exitCode=0 Feb 17 13:42:05 crc kubenswrapper[4955]: I0217 13:42:05.572755 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr" event={"ID":"e3944196-3712-4506-bde5-758adf20aa63","Type":"ContainerDied","Data":"6aa324d1361cd57e5a6a9797c1939e6d968539dbad7f337638ba6d0a3647bdd4"} Feb 17 13:42:06 crc kubenswrapper[4955]: I0217 13:42:06.982596 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.078809 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e3944196-3712-4506-bde5-758adf20aa63-ssh-key-openstack-edpm-ipam\") pod \"e3944196-3712-4506-bde5-758adf20aa63\" (UID: \"e3944196-3712-4506-bde5-758adf20aa63\") " Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.079115 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3944196-3712-4506-bde5-758adf20aa63-libvirt-combined-ca-bundle\") pod \"e3944196-3712-4506-bde5-758adf20aa63\" (UID: \"e3944196-3712-4506-bde5-758adf20aa63\") " Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.079470 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/e3944196-3712-4506-bde5-758adf20aa63-libvirt-secret-0\") pod \"e3944196-3712-4506-bde5-758adf20aa63\" (UID: \"e3944196-3712-4506-bde5-758adf20aa63\") " Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.079552 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mcbhl\" (UniqueName: \"kubernetes.io/projected/e3944196-3712-4506-bde5-758adf20aa63-kube-api-access-mcbhl\") pod \"e3944196-3712-4506-bde5-758adf20aa63\" (UID: \"e3944196-3712-4506-bde5-758adf20aa63\") " Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.079591 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3944196-3712-4506-bde5-758adf20aa63-inventory\") pod \"e3944196-3712-4506-bde5-758adf20aa63\" (UID: \"e3944196-3712-4506-bde5-758adf20aa63\") " Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.086273 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3944196-3712-4506-bde5-758adf20aa63-kube-api-access-mcbhl" (OuterVolumeSpecName: "kube-api-access-mcbhl") pod "e3944196-3712-4506-bde5-758adf20aa63" (UID: "e3944196-3712-4506-bde5-758adf20aa63"). InnerVolumeSpecName "kube-api-access-mcbhl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.088945 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3944196-3712-4506-bde5-758adf20aa63-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "e3944196-3712-4506-bde5-758adf20aa63" (UID: "e3944196-3712-4506-bde5-758adf20aa63"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.112164 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3944196-3712-4506-bde5-758adf20aa63-inventory" (OuterVolumeSpecName: "inventory") pod "e3944196-3712-4506-bde5-758adf20aa63" (UID: "e3944196-3712-4506-bde5-758adf20aa63"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.115249 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3944196-3712-4506-bde5-758adf20aa63-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "e3944196-3712-4506-bde5-758adf20aa63" (UID: "e3944196-3712-4506-bde5-758adf20aa63"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.125849 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3944196-3712-4506-bde5-758adf20aa63-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "e3944196-3712-4506-bde5-758adf20aa63" (UID: "e3944196-3712-4506-bde5-758adf20aa63"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.182029 4955 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e3944196-3712-4506-bde5-758adf20aa63-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.182071 4955 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3944196-3712-4506-bde5-758adf20aa63-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.182080 4955 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/e3944196-3712-4506-bde5-758adf20aa63-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.182091 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mcbhl\" (UniqueName: \"kubernetes.io/projected/e3944196-3712-4506-bde5-758adf20aa63-kube-api-access-mcbhl\") on node \"crc\" DevicePath \"\"" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.182100 4955 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3944196-3712-4506-bde5-758adf20aa63-inventory\") on node \"crc\" DevicePath \"\"" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.589184 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr" event={"ID":"e3944196-3712-4506-bde5-758adf20aa63","Type":"ContainerDied","Data":"121e2741a57c0572b32f5d7f2a51af51dbbdaa6d14016a144eb125c2b183dc06"} Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.589239 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="121e2741a57c0572b32f5d7f2a51af51dbbdaa6d14016a144eb125c2b183dc06" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.589291 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.699155 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l"] Feb 17 13:42:07 crc kubenswrapper[4955]: E0217 13:42:07.699646 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="172f71ef-1b49-41d9-83a3-0dd423f73cff" containerName="extract-utilities" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.699671 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="172f71ef-1b49-41d9-83a3-0dd423f73cff" containerName="extract-utilities" Feb 17 13:42:07 crc kubenswrapper[4955]: E0217 13:42:07.699705 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51037160-13be-429b-991a-5b1e2ecdd246" containerName="extract-utilities" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.699715 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="51037160-13be-429b-991a-5b1e2ecdd246" containerName="extract-utilities" Feb 17 13:42:07 crc kubenswrapper[4955]: E0217 13:42:07.699737 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3944196-3712-4506-bde5-758adf20aa63" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.699747 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3944196-3712-4506-bde5-758adf20aa63" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Feb 17 13:42:07 crc kubenswrapper[4955]: E0217 13:42:07.699758 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51037160-13be-429b-991a-5b1e2ecdd246" containerName="extract-content" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.699766 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="51037160-13be-429b-991a-5b1e2ecdd246" containerName="extract-content" Feb 17 13:42:07 crc kubenswrapper[4955]: E0217 13:42:07.699793 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51037160-13be-429b-991a-5b1e2ecdd246" containerName="registry-server" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.699802 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="51037160-13be-429b-991a-5b1e2ecdd246" containerName="registry-server" Feb 17 13:42:07 crc kubenswrapper[4955]: E0217 13:42:07.699813 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="172f71ef-1b49-41d9-83a3-0dd423f73cff" containerName="registry-server" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.699821 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="172f71ef-1b49-41d9-83a3-0dd423f73cff" containerName="registry-server" Feb 17 13:42:07 crc kubenswrapper[4955]: E0217 13:42:07.699834 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="172f71ef-1b49-41d9-83a3-0dd423f73cff" containerName="extract-content" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.699842 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="172f71ef-1b49-41d9-83a3-0dd423f73cff" containerName="extract-content" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.700066 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="51037160-13be-429b-991a-5b1e2ecdd246" containerName="registry-server" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.700089 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3944196-3712-4506-bde5-758adf20aa63" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.700102 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="172f71ef-1b49-41d9-83a3-0dd423f73cff" containerName="registry-server" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.700975 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.703413 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.703616 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.703849 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-t9lgc" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.703852 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.704107 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.704535 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.705986 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.709832 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l"] Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.793047 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v478l\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.793093 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v478l\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.793121 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v478l\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.793164 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v478l\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.793252 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v478l\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.793286 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v478l\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.793343 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v478l\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.793395 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwjdk\" (UniqueName: \"kubernetes.io/projected/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-kube-api-access-qwjdk\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v478l\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.793428 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v478l\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.793497 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v478l\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.793524 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v478l\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.894684 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v478l\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.894734 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v478l\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.894765 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v478l\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.894806 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v478l\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.894844 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v478l\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.894897 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v478l\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.894915 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v478l\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.894945 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v478l\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.894988 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwjdk\" (UniqueName: \"kubernetes.io/projected/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-kube-api-access-qwjdk\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v478l\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.895020 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v478l\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.895055 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v478l\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.896257 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v478l\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.900165 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v478l\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.901318 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v478l\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.901364 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v478l\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.901440 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v478l\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.901881 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v478l\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.902280 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v478l\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.902666 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v478l\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.907042 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v478l\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.910094 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v478l\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l" Feb 17 13:42:07 crc kubenswrapper[4955]: I0217 13:42:07.912374 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwjdk\" (UniqueName: \"kubernetes.io/projected/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-kube-api-access-qwjdk\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v478l\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l" Feb 17 13:42:08 crc kubenswrapper[4955]: I0217 13:42:08.030126 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l" Feb 17 13:42:08 crc kubenswrapper[4955]: I0217 13:42:08.642053 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l"] Feb 17 13:42:08 crc kubenswrapper[4955]: I0217 13:42:08.650342 4955 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 17 13:42:09 crc kubenswrapper[4955]: I0217 13:42:09.607402 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l" event={"ID":"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f","Type":"ContainerStarted","Data":"1b8ed0a96b75a05a34f011b52a4305926c1c6d7a647d95ec41c2c1ac42f551bc"} Feb 17 13:42:09 crc kubenswrapper[4955]: I0217 13:42:09.607758 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l" event={"ID":"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f","Type":"ContainerStarted","Data":"81ded8e2940e3a9f82f52353ccbf38d872390991a2a91c0b935708d024da6510"} Feb 17 13:42:09 crc kubenswrapper[4955]: I0217 13:42:09.634572 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l" podStartSLOduration=2.211043169 podStartE2EDuration="2.634550498s" podCreationTimestamp="2026-02-17 13:42:07 +0000 UTC" firstStartedPulling="2026-02-17 13:42:08.650061377 +0000 UTC m=+2267.172790920" lastFinishedPulling="2026-02-17 13:42:09.073568706 +0000 UTC m=+2267.596298249" observedRunningTime="2026-02-17 13:42:09.626333644 +0000 UTC m=+2268.149063187" watchObservedRunningTime="2026-02-17 13:42:09.634550498 +0000 UTC m=+2268.157280041" Feb 17 13:42:14 crc kubenswrapper[4955]: I0217 13:42:14.223210 4955 scope.go:117] "RemoveContainer" containerID="e81d66db8b1e6a694c97af3c245fca0ec09bc5dbf6e24a77ebe044515c97ad4e" Feb 17 13:42:14 crc kubenswrapper[4955]: E0217 13:42:14.223962 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:42:29 crc kubenswrapper[4955]: I0217 13:42:29.222643 4955 scope.go:117] "RemoveContainer" containerID="e81d66db8b1e6a694c97af3c245fca0ec09bc5dbf6e24a77ebe044515c97ad4e" Feb 17 13:42:29 crc kubenswrapper[4955]: E0217 13:42:29.223704 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:42:42 crc kubenswrapper[4955]: I0217 13:42:42.229213 4955 scope.go:117] "RemoveContainer" containerID="e81d66db8b1e6a694c97af3c245fca0ec09bc5dbf6e24a77ebe044515c97ad4e" Feb 17 13:42:42 crc kubenswrapper[4955]: E0217 13:42:42.230061 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:42:56 crc kubenswrapper[4955]: I0217 13:42:56.224131 4955 scope.go:117] "RemoveContainer" containerID="e81d66db8b1e6a694c97af3c245fca0ec09bc5dbf6e24a77ebe044515c97ad4e" Feb 17 13:42:56 crc kubenswrapper[4955]: E0217 13:42:56.225024 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:43:11 crc kubenswrapper[4955]: I0217 13:43:11.222771 4955 scope.go:117] "RemoveContainer" containerID="e81d66db8b1e6a694c97af3c245fca0ec09bc5dbf6e24a77ebe044515c97ad4e" Feb 17 13:43:11 crc kubenswrapper[4955]: E0217 13:43:11.223592 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:43:22 crc kubenswrapper[4955]: I0217 13:43:22.229978 4955 scope.go:117] "RemoveContainer" containerID="e81d66db8b1e6a694c97af3c245fca0ec09bc5dbf6e24a77ebe044515c97ad4e" Feb 17 13:43:22 crc kubenswrapper[4955]: E0217 13:43:22.230726 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:43:35 crc kubenswrapper[4955]: I0217 13:43:35.223618 4955 scope.go:117] "RemoveContainer" containerID="e81d66db8b1e6a694c97af3c245fca0ec09bc5dbf6e24a77ebe044515c97ad4e" Feb 17 13:43:35 crc kubenswrapper[4955]: E0217 13:43:35.224662 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:43:49 crc kubenswrapper[4955]: I0217 13:43:49.223084 4955 scope.go:117] "RemoveContainer" containerID="e81d66db8b1e6a694c97af3c245fca0ec09bc5dbf6e24a77ebe044515c97ad4e" Feb 17 13:43:49 crc kubenswrapper[4955]: E0217 13:43:49.228640 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:44:04 crc kubenswrapper[4955]: I0217 13:44:04.222994 4955 scope.go:117] "RemoveContainer" containerID="e81d66db8b1e6a694c97af3c245fca0ec09bc5dbf6e24a77ebe044515c97ad4e" Feb 17 13:44:04 crc kubenswrapper[4955]: E0217 13:44:04.223997 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:44:15 crc kubenswrapper[4955]: I0217 13:44:15.223883 4955 scope.go:117] "RemoveContainer" containerID="e81d66db8b1e6a694c97af3c245fca0ec09bc5dbf6e24a77ebe044515c97ad4e" Feb 17 13:44:15 crc kubenswrapper[4955]: E0217 13:44:15.224622 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:44:15 crc kubenswrapper[4955]: I0217 13:44:15.726888 4955 generic.go:334] "Generic (PLEG): container finished" podID="1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f" containerID="1b8ed0a96b75a05a34f011b52a4305926c1c6d7a647d95ec41c2c1ac42f551bc" exitCode=0 Feb 17 13:44:15 crc kubenswrapper[4955]: I0217 13:44:15.726973 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l" event={"ID":"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f","Type":"ContainerDied","Data":"1b8ed0a96b75a05a34f011b52a4305926c1c6d7a647d95ec41c2c1ac42f551bc"} Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.089530 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.179185 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-cell1-compute-config-2\") pod \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.179288 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-migration-ssh-key-0\") pod \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.179345 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwjdk\" (UniqueName: \"kubernetes.io/projected/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-kube-api-access-qwjdk\") pod \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.179382 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-migration-ssh-key-1\") pod \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.179480 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-inventory\") pod \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.179523 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-combined-ca-bundle\") pod \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.179559 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-ssh-key-openstack-edpm-ipam\") pod \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.179661 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-cell1-compute-config-1\") pod \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.179689 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-extra-config-0\") pod \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.179762 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-cell1-compute-config-3\") pod \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.179805 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-cell1-compute-config-0\") pod \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\" (UID: \"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f\") " Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.197824 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f" (UID: "1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.202172 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-kube-api-access-qwjdk" (OuterVolumeSpecName: "kube-api-access-qwjdk") pod "1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f" (UID: "1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f"). InnerVolumeSpecName "kube-api-access-qwjdk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.204800 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f" (UID: "1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.207012 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f" (UID: "1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.211879 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f" (UID: "1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.216989 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f" (UID: "1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.219164 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f" (UID: "1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.222289 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-cell1-compute-config-2" (OuterVolumeSpecName: "nova-cell1-compute-config-2") pod "1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f" (UID: "1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f"). InnerVolumeSpecName "nova-cell1-compute-config-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.229209 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f" (UID: "1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.230675 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-inventory" (OuterVolumeSpecName: "inventory") pod "1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f" (UID: "1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.237026 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-cell1-compute-config-3" (OuterVolumeSpecName: "nova-cell1-compute-config-3") pod "1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f" (UID: "1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f"). InnerVolumeSpecName "nova-cell1-compute-config-3". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.282551 4955 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.282593 4955 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.282602 4955 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-cell1-compute-config-3\") on node \"crc\" DevicePath \"\"" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.282612 4955 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.282622 4955 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-cell1-compute-config-2\") on node \"crc\" DevicePath \"\"" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.282634 4955 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.282647 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwjdk\" (UniqueName: \"kubernetes.io/projected/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-kube-api-access-qwjdk\") on node \"crc\" DevicePath \"\"" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.282658 4955 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.282672 4955 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-inventory\") on node \"crc\" DevicePath \"\"" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.282683 4955 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.282696 4955 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.741974 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l" event={"ID":"1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f","Type":"ContainerDied","Data":"81ded8e2940e3a9f82f52353ccbf38d872390991a2a91c0b935708d024da6510"} Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.742010 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="81ded8e2940e3a9f82f52353ccbf38d872390991a2a91c0b935708d024da6510" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.742078 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v478l" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.853263 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npz67"] Feb 17 13:44:17 crc kubenswrapper[4955]: E0217 13:44:17.853714 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f" containerName="nova-edpm-deployment-openstack-edpm-ipam" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.853729 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f" containerName="nova-edpm-deployment-openstack-edpm-ipam" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.854013 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f" containerName="nova-edpm-deployment-openstack-edpm-ipam" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.854734 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npz67" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.858240 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-t9lgc" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.858428 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.858649 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.859189 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.861557 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.863614 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npz67"] Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.893313 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnj7f\" (UniqueName: \"kubernetes.io/projected/0acb7fb8-671d-4e90-8be4-f5f507308f11-kube-api-access-gnj7f\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-npz67\" (UID: \"0acb7fb8-671d-4e90-8be4-f5f507308f11\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npz67" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.893689 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0acb7fb8-671d-4e90-8be4-f5f507308f11-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-npz67\" (UID: \"0acb7fb8-671d-4e90-8be4-f5f507308f11\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npz67" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.893744 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/0acb7fb8-671d-4e90-8be4-f5f507308f11-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-npz67\" (UID: \"0acb7fb8-671d-4e90-8be4-f5f507308f11\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npz67" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.893771 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/0acb7fb8-671d-4e90-8be4-f5f507308f11-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-npz67\" (UID: \"0acb7fb8-671d-4e90-8be4-f5f507308f11\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npz67" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.893824 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/0acb7fb8-671d-4e90-8be4-f5f507308f11-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-npz67\" (UID: \"0acb7fb8-671d-4e90-8be4-f5f507308f11\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npz67" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.893843 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0acb7fb8-671d-4e90-8be4-f5f507308f11-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-npz67\" (UID: \"0acb7fb8-671d-4e90-8be4-f5f507308f11\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npz67" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.893875 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0acb7fb8-671d-4e90-8be4-f5f507308f11-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-npz67\" (UID: \"0acb7fb8-671d-4e90-8be4-f5f507308f11\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npz67" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.995976 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/0acb7fb8-671d-4e90-8be4-f5f507308f11-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-npz67\" (UID: \"0acb7fb8-671d-4e90-8be4-f5f507308f11\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npz67" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.997008 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/0acb7fb8-671d-4e90-8be4-f5f507308f11-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-npz67\" (UID: \"0acb7fb8-671d-4e90-8be4-f5f507308f11\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npz67" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.997131 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/0acb7fb8-671d-4e90-8be4-f5f507308f11-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-npz67\" (UID: \"0acb7fb8-671d-4e90-8be4-f5f507308f11\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npz67" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.997168 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0acb7fb8-671d-4e90-8be4-f5f507308f11-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-npz67\" (UID: \"0acb7fb8-671d-4e90-8be4-f5f507308f11\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npz67" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.997248 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0acb7fb8-671d-4e90-8be4-f5f507308f11-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-npz67\" (UID: \"0acb7fb8-671d-4e90-8be4-f5f507308f11\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npz67" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.997394 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnj7f\" (UniqueName: \"kubernetes.io/projected/0acb7fb8-671d-4e90-8be4-f5f507308f11-kube-api-access-gnj7f\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-npz67\" (UID: \"0acb7fb8-671d-4e90-8be4-f5f507308f11\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npz67" Feb 17 13:44:17 crc kubenswrapper[4955]: I0217 13:44:17.997554 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0acb7fb8-671d-4e90-8be4-f5f507308f11-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-npz67\" (UID: \"0acb7fb8-671d-4e90-8be4-f5f507308f11\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npz67" Feb 17 13:44:18 crc kubenswrapper[4955]: I0217 13:44:18.002500 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/0acb7fb8-671d-4e90-8be4-f5f507308f11-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-npz67\" (UID: \"0acb7fb8-671d-4e90-8be4-f5f507308f11\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npz67" Feb 17 13:44:18 crc kubenswrapper[4955]: I0217 13:44:18.002609 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/0acb7fb8-671d-4e90-8be4-f5f507308f11-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-npz67\" (UID: \"0acb7fb8-671d-4e90-8be4-f5f507308f11\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npz67" Feb 17 13:44:18 crc kubenswrapper[4955]: I0217 13:44:18.002970 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0acb7fb8-671d-4e90-8be4-f5f507308f11-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-npz67\" (UID: \"0acb7fb8-671d-4e90-8be4-f5f507308f11\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npz67" Feb 17 13:44:18 crc kubenswrapper[4955]: I0217 13:44:18.003089 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/0acb7fb8-671d-4e90-8be4-f5f507308f11-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-npz67\" (UID: \"0acb7fb8-671d-4e90-8be4-f5f507308f11\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npz67" Feb 17 13:44:18 crc kubenswrapper[4955]: I0217 13:44:18.003910 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0acb7fb8-671d-4e90-8be4-f5f507308f11-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-npz67\" (UID: \"0acb7fb8-671d-4e90-8be4-f5f507308f11\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npz67" Feb 17 13:44:18 crc kubenswrapper[4955]: I0217 13:44:18.004444 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0acb7fb8-671d-4e90-8be4-f5f507308f11-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-npz67\" (UID: \"0acb7fb8-671d-4e90-8be4-f5f507308f11\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npz67" Feb 17 13:44:18 crc kubenswrapper[4955]: I0217 13:44:18.018326 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnj7f\" (UniqueName: \"kubernetes.io/projected/0acb7fb8-671d-4e90-8be4-f5f507308f11-kube-api-access-gnj7f\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-npz67\" (UID: \"0acb7fb8-671d-4e90-8be4-f5f507308f11\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npz67" Feb 17 13:44:18 crc kubenswrapper[4955]: I0217 13:44:18.170641 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npz67" Feb 17 13:44:18 crc kubenswrapper[4955]: I0217 13:44:18.673654 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npz67"] Feb 17 13:44:18 crc kubenswrapper[4955]: I0217 13:44:18.749913 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npz67" event={"ID":"0acb7fb8-671d-4e90-8be4-f5f507308f11","Type":"ContainerStarted","Data":"1da60edad11e2f0dfcd9de5c646859cd8cf1a720a01e6b2c70a5a7263ff73bc9"} Feb 17 13:44:19 crc kubenswrapper[4955]: I0217 13:44:19.759706 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npz67" event={"ID":"0acb7fb8-671d-4e90-8be4-f5f507308f11","Type":"ContainerStarted","Data":"ab524817b95eb90df2f8dfc9b797aba13ed2c4616293b5efba888e54d7c4ef43"} Feb 17 13:44:19 crc kubenswrapper[4955]: I0217 13:44:19.776211 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npz67" podStartSLOduration=2.358133522 podStartE2EDuration="2.776191013s" podCreationTimestamp="2026-02-17 13:44:17 +0000 UTC" firstStartedPulling="2026-02-17 13:44:18.675898238 +0000 UTC m=+2397.198627781" lastFinishedPulling="2026-02-17 13:44:19.093955729 +0000 UTC m=+2397.616685272" observedRunningTime="2026-02-17 13:44:19.773238599 +0000 UTC m=+2398.295968142" watchObservedRunningTime="2026-02-17 13:44:19.776191013 +0000 UTC m=+2398.298920556" Feb 17 13:44:27 crc kubenswrapper[4955]: I0217 13:44:27.222505 4955 scope.go:117] "RemoveContainer" containerID="e81d66db8b1e6a694c97af3c245fca0ec09bc5dbf6e24a77ebe044515c97ad4e" Feb 17 13:44:27 crc kubenswrapper[4955]: E0217 13:44:27.223232 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:44:39 crc kubenswrapper[4955]: I0217 13:44:39.222458 4955 scope.go:117] "RemoveContainer" containerID="e81d66db8b1e6a694c97af3c245fca0ec09bc5dbf6e24a77ebe044515c97ad4e" Feb 17 13:44:39 crc kubenswrapper[4955]: E0217 13:44:39.223211 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:44:52 crc kubenswrapper[4955]: I0217 13:44:52.228533 4955 scope.go:117] "RemoveContainer" containerID="e81d66db8b1e6a694c97af3c245fca0ec09bc5dbf6e24a77ebe044515c97ad4e" Feb 17 13:44:52 crc kubenswrapper[4955]: E0217 13:44:52.229411 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:45:00 crc kubenswrapper[4955]: I0217 13:45:00.145583 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29522265-hfdwg"] Feb 17 13:45:00 crc kubenswrapper[4955]: I0217 13:45:00.147259 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29522265-hfdwg" Feb 17 13:45:00 crc kubenswrapper[4955]: I0217 13:45:00.150860 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 17 13:45:00 crc kubenswrapper[4955]: I0217 13:45:00.150980 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 17 13:45:00 crc kubenswrapper[4955]: I0217 13:45:00.156956 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29522265-hfdwg"] Feb 17 13:45:00 crc kubenswrapper[4955]: I0217 13:45:00.331603 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3861a08f-b7f7-43b7-80b0-80f0f9289de9-secret-volume\") pod \"collect-profiles-29522265-hfdwg\" (UID: \"3861a08f-b7f7-43b7-80b0-80f0f9289de9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522265-hfdwg" Feb 17 13:45:00 crc kubenswrapper[4955]: I0217 13:45:00.332072 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3861a08f-b7f7-43b7-80b0-80f0f9289de9-config-volume\") pod \"collect-profiles-29522265-hfdwg\" (UID: \"3861a08f-b7f7-43b7-80b0-80f0f9289de9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522265-hfdwg" Feb 17 13:45:00 crc kubenswrapper[4955]: I0217 13:45:00.332120 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-md7df\" (UniqueName: \"kubernetes.io/projected/3861a08f-b7f7-43b7-80b0-80f0f9289de9-kube-api-access-md7df\") pod \"collect-profiles-29522265-hfdwg\" (UID: \"3861a08f-b7f7-43b7-80b0-80f0f9289de9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522265-hfdwg" Feb 17 13:45:00 crc kubenswrapper[4955]: I0217 13:45:00.434120 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3861a08f-b7f7-43b7-80b0-80f0f9289de9-config-volume\") pod \"collect-profiles-29522265-hfdwg\" (UID: \"3861a08f-b7f7-43b7-80b0-80f0f9289de9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522265-hfdwg" Feb 17 13:45:00 crc kubenswrapper[4955]: I0217 13:45:00.434204 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-md7df\" (UniqueName: \"kubernetes.io/projected/3861a08f-b7f7-43b7-80b0-80f0f9289de9-kube-api-access-md7df\") pod \"collect-profiles-29522265-hfdwg\" (UID: \"3861a08f-b7f7-43b7-80b0-80f0f9289de9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522265-hfdwg" Feb 17 13:45:00 crc kubenswrapper[4955]: I0217 13:45:00.434294 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3861a08f-b7f7-43b7-80b0-80f0f9289de9-secret-volume\") pod \"collect-profiles-29522265-hfdwg\" (UID: \"3861a08f-b7f7-43b7-80b0-80f0f9289de9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522265-hfdwg" Feb 17 13:45:00 crc kubenswrapper[4955]: I0217 13:45:00.435130 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3861a08f-b7f7-43b7-80b0-80f0f9289de9-config-volume\") pod \"collect-profiles-29522265-hfdwg\" (UID: \"3861a08f-b7f7-43b7-80b0-80f0f9289de9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522265-hfdwg" Feb 17 13:45:00 crc kubenswrapper[4955]: I0217 13:45:00.440806 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3861a08f-b7f7-43b7-80b0-80f0f9289de9-secret-volume\") pod \"collect-profiles-29522265-hfdwg\" (UID: \"3861a08f-b7f7-43b7-80b0-80f0f9289de9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522265-hfdwg" Feb 17 13:45:00 crc kubenswrapper[4955]: I0217 13:45:00.452850 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-md7df\" (UniqueName: \"kubernetes.io/projected/3861a08f-b7f7-43b7-80b0-80f0f9289de9-kube-api-access-md7df\") pod \"collect-profiles-29522265-hfdwg\" (UID: \"3861a08f-b7f7-43b7-80b0-80f0f9289de9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522265-hfdwg" Feb 17 13:45:00 crc kubenswrapper[4955]: I0217 13:45:00.467974 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29522265-hfdwg" Feb 17 13:45:00 crc kubenswrapper[4955]: I0217 13:45:00.971938 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29522265-hfdwg"] Feb 17 13:45:01 crc kubenswrapper[4955]: I0217 13:45:01.122339 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29522265-hfdwg" event={"ID":"3861a08f-b7f7-43b7-80b0-80f0f9289de9","Type":"ContainerStarted","Data":"3c6107c31bf69d40ff4bbdf4b31ed40b0222d7048d4d3dbf4f90dc3018488f6f"} Feb 17 13:45:02 crc kubenswrapper[4955]: I0217 13:45:02.131734 4955 generic.go:334] "Generic (PLEG): container finished" podID="3861a08f-b7f7-43b7-80b0-80f0f9289de9" containerID="f068340c79ece6cb95b698ea50b3506cca3fa3500c220875f0bba14b63df0c3c" exitCode=0 Feb 17 13:45:02 crc kubenswrapper[4955]: I0217 13:45:02.131818 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29522265-hfdwg" event={"ID":"3861a08f-b7f7-43b7-80b0-80f0f9289de9","Type":"ContainerDied","Data":"f068340c79ece6cb95b698ea50b3506cca3fa3500c220875f0bba14b63df0c3c"} Feb 17 13:45:03 crc kubenswrapper[4955]: I0217 13:45:03.453310 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29522265-hfdwg" Feb 17 13:45:03 crc kubenswrapper[4955]: I0217 13:45:03.593327 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3861a08f-b7f7-43b7-80b0-80f0f9289de9-config-volume\") pod \"3861a08f-b7f7-43b7-80b0-80f0f9289de9\" (UID: \"3861a08f-b7f7-43b7-80b0-80f0f9289de9\") " Feb 17 13:45:03 crc kubenswrapper[4955]: I0217 13:45:03.593538 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-md7df\" (UniqueName: \"kubernetes.io/projected/3861a08f-b7f7-43b7-80b0-80f0f9289de9-kube-api-access-md7df\") pod \"3861a08f-b7f7-43b7-80b0-80f0f9289de9\" (UID: \"3861a08f-b7f7-43b7-80b0-80f0f9289de9\") " Feb 17 13:45:03 crc kubenswrapper[4955]: I0217 13:45:03.593573 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3861a08f-b7f7-43b7-80b0-80f0f9289de9-secret-volume\") pod \"3861a08f-b7f7-43b7-80b0-80f0f9289de9\" (UID: \"3861a08f-b7f7-43b7-80b0-80f0f9289de9\") " Feb 17 13:45:03 crc kubenswrapper[4955]: I0217 13:45:03.594454 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3861a08f-b7f7-43b7-80b0-80f0f9289de9-config-volume" (OuterVolumeSpecName: "config-volume") pod "3861a08f-b7f7-43b7-80b0-80f0f9289de9" (UID: "3861a08f-b7f7-43b7-80b0-80f0f9289de9"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:45:03 crc kubenswrapper[4955]: I0217 13:45:03.603462 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3861a08f-b7f7-43b7-80b0-80f0f9289de9-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3861a08f-b7f7-43b7-80b0-80f0f9289de9" (UID: "3861a08f-b7f7-43b7-80b0-80f0f9289de9"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:45:03 crc kubenswrapper[4955]: I0217 13:45:03.604245 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3861a08f-b7f7-43b7-80b0-80f0f9289de9-kube-api-access-md7df" (OuterVolumeSpecName: "kube-api-access-md7df") pod "3861a08f-b7f7-43b7-80b0-80f0f9289de9" (UID: "3861a08f-b7f7-43b7-80b0-80f0f9289de9"). InnerVolumeSpecName "kube-api-access-md7df". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:45:03 crc kubenswrapper[4955]: I0217 13:45:03.695389 4955 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3861a08f-b7f7-43b7-80b0-80f0f9289de9-config-volume\") on node \"crc\" DevicePath \"\"" Feb 17 13:45:03 crc kubenswrapper[4955]: I0217 13:45:03.695430 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-md7df\" (UniqueName: \"kubernetes.io/projected/3861a08f-b7f7-43b7-80b0-80f0f9289de9-kube-api-access-md7df\") on node \"crc\" DevicePath \"\"" Feb 17 13:45:03 crc kubenswrapper[4955]: I0217 13:45:03.695440 4955 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3861a08f-b7f7-43b7-80b0-80f0f9289de9-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 17 13:45:04 crc kubenswrapper[4955]: I0217 13:45:04.147908 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29522265-hfdwg" event={"ID":"3861a08f-b7f7-43b7-80b0-80f0f9289de9","Type":"ContainerDied","Data":"3c6107c31bf69d40ff4bbdf4b31ed40b0222d7048d4d3dbf4f90dc3018488f6f"} Feb 17 13:45:04 crc kubenswrapper[4955]: I0217 13:45:04.147949 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c6107c31bf69d40ff4bbdf4b31ed40b0222d7048d4d3dbf4f90dc3018488f6f" Feb 17 13:45:04 crc kubenswrapper[4955]: I0217 13:45:04.148229 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29522265-hfdwg" Feb 17 13:45:04 crc kubenswrapper[4955]: I0217 13:45:04.541237 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29522220-v5twm"] Feb 17 13:45:04 crc kubenswrapper[4955]: I0217 13:45:04.549193 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29522220-v5twm"] Feb 17 13:45:06 crc kubenswrapper[4955]: I0217 13:45:06.235237 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93fe04d3-c430-4ed6-ae04-312cf0613834" path="/var/lib/kubelet/pods/93fe04d3-c430-4ed6-ae04-312cf0613834/volumes" Feb 17 13:45:07 crc kubenswrapper[4955]: I0217 13:45:07.224389 4955 scope.go:117] "RemoveContainer" containerID="e81d66db8b1e6a694c97af3c245fca0ec09bc5dbf6e24a77ebe044515c97ad4e" Feb 17 13:45:07 crc kubenswrapper[4955]: E0217 13:45:07.224650 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:45:18 crc kubenswrapper[4955]: I0217 13:45:18.223440 4955 scope.go:117] "RemoveContainer" containerID="e81d66db8b1e6a694c97af3c245fca0ec09bc5dbf6e24a77ebe044515c97ad4e" Feb 17 13:45:18 crc kubenswrapper[4955]: E0217 13:45:18.224193 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:45:29 crc kubenswrapper[4955]: I0217 13:45:29.222503 4955 scope.go:117] "RemoveContainer" containerID="e81d66db8b1e6a694c97af3c245fca0ec09bc5dbf6e24a77ebe044515c97ad4e" Feb 17 13:45:29 crc kubenswrapper[4955]: E0217 13:45:29.223564 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:45:41 crc kubenswrapper[4955]: I0217 13:45:41.223533 4955 scope.go:117] "RemoveContainer" containerID="e81d66db8b1e6a694c97af3c245fca0ec09bc5dbf6e24a77ebe044515c97ad4e" Feb 17 13:45:41 crc kubenswrapper[4955]: E0217 13:45:41.224307 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:45:51 crc kubenswrapper[4955]: I0217 13:45:51.184675 4955 scope.go:117] "RemoveContainer" containerID="a32124e8d3b973e4e5f3f6e8a4c8205bc70e8fba11e9335172e8be548ef9d4cb" Feb 17 13:45:54 crc kubenswrapper[4955]: I0217 13:45:54.223869 4955 scope.go:117] "RemoveContainer" containerID="e81d66db8b1e6a694c97af3c245fca0ec09bc5dbf6e24a77ebe044515c97ad4e" Feb 17 13:45:54 crc kubenswrapper[4955]: E0217 13:45:54.224505 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:46:04 crc kubenswrapper[4955]: I0217 13:46:04.187214 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-npxvr"] Feb 17 13:46:04 crc kubenswrapper[4955]: E0217 13:46:04.188253 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3861a08f-b7f7-43b7-80b0-80f0f9289de9" containerName="collect-profiles" Feb 17 13:46:04 crc kubenswrapper[4955]: I0217 13:46:04.188271 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="3861a08f-b7f7-43b7-80b0-80f0f9289de9" containerName="collect-profiles" Feb 17 13:46:04 crc kubenswrapper[4955]: I0217 13:46:04.188484 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="3861a08f-b7f7-43b7-80b0-80f0f9289de9" containerName="collect-profiles" Feb 17 13:46:04 crc kubenswrapper[4955]: I0217 13:46:04.190011 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-npxvr" Feb 17 13:46:04 crc kubenswrapper[4955]: I0217 13:46:04.202307 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-npxvr"] Feb 17 13:46:04 crc kubenswrapper[4955]: I0217 13:46:04.291814 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c4213ab-7ca9-44d3-b22e-fe40f14d44a0-catalog-content\") pod \"redhat-operators-npxvr\" (UID: \"4c4213ab-7ca9-44d3-b22e-fe40f14d44a0\") " pod="openshift-marketplace/redhat-operators-npxvr" Feb 17 13:46:04 crc kubenswrapper[4955]: I0217 13:46:04.291888 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c4213ab-7ca9-44d3-b22e-fe40f14d44a0-utilities\") pod \"redhat-operators-npxvr\" (UID: \"4c4213ab-7ca9-44d3-b22e-fe40f14d44a0\") " pod="openshift-marketplace/redhat-operators-npxvr" Feb 17 13:46:04 crc kubenswrapper[4955]: I0217 13:46:04.291969 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhwlk\" (UniqueName: \"kubernetes.io/projected/4c4213ab-7ca9-44d3-b22e-fe40f14d44a0-kube-api-access-jhwlk\") pod \"redhat-operators-npxvr\" (UID: \"4c4213ab-7ca9-44d3-b22e-fe40f14d44a0\") " pod="openshift-marketplace/redhat-operators-npxvr" Feb 17 13:46:04 crc kubenswrapper[4955]: I0217 13:46:04.393658 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c4213ab-7ca9-44d3-b22e-fe40f14d44a0-catalog-content\") pod \"redhat-operators-npxvr\" (UID: \"4c4213ab-7ca9-44d3-b22e-fe40f14d44a0\") " pod="openshift-marketplace/redhat-operators-npxvr" Feb 17 13:46:04 crc kubenswrapper[4955]: I0217 13:46:04.393719 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c4213ab-7ca9-44d3-b22e-fe40f14d44a0-utilities\") pod \"redhat-operators-npxvr\" (UID: \"4c4213ab-7ca9-44d3-b22e-fe40f14d44a0\") " pod="openshift-marketplace/redhat-operators-npxvr" Feb 17 13:46:04 crc kubenswrapper[4955]: I0217 13:46:04.393825 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhwlk\" (UniqueName: \"kubernetes.io/projected/4c4213ab-7ca9-44d3-b22e-fe40f14d44a0-kube-api-access-jhwlk\") pod \"redhat-operators-npxvr\" (UID: \"4c4213ab-7ca9-44d3-b22e-fe40f14d44a0\") " pod="openshift-marketplace/redhat-operators-npxvr" Feb 17 13:46:04 crc kubenswrapper[4955]: I0217 13:46:04.394245 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c4213ab-7ca9-44d3-b22e-fe40f14d44a0-catalog-content\") pod \"redhat-operators-npxvr\" (UID: \"4c4213ab-7ca9-44d3-b22e-fe40f14d44a0\") " pod="openshift-marketplace/redhat-operators-npxvr" Feb 17 13:46:04 crc kubenswrapper[4955]: I0217 13:46:04.394556 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c4213ab-7ca9-44d3-b22e-fe40f14d44a0-utilities\") pod \"redhat-operators-npxvr\" (UID: \"4c4213ab-7ca9-44d3-b22e-fe40f14d44a0\") " pod="openshift-marketplace/redhat-operators-npxvr" Feb 17 13:46:04 crc kubenswrapper[4955]: I0217 13:46:04.429033 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhwlk\" (UniqueName: \"kubernetes.io/projected/4c4213ab-7ca9-44d3-b22e-fe40f14d44a0-kube-api-access-jhwlk\") pod \"redhat-operators-npxvr\" (UID: \"4c4213ab-7ca9-44d3-b22e-fe40f14d44a0\") " pod="openshift-marketplace/redhat-operators-npxvr" Feb 17 13:46:04 crc kubenswrapper[4955]: I0217 13:46:04.515400 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-npxvr" Feb 17 13:46:05 crc kubenswrapper[4955]: I0217 13:46:05.003706 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-npxvr"] Feb 17 13:46:05 crc kubenswrapper[4955]: I0217 13:46:05.638474 4955 generic.go:334] "Generic (PLEG): container finished" podID="4c4213ab-7ca9-44d3-b22e-fe40f14d44a0" containerID="45ea8f6d947ea3df797f425e2d89d2a6b30cbc965586de062f15d092207ba234" exitCode=0 Feb 17 13:46:05 crc kubenswrapper[4955]: I0217 13:46:05.638589 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-npxvr" event={"ID":"4c4213ab-7ca9-44d3-b22e-fe40f14d44a0","Type":"ContainerDied","Data":"45ea8f6d947ea3df797f425e2d89d2a6b30cbc965586de062f15d092207ba234"} Feb 17 13:46:05 crc kubenswrapper[4955]: I0217 13:46:05.638744 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-npxvr" event={"ID":"4c4213ab-7ca9-44d3-b22e-fe40f14d44a0","Type":"ContainerStarted","Data":"7b52f9e833360d393b68cf9f1b19f21748e67f545a02b452e912050f19136089"} Feb 17 13:46:06 crc kubenswrapper[4955]: I0217 13:46:06.650735 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-npxvr" event={"ID":"4c4213ab-7ca9-44d3-b22e-fe40f14d44a0","Type":"ContainerStarted","Data":"99294a59e60f9aff37e39fd1b8f2768acf5d2a5ccbb75caf70e4c0049a34145a"} Feb 17 13:46:07 crc kubenswrapper[4955]: I0217 13:46:07.223011 4955 scope.go:117] "RemoveContainer" containerID="e81d66db8b1e6a694c97af3c245fca0ec09bc5dbf6e24a77ebe044515c97ad4e" Feb 17 13:46:07 crc kubenswrapper[4955]: E0217 13:46:07.223622 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:46:07 crc kubenswrapper[4955]: I0217 13:46:07.660916 4955 generic.go:334] "Generic (PLEG): container finished" podID="4c4213ab-7ca9-44d3-b22e-fe40f14d44a0" containerID="99294a59e60f9aff37e39fd1b8f2768acf5d2a5ccbb75caf70e4c0049a34145a" exitCode=0 Feb 17 13:46:07 crc kubenswrapper[4955]: I0217 13:46:07.660974 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-npxvr" event={"ID":"4c4213ab-7ca9-44d3-b22e-fe40f14d44a0","Type":"ContainerDied","Data":"99294a59e60f9aff37e39fd1b8f2768acf5d2a5ccbb75caf70e4c0049a34145a"} Feb 17 13:46:08 crc kubenswrapper[4955]: I0217 13:46:08.671590 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-npxvr" event={"ID":"4c4213ab-7ca9-44d3-b22e-fe40f14d44a0","Type":"ContainerStarted","Data":"38414ea7985afe2eff2c7f8cc5cecb77b02444b61fa680ad62882d3521322ee2"} Feb 17 13:46:08 crc kubenswrapper[4955]: I0217 13:46:08.691854 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-npxvr" podStartSLOduration=2.263129721 podStartE2EDuration="4.691835649s" podCreationTimestamp="2026-02-17 13:46:04 +0000 UTC" firstStartedPulling="2026-02-17 13:46:05.640282175 +0000 UTC m=+2504.163011718" lastFinishedPulling="2026-02-17 13:46:08.068988103 +0000 UTC m=+2506.591717646" observedRunningTime="2026-02-17 13:46:08.688978178 +0000 UTC m=+2507.211707731" watchObservedRunningTime="2026-02-17 13:46:08.691835649 +0000 UTC m=+2507.214565192" Feb 17 13:46:14 crc kubenswrapper[4955]: I0217 13:46:14.515799 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-npxvr" Feb 17 13:46:14 crc kubenswrapper[4955]: I0217 13:46:14.516375 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-npxvr" Feb 17 13:46:14 crc kubenswrapper[4955]: I0217 13:46:14.568117 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-npxvr" Feb 17 13:46:14 crc kubenswrapper[4955]: I0217 13:46:14.758635 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-npxvr" Feb 17 13:46:14 crc kubenswrapper[4955]: I0217 13:46:14.805454 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-npxvr"] Feb 17 13:46:16 crc kubenswrapper[4955]: I0217 13:46:16.730200 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-npxvr" podUID="4c4213ab-7ca9-44d3-b22e-fe40f14d44a0" containerName="registry-server" containerID="cri-o://38414ea7985afe2eff2c7f8cc5cecb77b02444b61fa680ad62882d3521322ee2" gracePeriod=2 Feb 17 13:46:17 crc kubenswrapper[4955]: I0217 13:46:17.168459 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-npxvr" Feb 17 13:46:17 crc kubenswrapper[4955]: I0217 13:46:17.320619 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c4213ab-7ca9-44d3-b22e-fe40f14d44a0-catalog-content\") pod \"4c4213ab-7ca9-44d3-b22e-fe40f14d44a0\" (UID: \"4c4213ab-7ca9-44d3-b22e-fe40f14d44a0\") " Feb 17 13:46:17 crc kubenswrapper[4955]: I0217 13:46:17.321963 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c4213ab-7ca9-44d3-b22e-fe40f14d44a0-utilities\") pod \"4c4213ab-7ca9-44d3-b22e-fe40f14d44a0\" (UID: \"4c4213ab-7ca9-44d3-b22e-fe40f14d44a0\") " Feb 17 13:46:17 crc kubenswrapper[4955]: I0217 13:46:17.322215 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhwlk\" (UniqueName: \"kubernetes.io/projected/4c4213ab-7ca9-44d3-b22e-fe40f14d44a0-kube-api-access-jhwlk\") pod \"4c4213ab-7ca9-44d3-b22e-fe40f14d44a0\" (UID: \"4c4213ab-7ca9-44d3-b22e-fe40f14d44a0\") " Feb 17 13:46:17 crc kubenswrapper[4955]: I0217 13:46:17.322962 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c4213ab-7ca9-44d3-b22e-fe40f14d44a0-utilities" (OuterVolumeSpecName: "utilities") pod "4c4213ab-7ca9-44d3-b22e-fe40f14d44a0" (UID: "4c4213ab-7ca9-44d3-b22e-fe40f14d44a0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:46:17 crc kubenswrapper[4955]: I0217 13:46:17.327026 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c4213ab-7ca9-44d3-b22e-fe40f14d44a0-kube-api-access-jhwlk" (OuterVolumeSpecName: "kube-api-access-jhwlk") pod "4c4213ab-7ca9-44d3-b22e-fe40f14d44a0" (UID: "4c4213ab-7ca9-44d3-b22e-fe40f14d44a0"). InnerVolumeSpecName "kube-api-access-jhwlk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:46:17 crc kubenswrapper[4955]: I0217 13:46:17.425444 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhwlk\" (UniqueName: \"kubernetes.io/projected/4c4213ab-7ca9-44d3-b22e-fe40f14d44a0-kube-api-access-jhwlk\") on node \"crc\" DevicePath \"\"" Feb 17 13:46:17 crc kubenswrapper[4955]: I0217 13:46:17.425494 4955 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c4213ab-7ca9-44d3-b22e-fe40f14d44a0-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 13:46:17 crc kubenswrapper[4955]: I0217 13:46:17.445944 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c4213ab-7ca9-44d3-b22e-fe40f14d44a0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4c4213ab-7ca9-44d3-b22e-fe40f14d44a0" (UID: "4c4213ab-7ca9-44d3-b22e-fe40f14d44a0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:46:17 crc kubenswrapper[4955]: I0217 13:46:17.527196 4955 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c4213ab-7ca9-44d3-b22e-fe40f14d44a0-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 13:46:17 crc kubenswrapper[4955]: I0217 13:46:17.739622 4955 generic.go:334] "Generic (PLEG): container finished" podID="4c4213ab-7ca9-44d3-b22e-fe40f14d44a0" containerID="38414ea7985afe2eff2c7f8cc5cecb77b02444b61fa680ad62882d3521322ee2" exitCode=0 Feb 17 13:46:17 crc kubenswrapper[4955]: I0217 13:46:17.739677 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-npxvr" event={"ID":"4c4213ab-7ca9-44d3-b22e-fe40f14d44a0","Type":"ContainerDied","Data":"38414ea7985afe2eff2c7f8cc5cecb77b02444b61fa680ad62882d3521322ee2"} Feb 17 13:46:17 crc kubenswrapper[4955]: I0217 13:46:17.739733 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-npxvr" event={"ID":"4c4213ab-7ca9-44d3-b22e-fe40f14d44a0","Type":"ContainerDied","Data":"7b52f9e833360d393b68cf9f1b19f21748e67f545a02b452e912050f19136089"} Feb 17 13:46:17 crc kubenswrapper[4955]: I0217 13:46:17.739757 4955 scope.go:117] "RemoveContainer" containerID="38414ea7985afe2eff2c7f8cc5cecb77b02444b61fa680ad62882d3521322ee2" Feb 17 13:46:17 crc kubenswrapper[4955]: I0217 13:46:17.740652 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-npxvr" Feb 17 13:46:17 crc kubenswrapper[4955]: I0217 13:46:17.757656 4955 scope.go:117] "RemoveContainer" containerID="99294a59e60f9aff37e39fd1b8f2768acf5d2a5ccbb75caf70e4c0049a34145a" Feb 17 13:46:17 crc kubenswrapper[4955]: I0217 13:46:17.786858 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-npxvr"] Feb 17 13:46:17 crc kubenswrapper[4955]: I0217 13:46:17.796611 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-npxvr"] Feb 17 13:46:17 crc kubenswrapper[4955]: I0217 13:46:17.805235 4955 scope.go:117] "RemoveContainer" containerID="45ea8f6d947ea3df797f425e2d89d2a6b30cbc965586de062f15d092207ba234" Feb 17 13:46:17 crc kubenswrapper[4955]: I0217 13:46:17.828899 4955 scope.go:117] "RemoveContainer" containerID="38414ea7985afe2eff2c7f8cc5cecb77b02444b61fa680ad62882d3521322ee2" Feb 17 13:46:17 crc kubenswrapper[4955]: E0217 13:46:17.829387 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38414ea7985afe2eff2c7f8cc5cecb77b02444b61fa680ad62882d3521322ee2\": container with ID starting with 38414ea7985afe2eff2c7f8cc5cecb77b02444b61fa680ad62882d3521322ee2 not found: ID does not exist" containerID="38414ea7985afe2eff2c7f8cc5cecb77b02444b61fa680ad62882d3521322ee2" Feb 17 13:46:17 crc kubenswrapper[4955]: I0217 13:46:17.829447 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38414ea7985afe2eff2c7f8cc5cecb77b02444b61fa680ad62882d3521322ee2"} err="failed to get container status \"38414ea7985afe2eff2c7f8cc5cecb77b02444b61fa680ad62882d3521322ee2\": rpc error: code = NotFound desc = could not find container \"38414ea7985afe2eff2c7f8cc5cecb77b02444b61fa680ad62882d3521322ee2\": container with ID starting with 38414ea7985afe2eff2c7f8cc5cecb77b02444b61fa680ad62882d3521322ee2 not found: ID does not exist" Feb 17 13:46:17 crc kubenswrapper[4955]: I0217 13:46:17.829472 4955 scope.go:117] "RemoveContainer" containerID="99294a59e60f9aff37e39fd1b8f2768acf5d2a5ccbb75caf70e4c0049a34145a" Feb 17 13:46:17 crc kubenswrapper[4955]: E0217 13:46:17.829725 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99294a59e60f9aff37e39fd1b8f2768acf5d2a5ccbb75caf70e4c0049a34145a\": container with ID starting with 99294a59e60f9aff37e39fd1b8f2768acf5d2a5ccbb75caf70e4c0049a34145a not found: ID does not exist" containerID="99294a59e60f9aff37e39fd1b8f2768acf5d2a5ccbb75caf70e4c0049a34145a" Feb 17 13:46:17 crc kubenswrapper[4955]: I0217 13:46:17.829827 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99294a59e60f9aff37e39fd1b8f2768acf5d2a5ccbb75caf70e4c0049a34145a"} err="failed to get container status \"99294a59e60f9aff37e39fd1b8f2768acf5d2a5ccbb75caf70e4c0049a34145a\": rpc error: code = NotFound desc = could not find container \"99294a59e60f9aff37e39fd1b8f2768acf5d2a5ccbb75caf70e4c0049a34145a\": container with ID starting with 99294a59e60f9aff37e39fd1b8f2768acf5d2a5ccbb75caf70e4c0049a34145a not found: ID does not exist" Feb 17 13:46:17 crc kubenswrapper[4955]: I0217 13:46:17.829934 4955 scope.go:117] "RemoveContainer" containerID="45ea8f6d947ea3df797f425e2d89d2a6b30cbc965586de062f15d092207ba234" Feb 17 13:46:17 crc kubenswrapper[4955]: E0217 13:46:17.830310 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45ea8f6d947ea3df797f425e2d89d2a6b30cbc965586de062f15d092207ba234\": container with ID starting with 45ea8f6d947ea3df797f425e2d89d2a6b30cbc965586de062f15d092207ba234 not found: ID does not exist" containerID="45ea8f6d947ea3df797f425e2d89d2a6b30cbc965586de062f15d092207ba234" Feb 17 13:46:17 crc kubenswrapper[4955]: I0217 13:46:17.830332 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45ea8f6d947ea3df797f425e2d89d2a6b30cbc965586de062f15d092207ba234"} err="failed to get container status \"45ea8f6d947ea3df797f425e2d89d2a6b30cbc965586de062f15d092207ba234\": rpc error: code = NotFound desc = could not find container \"45ea8f6d947ea3df797f425e2d89d2a6b30cbc965586de062f15d092207ba234\": container with ID starting with 45ea8f6d947ea3df797f425e2d89d2a6b30cbc965586de062f15d092207ba234 not found: ID does not exist" Feb 17 13:46:18 crc kubenswrapper[4955]: I0217 13:46:18.235610 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c4213ab-7ca9-44d3-b22e-fe40f14d44a0" path="/var/lib/kubelet/pods/4c4213ab-7ca9-44d3-b22e-fe40f14d44a0/volumes" Feb 17 13:46:20 crc kubenswrapper[4955]: I0217 13:46:20.766234 4955 generic.go:334] "Generic (PLEG): container finished" podID="0acb7fb8-671d-4e90-8be4-f5f507308f11" containerID="ab524817b95eb90df2f8dfc9b797aba13ed2c4616293b5efba888e54d7c4ef43" exitCode=0 Feb 17 13:46:20 crc kubenswrapper[4955]: I0217 13:46:20.766344 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npz67" event={"ID":"0acb7fb8-671d-4e90-8be4-f5f507308f11","Type":"ContainerDied","Data":"ab524817b95eb90df2f8dfc9b797aba13ed2c4616293b5efba888e54d7c4ef43"} Feb 17 13:46:21 crc kubenswrapper[4955]: I0217 13:46:21.223838 4955 scope.go:117] "RemoveContainer" containerID="e81d66db8b1e6a694c97af3c245fca0ec09bc5dbf6e24a77ebe044515c97ad4e" Feb 17 13:46:21 crc kubenswrapper[4955]: E0217 13:46:21.224112 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:46:22 crc kubenswrapper[4955]: I0217 13:46:22.129905 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npz67" Feb 17 13:46:22 crc kubenswrapper[4955]: I0217 13:46:22.213336 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0acb7fb8-671d-4e90-8be4-f5f507308f11-telemetry-combined-ca-bundle\") pod \"0acb7fb8-671d-4e90-8be4-f5f507308f11\" (UID: \"0acb7fb8-671d-4e90-8be4-f5f507308f11\") " Feb 17 13:46:22 crc kubenswrapper[4955]: I0217 13:46:22.213440 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0acb7fb8-671d-4e90-8be4-f5f507308f11-inventory\") pod \"0acb7fb8-671d-4e90-8be4-f5f507308f11\" (UID: \"0acb7fb8-671d-4e90-8be4-f5f507308f11\") " Feb 17 13:46:22 crc kubenswrapper[4955]: I0217 13:46:22.213501 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/0acb7fb8-671d-4e90-8be4-f5f507308f11-ceilometer-compute-config-data-1\") pod \"0acb7fb8-671d-4e90-8be4-f5f507308f11\" (UID: \"0acb7fb8-671d-4e90-8be4-f5f507308f11\") " Feb 17 13:46:22 crc kubenswrapper[4955]: I0217 13:46:22.213535 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gnj7f\" (UniqueName: \"kubernetes.io/projected/0acb7fb8-671d-4e90-8be4-f5f507308f11-kube-api-access-gnj7f\") pod \"0acb7fb8-671d-4e90-8be4-f5f507308f11\" (UID: \"0acb7fb8-671d-4e90-8be4-f5f507308f11\") " Feb 17 13:46:22 crc kubenswrapper[4955]: I0217 13:46:22.213580 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/0acb7fb8-671d-4e90-8be4-f5f507308f11-ceilometer-compute-config-data-0\") pod \"0acb7fb8-671d-4e90-8be4-f5f507308f11\" (UID: \"0acb7fb8-671d-4e90-8be4-f5f507308f11\") " Feb 17 13:46:22 crc kubenswrapper[4955]: I0217 13:46:22.213639 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0acb7fb8-671d-4e90-8be4-f5f507308f11-ssh-key-openstack-edpm-ipam\") pod \"0acb7fb8-671d-4e90-8be4-f5f507308f11\" (UID: \"0acb7fb8-671d-4e90-8be4-f5f507308f11\") " Feb 17 13:46:22 crc kubenswrapper[4955]: I0217 13:46:22.213671 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/0acb7fb8-671d-4e90-8be4-f5f507308f11-ceilometer-compute-config-data-2\") pod \"0acb7fb8-671d-4e90-8be4-f5f507308f11\" (UID: \"0acb7fb8-671d-4e90-8be4-f5f507308f11\") " Feb 17 13:46:22 crc kubenswrapper[4955]: I0217 13:46:22.218778 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0acb7fb8-671d-4e90-8be4-f5f507308f11-kube-api-access-gnj7f" (OuterVolumeSpecName: "kube-api-access-gnj7f") pod "0acb7fb8-671d-4e90-8be4-f5f507308f11" (UID: "0acb7fb8-671d-4e90-8be4-f5f507308f11"). InnerVolumeSpecName "kube-api-access-gnj7f". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:46:22 crc kubenswrapper[4955]: I0217 13:46:22.218903 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0acb7fb8-671d-4e90-8be4-f5f507308f11-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "0acb7fb8-671d-4e90-8be4-f5f507308f11" (UID: "0acb7fb8-671d-4e90-8be4-f5f507308f11"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:46:22 crc kubenswrapper[4955]: I0217 13:46:22.248002 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0acb7fb8-671d-4e90-8be4-f5f507308f11-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "0acb7fb8-671d-4e90-8be4-f5f507308f11" (UID: "0acb7fb8-671d-4e90-8be4-f5f507308f11"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:46:22 crc kubenswrapper[4955]: I0217 13:46:22.251125 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0acb7fb8-671d-4e90-8be4-f5f507308f11-inventory" (OuterVolumeSpecName: "inventory") pod "0acb7fb8-671d-4e90-8be4-f5f507308f11" (UID: "0acb7fb8-671d-4e90-8be4-f5f507308f11"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:46:22 crc kubenswrapper[4955]: I0217 13:46:22.251651 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0acb7fb8-671d-4e90-8be4-f5f507308f11-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "0acb7fb8-671d-4e90-8be4-f5f507308f11" (UID: "0acb7fb8-671d-4e90-8be4-f5f507308f11"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:46:22 crc kubenswrapper[4955]: I0217 13:46:22.254892 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0acb7fb8-671d-4e90-8be4-f5f507308f11-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "0acb7fb8-671d-4e90-8be4-f5f507308f11" (UID: "0acb7fb8-671d-4e90-8be4-f5f507308f11"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:46:22 crc kubenswrapper[4955]: I0217 13:46:22.265406 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0acb7fb8-671d-4e90-8be4-f5f507308f11-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "0acb7fb8-671d-4e90-8be4-f5f507308f11" (UID: "0acb7fb8-671d-4e90-8be4-f5f507308f11"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:46:22 crc kubenswrapper[4955]: I0217 13:46:22.316161 4955 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/0acb7fb8-671d-4e90-8be4-f5f507308f11-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Feb 17 13:46:22 crc kubenswrapper[4955]: I0217 13:46:22.316187 4955 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0acb7fb8-671d-4e90-8be4-f5f507308f11-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 13:46:22 crc kubenswrapper[4955]: I0217 13:46:22.316219 4955 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0acb7fb8-671d-4e90-8be4-f5f507308f11-inventory\") on node \"crc\" DevicePath \"\"" Feb 17 13:46:22 crc kubenswrapper[4955]: I0217 13:46:22.316229 4955 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/0acb7fb8-671d-4e90-8be4-f5f507308f11-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Feb 17 13:46:22 crc kubenswrapper[4955]: I0217 13:46:22.316239 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gnj7f\" (UniqueName: \"kubernetes.io/projected/0acb7fb8-671d-4e90-8be4-f5f507308f11-kube-api-access-gnj7f\") on node \"crc\" DevicePath \"\"" Feb 17 13:46:22 crc kubenswrapper[4955]: I0217 13:46:22.316248 4955 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/0acb7fb8-671d-4e90-8be4-f5f507308f11-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Feb 17 13:46:22 crc kubenswrapper[4955]: I0217 13:46:22.316258 4955 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0acb7fb8-671d-4e90-8be4-f5f507308f11-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 17 13:46:22 crc kubenswrapper[4955]: I0217 13:46:22.783276 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npz67" event={"ID":"0acb7fb8-671d-4e90-8be4-f5f507308f11","Type":"ContainerDied","Data":"1da60edad11e2f0dfcd9de5c646859cd8cf1a720a01e6b2c70a5a7263ff73bc9"} Feb 17 13:46:22 crc kubenswrapper[4955]: I0217 13:46:22.783314 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1da60edad11e2f0dfcd9de5c646859cd8cf1a720a01e6b2c70a5a7263ff73bc9" Feb 17 13:46:22 crc kubenswrapper[4955]: I0217 13:46:22.783384 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npz67" Feb 17 13:46:35 crc kubenswrapper[4955]: I0217 13:46:35.222831 4955 scope.go:117] "RemoveContainer" containerID="e81d66db8b1e6a694c97af3c245fca0ec09bc5dbf6e24a77ebe044515c97ad4e" Feb 17 13:46:35 crc kubenswrapper[4955]: I0217 13:46:35.888589 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" event={"ID":"5bf36887-2a2b-4193-abdb-a15beaf46300","Type":"ContainerStarted","Data":"e51d645860a15fd43cd83c6bac8255e568dfbde7c066e876b04046df9a6623fc"} Feb 17 13:47:16 crc kubenswrapper[4955]: I0217 13:47:16.961231 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Feb 17 13:47:16 crc kubenswrapper[4955]: E0217 13:47:16.962179 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c4213ab-7ca9-44d3-b22e-fe40f14d44a0" containerName="registry-server" Feb 17 13:47:16 crc kubenswrapper[4955]: I0217 13:47:16.962196 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c4213ab-7ca9-44d3-b22e-fe40f14d44a0" containerName="registry-server" Feb 17 13:47:16 crc kubenswrapper[4955]: E0217 13:47:16.962232 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c4213ab-7ca9-44d3-b22e-fe40f14d44a0" containerName="extract-content" Feb 17 13:47:16 crc kubenswrapper[4955]: I0217 13:47:16.962241 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c4213ab-7ca9-44d3-b22e-fe40f14d44a0" containerName="extract-content" Feb 17 13:47:16 crc kubenswrapper[4955]: E0217 13:47:16.962261 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0acb7fb8-671d-4e90-8be4-f5f507308f11" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Feb 17 13:47:16 crc kubenswrapper[4955]: I0217 13:47:16.962271 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="0acb7fb8-671d-4e90-8be4-f5f507308f11" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Feb 17 13:47:16 crc kubenswrapper[4955]: E0217 13:47:16.962288 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c4213ab-7ca9-44d3-b22e-fe40f14d44a0" containerName="extract-utilities" Feb 17 13:47:16 crc kubenswrapper[4955]: I0217 13:47:16.962295 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c4213ab-7ca9-44d3-b22e-fe40f14d44a0" containerName="extract-utilities" Feb 17 13:47:16 crc kubenswrapper[4955]: I0217 13:47:16.962490 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c4213ab-7ca9-44d3-b22e-fe40f14d44a0" containerName="registry-server" Feb 17 13:47:16 crc kubenswrapper[4955]: I0217 13:47:16.962510 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="0acb7fb8-671d-4e90-8be4-f5f507308f11" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Feb 17 13:47:16 crc kubenswrapper[4955]: I0217 13:47:16.963281 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 17 13:47:16 crc kubenswrapper[4955]: I0217 13:47:16.966027 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Feb 17 13:47:16 crc kubenswrapper[4955]: I0217 13:47:16.966080 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-kwzmd" Feb 17 13:47:16 crc kubenswrapper[4955]: I0217 13:47:16.966032 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Feb 17 13:47:16 crc kubenswrapper[4955]: I0217 13:47:16.966449 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Feb 17 13:47:16 crc kubenswrapper[4955]: I0217 13:47:16.977146 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Feb 17 13:47:17 crc kubenswrapper[4955]: I0217 13:47:17.074742 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/de1d0057-d623-4185-b31a-572dfaabcce2-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"de1d0057-d623-4185-b31a-572dfaabcce2\") " pod="openstack/tempest-tests-tempest" Feb 17 13:47:17 crc kubenswrapper[4955]: I0217 13:47:17.074820 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/de1d0057-d623-4185-b31a-572dfaabcce2-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"de1d0057-d623-4185-b31a-572dfaabcce2\") " pod="openstack/tempest-tests-tempest" Feb 17 13:47:17 crc kubenswrapper[4955]: I0217 13:47:17.074908 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"de1d0057-d623-4185-b31a-572dfaabcce2\") " pod="openstack/tempest-tests-tempest" Feb 17 13:47:17 crc kubenswrapper[4955]: I0217 13:47:17.074940 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vt7w\" (UniqueName: \"kubernetes.io/projected/de1d0057-d623-4185-b31a-572dfaabcce2-kube-api-access-2vt7w\") pod \"tempest-tests-tempest\" (UID: \"de1d0057-d623-4185-b31a-572dfaabcce2\") " pod="openstack/tempest-tests-tempest" Feb 17 13:47:17 crc kubenswrapper[4955]: I0217 13:47:17.075006 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de1d0057-d623-4185-b31a-572dfaabcce2-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"de1d0057-d623-4185-b31a-572dfaabcce2\") " pod="openstack/tempest-tests-tempest" Feb 17 13:47:17 crc kubenswrapper[4955]: I0217 13:47:17.075039 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/de1d0057-d623-4185-b31a-572dfaabcce2-config-data\") pod \"tempest-tests-tempest\" (UID: \"de1d0057-d623-4185-b31a-572dfaabcce2\") " pod="openstack/tempest-tests-tempest" Feb 17 13:47:17 crc kubenswrapper[4955]: I0217 13:47:17.075082 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/de1d0057-d623-4185-b31a-572dfaabcce2-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"de1d0057-d623-4185-b31a-572dfaabcce2\") " pod="openstack/tempest-tests-tempest" Feb 17 13:47:17 crc kubenswrapper[4955]: I0217 13:47:17.075124 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/de1d0057-d623-4185-b31a-572dfaabcce2-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"de1d0057-d623-4185-b31a-572dfaabcce2\") " pod="openstack/tempest-tests-tempest" Feb 17 13:47:17 crc kubenswrapper[4955]: I0217 13:47:17.075170 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/de1d0057-d623-4185-b31a-572dfaabcce2-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"de1d0057-d623-4185-b31a-572dfaabcce2\") " pod="openstack/tempest-tests-tempest" Feb 17 13:47:17 crc kubenswrapper[4955]: I0217 13:47:17.176976 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/de1d0057-d623-4185-b31a-572dfaabcce2-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"de1d0057-d623-4185-b31a-572dfaabcce2\") " pod="openstack/tempest-tests-tempest" Feb 17 13:47:17 crc kubenswrapper[4955]: I0217 13:47:17.177046 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/de1d0057-d623-4185-b31a-572dfaabcce2-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"de1d0057-d623-4185-b31a-572dfaabcce2\") " pod="openstack/tempest-tests-tempest" Feb 17 13:47:17 crc kubenswrapper[4955]: I0217 13:47:17.177103 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/de1d0057-d623-4185-b31a-572dfaabcce2-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"de1d0057-d623-4185-b31a-572dfaabcce2\") " pod="openstack/tempest-tests-tempest" Feb 17 13:47:17 crc kubenswrapper[4955]: I0217 13:47:17.177124 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/de1d0057-d623-4185-b31a-572dfaabcce2-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"de1d0057-d623-4185-b31a-572dfaabcce2\") " pod="openstack/tempest-tests-tempest" Feb 17 13:47:17 crc kubenswrapper[4955]: I0217 13:47:17.177157 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"de1d0057-d623-4185-b31a-572dfaabcce2\") " pod="openstack/tempest-tests-tempest" Feb 17 13:47:17 crc kubenswrapper[4955]: I0217 13:47:17.177178 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vt7w\" (UniqueName: \"kubernetes.io/projected/de1d0057-d623-4185-b31a-572dfaabcce2-kube-api-access-2vt7w\") pod \"tempest-tests-tempest\" (UID: \"de1d0057-d623-4185-b31a-572dfaabcce2\") " pod="openstack/tempest-tests-tempest" Feb 17 13:47:17 crc kubenswrapper[4955]: I0217 13:47:17.177223 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de1d0057-d623-4185-b31a-572dfaabcce2-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"de1d0057-d623-4185-b31a-572dfaabcce2\") " pod="openstack/tempest-tests-tempest" Feb 17 13:47:17 crc kubenswrapper[4955]: I0217 13:47:17.177251 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/de1d0057-d623-4185-b31a-572dfaabcce2-config-data\") pod \"tempest-tests-tempest\" (UID: \"de1d0057-d623-4185-b31a-572dfaabcce2\") " pod="openstack/tempest-tests-tempest" Feb 17 13:47:17 crc kubenswrapper[4955]: I0217 13:47:17.177277 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/de1d0057-d623-4185-b31a-572dfaabcce2-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"de1d0057-d623-4185-b31a-572dfaabcce2\") " pod="openstack/tempest-tests-tempest" Feb 17 13:47:17 crc kubenswrapper[4955]: I0217 13:47:17.177547 4955 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"de1d0057-d623-4185-b31a-572dfaabcce2\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/tempest-tests-tempest" Feb 17 13:47:17 crc kubenswrapper[4955]: I0217 13:47:17.177858 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/de1d0057-d623-4185-b31a-572dfaabcce2-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"de1d0057-d623-4185-b31a-572dfaabcce2\") " pod="openstack/tempest-tests-tempest" Feb 17 13:47:17 crc kubenswrapper[4955]: I0217 13:47:17.177962 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/de1d0057-d623-4185-b31a-572dfaabcce2-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"de1d0057-d623-4185-b31a-572dfaabcce2\") " pod="openstack/tempest-tests-tempest" Feb 17 13:47:17 crc kubenswrapper[4955]: I0217 13:47:17.178418 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/de1d0057-d623-4185-b31a-572dfaabcce2-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"de1d0057-d623-4185-b31a-572dfaabcce2\") " pod="openstack/tempest-tests-tempest" Feb 17 13:47:17 crc kubenswrapper[4955]: I0217 13:47:17.179026 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/de1d0057-d623-4185-b31a-572dfaabcce2-config-data\") pod \"tempest-tests-tempest\" (UID: \"de1d0057-d623-4185-b31a-572dfaabcce2\") " pod="openstack/tempest-tests-tempest" Feb 17 13:47:17 crc kubenswrapper[4955]: I0217 13:47:17.184114 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/de1d0057-d623-4185-b31a-572dfaabcce2-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"de1d0057-d623-4185-b31a-572dfaabcce2\") " pod="openstack/tempest-tests-tempest" Feb 17 13:47:17 crc kubenswrapper[4955]: I0217 13:47:17.184502 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/de1d0057-d623-4185-b31a-572dfaabcce2-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"de1d0057-d623-4185-b31a-572dfaabcce2\") " pod="openstack/tempest-tests-tempest" Feb 17 13:47:17 crc kubenswrapper[4955]: I0217 13:47:17.185609 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de1d0057-d623-4185-b31a-572dfaabcce2-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"de1d0057-d623-4185-b31a-572dfaabcce2\") " pod="openstack/tempest-tests-tempest" Feb 17 13:47:17 crc kubenswrapper[4955]: I0217 13:47:17.202333 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vt7w\" (UniqueName: \"kubernetes.io/projected/de1d0057-d623-4185-b31a-572dfaabcce2-kube-api-access-2vt7w\") pod \"tempest-tests-tempest\" (UID: \"de1d0057-d623-4185-b31a-572dfaabcce2\") " pod="openstack/tempest-tests-tempest" Feb 17 13:47:17 crc kubenswrapper[4955]: I0217 13:47:17.213911 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"de1d0057-d623-4185-b31a-572dfaabcce2\") " pod="openstack/tempest-tests-tempest" Feb 17 13:47:17 crc kubenswrapper[4955]: I0217 13:47:17.292002 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 17 13:47:17 crc kubenswrapper[4955]: I0217 13:47:17.798438 4955 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 17 13:47:17 crc kubenswrapper[4955]: I0217 13:47:17.801277 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Feb 17 13:47:18 crc kubenswrapper[4955]: I0217 13:47:18.309190 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"de1d0057-d623-4185-b31a-572dfaabcce2","Type":"ContainerStarted","Data":"1b6db8750e4fdbf7eb868f485323ebcabc6e6a0f3301eae8715e83791a8df991"} Feb 17 13:47:44 crc kubenswrapper[4955]: E0217 13:47:44.245622 4955 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Feb 17 13:47:44 crc kubenswrapper[4955]: E0217 13:47:44.247436 4955 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2vt7w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(de1d0057-d623-4185-b31a-572dfaabcce2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 17 13:47:44 crc kubenswrapper[4955]: E0217 13:47:44.248601 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="de1d0057-d623-4185-b31a-572dfaabcce2" Feb 17 13:47:44 crc kubenswrapper[4955]: E0217 13:47:44.560669 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="de1d0057-d623-4185-b31a-572dfaabcce2" Feb 17 13:48:00 crc kubenswrapper[4955]: I0217 13:48:00.649957 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Feb 17 13:48:02 crc kubenswrapper[4955]: I0217 13:48:02.694520 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"de1d0057-d623-4185-b31a-572dfaabcce2","Type":"ContainerStarted","Data":"bdbd1017b2f134a5c55399b939afe437fbca6626b6654182d4ac13d5dde9f701"} Feb 17 13:48:02 crc kubenswrapper[4955]: I0217 13:48:02.719469 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.870396246 podStartE2EDuration="47.719451503s" podCreationTimestamp="2026-02-17 13:47:15 +0000 UTC" firstStartedPulling="2026-02-17 13:47:17.798078415 +0000 UTC m=+2576.320807958" lastFinishedPulling="2026-02-17 13:48:00.647133672 +0000 UTC m=+2619.169863215" observedRunningTime="2026-02-17 13:48:02.709769177 +0000 UTC m=+2621.232498720" watchObservedRunningTime="2026-02-17 13:48:02.719451503 +0000 UTC m=+2621.242181046" Feb 17 13:48:32 crc kubenswrapper[4955]: I0217 13:48:32.372296 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nr6s5"] Feb 17 13:48:32 crc kubenswrapper[4955]: I0217 13:48:32.376292 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nr6s5" Feb 17 13:48:32 crc kubenswrapper[4955]: I0217 13:48:32.388900 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nr6s5"] Feb 17 13:48:32 crc kubenswrapper[4955]: I0217 13:48:32.509692 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b99347d9-5973-4f7e-8d8c-dad6df3e0c92-utilities\") pod \"redhat-marketplace-nr6s5\" (UID: \"b99347d9-5973-4f7e-8d8c-dad6df3e0c92\") " pod="openshift-marketplace/redhat-marketplace-nr6s5" Feb 17 13:48:32 crc kubenswrapper[4955]: I0217 13:48:32.509754 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b99347d9-5973-4f7e-8d8c-dad6df3e0c92-catalog-content\") pod \"redhat-marketplace-nr6s5\" (UID: \"b99347d9-5973-4f7e-8d8c-dad6df3e0c92\") " pod="openshift-marketplace/redhat-marketplace-nr6s5" Feb 17 13:48:32 crc kubenswrapper[4955]: I0217 13:48:32.509885 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fhmk\" (UniqueName: \"kubernetes.io/projected/b99347d9-5973-4f7e-8d8c-dad6df3e0c92-kube-api-access-5fhmk\") pod \"redhat-marketplace-nr6s5\" (UID: \"b99347d9-5973-4f7e-8d8c-dad6df3e0c92\") " pod="openshift-marketplace/redhat-marketplace-nr6s5" Feb 17 13:48:32 crc kubenswrapper[4955]: I0217 13:48:32.611281 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b99347d9-5973-4f7e-8d8c-dad6df3e0c92-utilities\") pod \"redhat-marketplace-nr6s5\" (UID: \"b99347d9-5973-4f7e-8d8c-dad6df3e0c92\") " pod="openshift-marketplace/redhat-marketplace-nr6s5" Feb 17 13:48:32 crc kubenswrapper[4955]: I0217 13:48:32.611350 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b99347d9-5973-4f7e-8d8c-dad6df3e0c92-catalog-content\") pod \"redhat-marketplace-nr6s5\" (UID: \"b99347d9-5973-4f7e-8d8c-dad6df3e0c92\") " pod="openshift-marketplace/redhat-marketplace-nr6s5" Feb 17 13:48:32 crc kubenswrapper[4955]: I0217 13:48:32.611449 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fhmk\" (UniqueName: \"kubernetes.io/projected/b99347d9-5973-4f7e-8d8c-dad6df3e0c92-kube-api-access-5fhmk\") pod \"redhat-marketplace-nr6s5\" (UID: \"b99347d9-5973-4f7e-8d8c-dad6df3e0c92\") " pod="openshift-marketplace/redhat-marketplace-nr6s5" Feb 17 13:48:32 crc kubenswrapper[4955]: I0217 13:48:32.612355 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b99347d9-5973-4f7e-8d8c-dad6df3e0c92-utilities\") pod \"redhat-marketplace-nr6s5\" (UID: \"b99347d9-5973-4f7e-8d8c-dad6df3e0c92\") " pod="openshift-marketplace/redhat-marketplace-nr6s5" Feb 17 13:48:32 crc kubenswrapper[4955]: I0217 13:48:32.612420 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b99347d9-5973-4f7e-8d8c-dad6df3e0c92-catalog-content\") pod \"redhat-marketplace-nr6s5\" (UID: \"b99347d9-5973-4f7e-8d8c-dad6df3e0c92\") " pod="openshift-marketplace/redhat-marketplace-nr6s5" Feb 17 13:48:32 crc kubenswrapper[4955]: I0217 13:48:32.631974 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fhmk\" (UniqueName: \"kubernetes.io/projected/b99347d9-5973-4f7e-8d8c-dad6df3e0c92-kube-api-access-5fhmk\") pod \"redhat-marketplace-nr6s5\" (UID: \"b99347d9-5973-4f7e-8d8c-dad6df3e0c92\") " pod="openshift-marketplace/redhat-marketplace-nr6s5" Feb 17 13:48:32 crc kubenswrapper[4955]: I0217 13:48:32.722191 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nr6s5" Feb 17 13:48:33 crc kubenswrapper[4955]: I0217 13:48:33.664289 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nr6s5"] Feb 17 13:48:33 crc kubenswrapper[4955]: I0217 13:48:33.962123 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nr6s5" event={"ID":"b99347d9-5973-4f7e-8d8c-dad6df3e0c92","Type":"ContainerStarted","Data":"9d4a3124e545eef12aa2ae869bbe177d19d8420053fe6530f75a15847cab3b92"} Feb 17 13:48:34 crc kubenswrapper[4955]: I0217 13:48:34.971685 4955 generic.go:334] "Generic (PLEG): container finished" podID="b99347d9-5973-4f7e-8d8c-dad6df3e0c92" containerID="fe90edb414fd019576803f6d48082a2db875cb1d1065f73f9137e018ae75436c" exitCode=0 Feb 17 13:48:34 crc kubenswrapper[4955]: I0217 13:48:34.971803 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nr6s5" event={"ID":"b99347d9-5973-4f7e-8d8c-dad6df3e0c92","Type":"ContainerDied","Data":"fe90edb414fd019576803f6d48082a2db875cb1d1065f73f9137e018ae75436c"} Feb 17 13:48:36 crc kubenswrapper[4955]: I0217 13:48:36.991755 4955 generic.go:334] "Generic (PLEG): container finished" podID="b99347d9-5973-4f7e-8d8c-dad6df3e0c92" containerID="66d650031e7378665b7b68cdf799e1d9bc14680c42f150232a883f8d4869ff3d" exitCode=0 Feb 17 13:48:36 crc kubenswrapper[4955]: I0217 13:48:36.991812 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nr6s5" event={"ID":"b99347d9-5973-4f7e-8d8c-dad6df3e0c92","Type":"ContainerDied","Data":"66d650031e7378665b7b68cdf799e1d9bc14680c42f150232a883f8d4869ff3d"} Feb 17 13:48:38 crc kubenswrapper[4955]: I0217 13:48:38.002949 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nr6s5" event={"ID":"b99347d9-5973-4f7e-8d8c-dad6df3e0c92","Type":"ContainerStarted","Data":"41721e38bd5094f57bf243115297e935c9339cc99e9ec67c29dfde875796cb19"} Feb 17 13:48:38 crc kubenswrapper[4955]: I0217 13:48:38.028539 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nr6s5" podStartSLOduration=3.538133949 podStartE2EDuration="6.028516123s" podCreationTimestamp="2026-02-17 13:48:32 +0000 UTC" firstStartedPulling="2026-02-17 13:48:34.975069505 +0000 UTC m=+2653.497799048" lastFinishedPulling="2026-02-17 13:48:37.465451679 +0000 UTC m=+2655.988181222" observedRunningTime="2026-02-17 13:48:38.020591087 +0000 UTC m=+2656.543320630" watchObservedRunningTime="2026-02-17 13:48:38.028516123 +0000 UTC m=+2656.551245666" Feb 17 13:48:42 crc kubenswrapper[4955]: I0217 13:48:42.722372 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nr6s5" Feb 17 13:48:42 crc kubenswrapper[4955]: I0217 13:48:42.723148 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nr6s5" Feb 17 13:48:42 crc kubenswrapper[4955]: I0217 13:48:42.767151 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nr6s5" Feb 17 13:48:43 crc kubenswrapper[4955]: I0217 13:48:43.090550 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nr6s5" Feb 17 13:48:43 crc kubenswrapper[4955]: I0217 13:48:43.134972 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nr6s5"] Feb 17 13:48:45 crc kubenswrapper[4955]: I0217 13:48:45.060356 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nr6s5" podUID="b99347d9-5973-4f7e-8d8c-dad6df3e0c92" containerName="registry-server" containerID="cri-o://41721e38bd5094f57bf243115297e935c9339cc99e9ec67c29dfde875796cb19" gracePeriod=2 Feb 17 13:48:45 crc kubenswrapper[4955]: I0217 13:48:45.536320 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nr6s5" Feb 17 13:48:45 crc kubenswrapper[4955]: I0217 13:48:45.704553 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5fhmk\" (UniqueName: \"kubernetes.io/projected/b99347d9-5973-4f7e-8d8c-dad6df3e0c92-kube-api-access-5fhmk\") pod \"b99347d9-5973-4f7e-8d8c-dad6df3e0c92\" (UID: \"b99347d9-5973-4f7e-8d8c-dad6df3e0c92\") " Feb 17 13:48:45 crc kubenswrapper[4955]: I0217 13:48:45.704662 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b99347d9-5973-4f7e-8d8c-dad6df3e0c92-catalog-content\") pod \"b99347d9-5973-4f7e-8d8c-dad6df3e0c92\" (UID: \"b99347d9-5973-4f7e-8d8c-dad6df3e0c92\") " Feb 17 13:48:45 crc kubenswrapper[4955]: I0217 13:48:45.704692 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b99347d9-5973-4f7e-8d8c-dad6df3e0c92-utilities\") pod \"b99347d9-5973-4f7e-8d8c-dad6df3e0c92\" (UID: \"b99347d9-5973-4f7e-8d8c-dad6df3e0c92\") " Feb 17 13:48:45 crc kubenswrapper[4955]: I0217 13:48:45.705605 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b99347d9-5973-4f7e-8d8c-dad6df3e0c92-utilities" (OuterVolumeSpecName: "utilities") pod "b99347d9-5973-4f7e-8d8c-dad6df3e0c92" (UID: "b99347d9-5973-4f7e-8d8c-dad6df3e0c92"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:48:45 crc kubenswrapper[4955]: I0217 13:48:45.710589 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b99347d9-5973-4f7e-8d8c-dad6df3e0c92-kube-api-access-5fhmk" (OuterVolumeSpecName: "kube-api-access-5fhmk") pod "b99347d9-5973-4f7e-8d8c-dad6df3e0c92" (UID: "b99347d9-5973-4f7e-8d8c-dad6df3e0c92"). InnerVolumeSpecName "kube-api-access-5fhmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:48:45 crc kubenswrapper[4955]: I0217 13:48:45.732947 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b99347d9-5973-4f7e-8d8c-dad6df3e0c92-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b99347d9-5973-4f7e-8d8c-dad6df3e0c92" (UID: "b99347d9-5973-4f7e-8d8c-dad6df3e0c92"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:48:45 crc kubenswrapper[4955]: I0217 13:48:45.806726 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5fhmk\" (UniqueName: \"kubernetes.io/projected/b99347d9-5973-4f7e-8d8c-dad6df3e0c92-kube-api-access-5fhmk\") on node \"crc\" DevicePath \"\"" Feb 17 13:48:45 crc kubenswrapper[4955]: I0217 13:48:45.806762 4955 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b99347d9-5973-4f7e-8d8c-dad6df3e0c92-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 13:48:45 crc kubenswrapper[4955]: I0217 13:48:45.806774 4955 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b99347d9-5973-4f7e-8d8c-dad6df3e0c92-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 13:48:46 crc kubenswrapper[4955]: I0217 13:48:46.072140 4955 generic.go:334] "Generic (PLEG): container finished" podID="b99347d9-5973-4f7e-8d8c-dad6df3e0c92" containerID="41721e38bd5094f57bf243115297e935c9339cc99e9ec67c29dfde875796cb19" exitCode=0 Feb 17 13:48:46 crc kubenswrapper[4955]: I0217 13:48:46.072334 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nr6s5" event={"ID":"b99347d9-5973-4f7e-8d8c-dad6df3e0c92","Type":"ContainerDied","Data":"41721e38bd5094f57bf243115297e935c9339cc99e9ec67c29dfde875796cb19"} Feb 17 13:48:46 crc kubenswrapper[4955]: I0217 13:48:46.072474 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nr6s5" event={"ID":"b99347d9-5973-4f7e-8d8c-dad6df3e0c92","Type":"ContainerDied","Data":"9d4a3124e545eef12aa2ae869bbe177d19d8420053fe6530f75a15847cab3b92"} Feb 17 13:48:46 crc kubenswrapper[4955]: I0217 13:48:46.072398 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nr6s5" Feb 17 13:48:46 crc kubenswrapper[4955]: I0217 13:48:46.072495 4955 scope.go:117] "RemoveContainer" containerID="41721e38bd5094f57bf243115297e935c9339cc99e9ec67c29dfde875796cb19" Feb 17 13:48:46 crc kubenswrapper[4955]: I0217 13:48:46.105975 4955 scope.go:117] "RemoveContainer" containerID="66d650031e7378665b7b68cdf799e1d9bc14680c42f150232a883f8d4869ff3d" Feb 17 13:48:46 crc kubenswrapper[4955]: I0217 13:48:46.108530 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nr6s5"] Feb 17 13:48:46 crc kubenswrapper[4955]: I0217 13:48:46.116592 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nr6s5"] Feb 17 13:48:46 crc kubenswrapper[4955]: I0217 13:48:46.136643 4955 scope.go:117] "RemoveContainer" containerID="fe90edb414fd019576803f6d48082a2db875cb1d1065f73f9137e018ae75436c" Feb 17 13:48:46 crc kubenswrapper[4955]: I0217 13:48:46.187953 4955 scope.go:117] "RemoveContainer" containerID="41721e38bd5094f57bf243115297e935c9339cc99e9ec67c29dfde875796cb19" Feb 17 13:48:46 crc kubenswrapper[4955]: E0217 13:48:46.188465 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41721e38bd5094f57bf243115297e935c9339cc99e9ec67c29dfde875796cb19\": container with ID starting with 41721e38bd5094f57bf243115297e935c9339cc99e9ec67c29dfde875796cb19 not found: ID does not exist" containerID="41721e38bd5094f57bf243115297e935c9339cc99e9ec67c29dfde875796cb19" Feb 17 13:48:46 crc kubenswrapper[4955]: I0217 13:48:46.188530 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41721e38bd5094f57bf243115297e935c9339cc99e9ec67c29dfde875796cb19"} err="failed to get container status \"41721e38bd5094f57bf243115297e935c9339cc99e9ec67c29dfde875796cb19\": rpc error: code = NotFound desc = could not find container \"41721e38bd5094f57bf243115297e935c9339cc99e9ec67c29dfde875796cb19\": container with ID starting with 41721e38bd5094f57bf243115297e935c9339cc99e9ec67c29dfde875796cb19 not found: ID does not exist" Feb 17 13:48:46 crc kubenswrapper[4955]: I0217 13:48:46.188564 4955 scope.go:117] "RemoveContainer" containerID="66d650031e7378665b7b68cdf799e1d9bc14680c42f150232a883f8d4869ff3d" Feb 17 13:48:46 crc kubenswrapper[4955]: E0217 13:48:46.189166 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66d650031e7378665b7b68cdf799e1d9bc14680c42f150232a883f8d4869ff3d\": container with ID starting with 66d650031e7378665b7b68cdf799e1d9bc14680c42f150232a883f8d4869ff3d not found: ID does not exist" containerID="66d650031e7378665b7b68cdf799e1d9bc14680c42f150232a883f8d4869ff3d" Feb 17 13:48:46 crc kubenswrapper[4955]: I0217 13:48:46.189218 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66d650031e7378665b7b68cdf799e1d9bc14680c42f150232a883f8d4869ff3d"} err="failed to get container status \"66d650031e7378665b7b68cdf799e1d9bc14680c42f150232a883f8d4869ff3d\": rpc error: code = NotFound desc = could not find container \"66d650031e7378665b7b68cdf799e1d9bc14680c42f150232a883f8d4869ff3d\": container with ID starting with 66d650031e7378665b7b68cdf799e1d9bc14680c42f150232a883f8d4869ff3d not found: ID does not exist" Feb 17 13:48:46 crc kubenswrapper[4955]: I0217 13:48:46.189251 4955 scope.go:117] "RemoveContainer" containerID="fe90edb414fd019576803f6d48082a2db875cb1d1065f73f9137e018ae75436c" Feb 17 13:48:46 crc kubenswrapper[4955]: E0217 13:48:46.189523 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe90edb414fd019576803f6d48082a2db875cb1d1065f73f9137e018ae75436c\": container with ID starting with fe90edb414fd019576803f6d48082a2db875cb1d1065f73f9137e018ae75436c not found: ID does not exist" containerID="fe90edb414fd019576803f6d48082a2db875cb1d1065f73f9137e018ae75436c" Feb 17 13:48:46 crc kubenswrapper[4955]: I0217 13:48:46.189551 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe90edb414fd019576803f6d48082a2db875cb1d1065f73f9137e018ae75436c"} err="failed to get container status \"fe90edb414fd019576803f6d48082a2db875cb1d1065f73f9137e018ae75436c\": rpc error: code = NotFound desc = could not find container \"fe90edb414fd019576803f6d48082a2db875cb1d1065f73f9137e018ae75436c\": container with ID starting with fe90edb414fd019576803f6d48082a2db875cb1d1065f73f9137e018ae75436c not found: ID does not exist" Feb 17 13:48:46 crc kubenswrapper[4955]: I0217 13:48:46.235472 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b99347d9-5973-4f7e-8d8c-dad6df3e0c92" path="/var/lib/kubelet/pods/b99347d9-5973-4f7e-8d8c-dad6df3e0c92/volumes" Feb 17 13:49:04 crc kubenswrapper[4955]: I0217 13:49:04.575039 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 13:49:04 crc kubenswrapper[4955]: I0217 13:49:04.575708 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 13:49:09 crc kubenswrapper[4955]: I0217 13:49:09.770383 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6m6b5"] Feb 17 13:49:09 crc kubenswrapper[4955]: E0217 13:49:09.771451 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b99347d9-5973-4f7e-8d8c-dad6df3e0c92" containerName="extract-content" Feb 17 13:49:09 crc kubenswrapper[4955]: I0217 13:49:09.771470 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="b99347d9-5973-4f7e-8d8c-dad6df3e0c92" containerName="extract-content" Feb 17 13:49:09 crc kubenswrapper[4955]: E0217 13:49:09.771480 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b99347d9-5973-4f7e-8d8c-dad6df3e0c92" containerName="registry-server" Feb 17 13:49:09 crc kubenswrapper[4955]: I0217 13:49:09.771488 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="b99347d9-5973-4f7e-8d8c-dad6df3e0c92" containerName="registry-server" Feb 17 13:49:09 crc kubenswrapper[4955]: E0217 13:49:09.771521 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b99347d9-5973-4f7e-8d8c-dad6df3e0c92" containerName="extract-utilities" Feb 17 13:49:09 crc kubenswrapper[4955]: I0217 13:49:09.771530 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="b99347d9-5973-4f7e-8d8c-dad6df3e0c92" containerName="extract-utilities" Feb 17 13:49:09 crc kubenswrapper[4955]: I0217 13:49:09.771707 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="b99347d9-5973-4f7e-8d8c-dad6df3e0c92" containerName="registry-server" Feb 17 13:49:09 crc kubenswrapper[4955]: I0217 13:49:09.775297 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6m6b5" Feb 17 13:49:09 crc kubenswrapper[4955]: I0217 13:49:09.784168 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6m6b5"] Feb 17 13:49:09 crc kubenswrapper[4955]: I0217 13:49:09.900709 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/927236ac-2b15-4be4-8688-cf4494211b7e-utilities\") pod \"community-operators-6m6b5\" (UID: \"927236ac-2b15-4be4-8688-cf4494211b7e\") " pod="openshift-marketplace/community-operators-6m6b5" Feb 17 13:49:09 crc kubenswrapper[4955]: I0217 13:49:09.900879 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/927236ac-2b15-4be4-8688-cf4494211b7e-catalog-content\") pod \"community-operators-6m6b5\" (UID: \"927236ac-2b15-4be4-8688-cf4494211b7e\") " pod="openshift-marketplace/community-operators-6m6b5" Feb 17 13:49:09 crc kubenswrapper[4955]: I0217 13:49:09.901066 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cw7cp\" (UniqueName: \"kubernetes.io/projected/927236ac-2b15-4be4-8688-cf4494211b7e-kube-api-access-cw7cp\") pod \"community-operators-6m6b5\" (UID: \"927236ac-2b15-4be4-8688-cf4494211b7e\") " pod="openshift-marketplace/community-operators-6m6b5" Feb 17 13:49:10 crc kubenswrapper[4955]: I0217 13:49:10.002608 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/927236ac-2b15-4be4-8688-cf4494211b7e-catalog-content\") pod \"community-operators-6m6b5\" (UID: \"927236ac-2b15-4be4-8688-cf4494211b7e\") " pod="openshift-marketplace/community-operators-6m6b5" Feb 17 13:49:10 crc kubenswrapper[4955]: I0217 13:49:10.002719 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cw7cp\" (UniqueName: \"kubernetes.io/projected/927236ac-2b15-4be4-8688-cf4494211b7e-kube-api-access-cw7cp\") pod \"community-operators-6m6b5\" (UID: \"927236ac-2b15-4be4-8688-cf4494211b7e\") " pod="openshift-marketplace/community-operators-6m6b5" Feb 17 13:49:10 crc kubenswrapper[4955]: I0217 13:49:10.002888 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/927236ac-2b15-4be4-8688-cf4494211b7e-utilities\") pod \"community-operators-6m6b5\" (UID: \"927236ac-2b15-4be4-8688-cf4494211b7e\") " pod="openshift-marketplace/community-operators-6m6b5" Feb 17 13:49:10 crc kubenswrapper[4955]: I0217 13:49:10.003336 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/927236ac-2b15-4be4-8688-cf4494211b7e-utilities\") pod \"community-operators-6m6b5\" (UID: \"927236ac-2b15-4be4-8688-cf4494211b7e\") " pod="openshift-marketplace/community-operators-6m6b5" Feb 17 13:49:10 crc kubenswrapper[4955]: I0217 13:49:10.003568 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/927236ac-2b15-4be4-8688-cf4494211b7e-catalog-content\") pod \"community-operators-6m6b5\" (UID: \"927236ac-2b15-4be4-8688-cf4494211b7e\") " pod="openshift-marketplace/community-operators-6m6b5" Feb 17 13:49:10 crc kubenswrapper[4955]: I0217 13:49:10.022232 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cw7cp\" (UniqueName: \"kubernetes.io/projected/927236ac-2b15-4be4-8688-cf4494211b7e-kube-api-access-cw7cp\") pod \"community-operators-6m6b5\" (UID: \"927236ac-2b15-4be4-8688-cf4494211b7e\") " pod="openshift-marketplace/community-operators-6m6b5" Feb 17 13:49:10 crc kubenswrapper[4955]: I0217 13:49:10.102168 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6m6b5" Feb 17 13:49:10 crc kubenswrapper[4955]: I0217 13:49:10.652765 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6m6b5"] Feb 17 13:49:11 crc kubenswrapper[4955]: I0217 13:49:11.330028 4955 generic.go:334] "Generic (PLEG): container finished" podID="927236ac-2b15-4be4-8688-cf4494211b7e" containerID="f31630f9654f8f27f7eccbc6dd8e7e6f3cf6872bf64bae9ab8dca23a5e241b8e" exitCode=0 Feb 17 13:49:11 crc kubenswrapper[4955]: I0217 13:49:11.330089 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6m6b5" event={"ID":"927236ac-2b15-4be4-8688-cf4494211b7e","Type":"ContainerDied","Data":"f31630f9654f8f27f7eccbc6dd8e7e6f3cf6872bf64bae9ab8dca23a5e241b8e"} Feb 17 13:49:11 crc kubenswrapper[4955]: I0217 13:49:11.330316 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6m6b5" event={"ID":"927236ac-2b15-4be4-8688-cf4494211b7e","Type":"ContainerStarted","Data":"f831ea12e39c6c7209740a23d7cc8b976217ad08b7cc2a375bdf79ecfffacd17"} Feb 17 13:49:12 crc kubenswrapper[4955]: I0217 13:49:12.340801 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6m6b5" event={"ID":"927236ac-2b15-4be4-8688-cf4494211b7e","Type":"ContainerStarted","Data":"3f5a2b945382e487bd6a432f00b616ecb4b94a7a6237d09de75cf03055fa0ea4"} Feb 17 13:49:13 crc kubenswrapper[4955]: I0217 13:49:13.367052 4955 generic.go:334] "Generic (PLEG): container finished" podID="927236ac-2b15-4be4-8688-cf4494211b7e" containerID="3f5a2b945382e487bd6a432f00b616ecb4b94a7a6237d09de75cf03055fa0ea4" exitCode=0 Feb 17 13:49:13 crc kubenswrapper[4955]: I0217 13:49:13.367115 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6m6b5" event={"ID":"927236ac-2b15-4be4-8688-cf4494211b7e","Type":"ContainerDied","Data":"3f5a2b945382e487bd6a432f00b616ecb4b94a7a6237d09de75cf03055fa0ea4"} Feb 17 13:49:14 crc kubenswrapper[4955]: I0217 13:49:14.376735 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6m6b5" event={"ID":"927236ac-2b15-4be4-8688-cf4494211b7e","Type":"ContainerStarted","Data":"ab24368181bb7eb177ee3d0e0e4ad9f554b15a78230283333c3a10e38aee27ff"} Feb 17 13:49:14 crc kubenswrapper[4955]: I0217 13:49:14.401046 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6m6b5" podStartSLOduration=2.824714343 podStartE2EDuration="5.401020242s" podCreationTimestamp="2026-02-17 13:49:09 +0000 UTC" firstStartedPulling="2026-02-17 13:49:11.332506848 +0000 UTC m=+2689.855236391" lastFinishedPulling="2026-02-17 13:49:13.908812707 +0000 UTC m=+2692.431542290" observedRunningTime="2026-02-17 13:49:14.394937108 +0000 UTC m=+2692.917666651" watchObservedRunningTime="2026-02-17 13:49:14.401020242 +0000 UTC m=+2692.923749785" Feb 17 13:49:18 crc kubenswrapper[4955]: I0217 13:49:18.796731 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-w9gql"] Feb 17 13:49:18 crc kubenswrapper[4955]: I0217 13:49:18.799373 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w9gql" Feb 17 13:49:18 crc kubenswrapper[4955]: I0217 13:49:18.808219 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-w9gql"] Feb 17 13:49:18 crc kubenswrapper[4955]: I0217 13:49:18.898402 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fb91c58-93f6-46f1-b4e7-2fed9c334e96-catalog-content\") pod \"certified-operators-w9gql\" (UID: \"9fb91c58-93f6-46f1-b4e7-2fed9c334e96\") " pod="openshift-marketplace/certified-operators-w9gql" Feb 17 13:49:18 crc kubenswrapper[4955]: I0217 13:49:18.898497 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9f8md\" (UniqueName: \"kubernetes.io/projected/9fb91c58-93f6-46f1-b4e7-2fed9c334e96-kube-api-access-9f8md\") pod \"certified-operators-w9gql\" (UID: \"9fb91c58-93f6-46f1-b4e7-2fed9c334e96\") " pod="openshift-marketplace/certified-operators-w9gql" Feb 17 13:49:18 crc kubenswrapper[4955]: I0217 13:49:18.898553 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fb91c58-93f6-46f1-b4e7-2fed9c334e96-utilities\") pod \"certified-operators-w9gql\" (UID: \"9fb91c58-93f6-46f1-b4e7-2fed9c334e96\") " pod="openshift-marketplace/certified-operators-w9gql" Feb 17 13:49:19 crc kubenswrapper[4955]: I0217 13:49:19.000951 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fb91c58-93f6-46f1-b4e7-2fed9c334e96-catalog-content\") pod \"certified-operators-w9gql\" (UID: \"9fb91c58-93f6-46f1-b4e7-2fed9c334e96\") " pod="openshift-marketplace/certified-operators-w9gql" Feb 17 13:49:19 crc kubenswrapper[4955]: I0217 13:49:19.001028 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9f8md\" (UniqueName: \"kubernetes.io/projected/9fb91c58-93f6-46f1-b4e7-2fed9c334e96-kube-api-access-9f8md\") pod \"certified-operators-w9gql\" (UID: \"9fb91c58-93f6-46f1-b4e7-2fed9c334e96\") " pod="openshift-marketplace/certified-operators-w9gql" Feb 17 13:49:19 crc kubenswrapper[4955]: I0217 13:49:19.001075 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fb91c58-93f6-46f1-b4e7-2fed9c334e96-utilities\") pod \"certified-operators-w9gql\" (UID: \"9fb91c58-93f6-46f1-b4e7-2fed9c334e96\") " pod="openshift-marketplace/certified-operators-w9gql" Feb 17 13:49:19 crc kubenswrapper[4955]: I0217 13:49:19.001658 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fb91c58-93f6-46f1-b4e7-2fed9c334e96-utilities\") pod \"certified-operators-w9gql\" (UID: \"9fb91c58-93f6-46f1-b4e7-2fed9c334e96\") " pod="openshift-marketplace/certified-operators-w9gql" Feb 17 13:49:19 crc kubenswrapper[4955]: I0217 13:49:19.001690 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fb91c58-93f6-46f1-b4e7-2fed9c334e96-catalog-content\") pod \"certified-operators-w9gql\" (UID: \"9fb91c58-93f6-46f1-b4e7-2fed9c334e96\") " pod="openshift-marketplace/certified-operators-w9gql" Feb 17 13:49:19 crc kubenswrapper[4955]: I0217 13:49:19.029010 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9f8md\" (UniqueName: \"kubernetes.io/projected/9fb91c58-93f6-46f1-b4e7-2fed9c334e96-kube-api-access-9f8md\") pod \"certified-operators-w9gql\" (UID: \"9fb91c58-93f6-46f1-b4e7-2fed9c334e96\") " pod="openshift-marketplace/certified-operators-w9gql" Feb 17 13:49:19 crc kubenswrapper[4955]: I0217 13:49:19.163149 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w9gql" Feb 17 13:49:19 crc kubenswrapper[4955]: I0217 13:49:19.778097 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-w9gql"] Feb 17 13:49:20 crc kubenswrapper[4955]: I0217 13:49:20.102462 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6m6b5" Feb 17 13:49:20 crc kubenswrapper[4955]: I0217 13:49:20.102845 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6m6b5" Feb 17 13:49:20 crc kubenswrapper[4955]: I0217 13:49:20.153858 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6m6b5" Feb 17 13:49:20 crc kubenswrapper[4955]: I0217 13:49:20.443386 4955 generic.go:334] "Generic (PLEG): container finished" podID="9fb91c58-93f6-46f1-b4e7-2fed9c334e96" containerID="9d3b7ec8baeba3e74be585817fede85bd91632f7e53ad1a53f9754bd504c8019" exitCode=0 Feb 17 13:49:20 crc kubenswrapper[4955]: I0217 13:49:20.443597 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w9gql" event={"ID":"9fb91c58-93f6-46f1-b4e7-2fed9c334e96","Type":"ContainerDied","Data":"9d3b7ec8baeba3e74be585817fede85bd91632f7e53ad1a53f9754bd504c8019"} Feb 17 13:49:20 crc kubenswrapper[4955]: I0217 13:49:20.443634 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w9gql" event={"ID":"9fb91c58-93f6-46f1-b4e7-2fed9c334e96","Type":"ContainerStarted","Data":"d85a6ad15e6b6c9374cd32fd81c7ec8aa720ef04e1c31576e202a8d0efac1a3c"} Feb 17 13:49:20 crc kubenswrapper[4955]: I0217 13:49:20.503310 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6m6b5" Feb 17 13:49:21 crc kubenswrapper[4955]: I0217 13:49:21.196669 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6m6b5"] Feb 17 13:49:22 crc kubenswrapper[4955]: I0217 13:49:22.460589 4955 generic.go:334] "Generic (PLEG): container finished" podID="9fb91c58-93f6-46f1-b4e7-2fed9c334e96" containerID="dba06894d0feac3b5fa3187ccd46d2b447cfb5254bd78eed841ec958f7612625" exitCode=0 Feb 17 13:49:22 crc kubenswrapper[4955]: I0217 13:49:22.460667 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w9gql" event={"ID":"9fb91c58-93f6-46f1-b4e7-2fed9c334e96","Type":"ContainerDied","Data":"dba06894d0feac3b5fa3187ccd46d2b447cfb5254bd78eed841ec958f7612625"} Feb 17 13:49:22 crc kubenswrapper[4955]: I0217 13:49:22.461123 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6m6b5" podUID="927236ac-2b15-4be4-8688-cf4494211b7e" containerName="registry-server" containerID="cri-o://ab24368181bb7eb177ee3d0e0e4ad9f554b15a78230283333c3a10e38aee27ff" gracePeriod=2 Feb 17 13:49:22 crc kubenswrapper[4955]: E0217 13:49:22.793746 4955 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod927236ac_2b15_4be4_8688_cf4494211b7e.slice/crio-conmon-ab24368181bb7eb177ee3d0e0e4ad9f554b15a78230283333c3a10e38aee27ff.scope\": RecentStats: unable to find data in memory cache]" Feb 17 13:49:23 crc kubenswrapper[4955]: I0217 13:49:23.491705 4955 generic.go:334] "Generic (PLEG): container finished" podID="927236ac-2b15-4be4-8688-cf4494211b7e" containerID="ab24368181bb7eb177ee3d0e0e4ad9f554b15a78230283333c3a10e38aee27ff" exitCode=0 Feb 17 13:49:23 crc kubenswrapper[4955]: I0217 13:49:23.491835 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6m6b5" event={"ID":"927236ac-2b15-4be4-8688-cf4494211b7e","Type":"ContainerDied","Data":"ab24368181bb7eb177ee3d0e0e4ad9f554b15a78230283333c3a10e38aee27ff"} Feb 17 13:49:23 crc kubenswrapper[4955]: I0217 13:49:23.492272 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6m6b5" event={"ID":"927236ac-2b15-4be4-8688-cf4494211b7e","Type":"ContainerDied","Data":"f831ea12e39c6c7209740a23d7cc8b976217ad08b7cc2a375bdf79ecfffacd17"} Feb 17 13:49:23 crc kubenswrapper[4955]: I0217 13:49:23.492295 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f831ea12e39c6c7209740a23d7cc8b976217ad08b7cc2a375bdf79ecfffacd17" Feb 17 13:49:23 crc kubenswrapper[4955]: I0217 13:49:23.495187 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6m6b5" Feb 17 13:49:23 crc kubenswrapper[4955]: I0217 13:49:23.501658 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w9gql" event={"ID":"9fb91c58-93f6-46f1-b4e7-2fed9c334e96","Type":"ContainerStarted","Data":"3194e22d49962cfbe20e0c69579523024be60c1fe9117cbbe6e7d68695d75c9f"} Feb 17 13:49:23 crc kubenswrapper[4955]: I0217 13:49:23.543415 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cw7cp\" (UniqueName: \"kubernetes.io/projected/927236ac-2b15-4be4-8688-cf4494211b7e-kube-api-access-cw7cp\") pod \"927236ac-2b15-4be4-8688-cf4494211b7e\" (UID: \"927236ac-2b15-4be4-8688-cf4494211b7e\") " Feb 17 13:49:23 crc kubenswrapper[4955]: I0217 13:49:23.543587 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/927236ac-2b15-4be4-8688-cf4494211b7e-catalog-content\") pod \"927236ac-2b15-4be4-8688-cf4494211b7e\" (UID: \"927236ac-2b15-4be4-8688-cf4494211b7e\") " Feb 17 13:49:23 crc kubenswrapper[4955]: I0217 13:49:23.543650 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/927236ac-2b15-4be4-8688-cf4494211b7e-utilities\") pod \"927236ac-2b15-4be4-8688-cf4494211b7e\" (UID: \"927236ac-2b15-4be4-8688-cf4494211b7e\") " Feb 17 13:49:23 crc kubenswrapper[4955]: I0217 13:49:23.544559 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/927236ac-2b15-4be4-8688-cf4494211b7e-utilities" (OuterVolumeSpecName: "utilities") pod "927236ac-2b15-4be4-8688-cf4494211b7e" (UID: "927236ac-2b15-4be4-8688-cf4494211b7e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:49:23 crc kubenswrapper[4955]: I0217 13:49:23.563314 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/927236ac-2b15-4be4-8688-cf4494211b7e-kube-api-access-cw7cp" (OuterVolumeSpecName: "kube-api-access-cw7cp") pod "927236ac-2b15-4be4-8688-cf4494211b7e" (UID: "927236ac-2b15-4be4-8688-cf4494211b7e"). InnerVolumeSpecName "kube-api-access-cw7cp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:49:23 crc kubenswrapper[4955]: I0217 13:49:23.580432 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-w9gql" podStartSLOduration=3.161504588 podStartE2EDuration="5.580402485s" podCreationTimestamp="2026-02-17 13:49:18 +0000 UTC" firstStartedPulling="2026-02-17 13:49:20.453595969 +0000 UTC m=+2698.976325512" lastFinishedPulling="2026-02-17 13:49:22.872493866 +0000 UTC m=+2701.395223409" observedRunningTime="2026-02-17 13:49:23.563893674 +0000 UTC m=+2702.086623217" watchObservedRunningTime="2026-02-17 13:49:23.580402485 +0000 UTC m=+2702.103132028" Feb 17 13:49:23 crc kubenswrapper[4955]: I0217 13:49:23.600183 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/927236ac-2b15-4be4-8688-cf4494211b7e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "927236ac-2b15-4be4-8688-cf4494211b7e" (UID: "927236ac-2b15-4be4-8688-cf4494211b7e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:49:23 crc kubenswrapper[4955]: I0217 13:49:23.645022 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cw7cp\" (UniqueName: \"kubernetes.io/projected/927236ac-2b15-4be4-8688-cf4494211b7e-kube-api-access-cw7cp\") on node \"crc\" DevicePath \"\"" Feb 17 13:49:23 crc kubenswrapper[4955]: I0217 13:49:23.645062 4955 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/927236ac-2b15-4be4-8688-cf4494211b7e-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 13:49:23 crc kubenswrapper[4955]: I0217 13:49:23.645074 4955 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/927236ac-2b15-4be4-8688-cf4494211b7e-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 13:49:24 crc kubenswrapper[4955]: I0217 13:49:24.507944 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6m6b5" Feb 17 13:49:24 crc kubenswrapper[4955]: I0217 13:49:24.530750 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6m6b5"] Feb 17 13:49:24 crc kubenswrapper[4955]: I0217 13:49:24.544934 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6m6b5"] Feb 17 13:49:26 crc kubenswrapper[4955]: I0217 13:49:26.238349 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="927236ac-2b15-4be4-8688-cf4494211b7e" path="/var/lib/kubelet/pods/927236ac-2b15-4be4-8688-cf4494211b7e/volumes" Feb 17 13:49:29 crc kubenswrapper[4955]: I0217 13:49:29.163795 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-w9gql" Feb 17 13:49:29 crc kubenswrapper[4955]: I0217 13:49:29.165351 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-w9gql" Feb 17 13:49:29 crc kubenswrapper[4955]: I0217 13:49:29.211278 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-w9gql" Feb 17 13:49:29 crc kubenswrapper[4955]: I0217 13:49:29.611219 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-w9gql" Feb 17 13:49:29 crc kubenswrapper[4955]: I0217 13:49:29.982170 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-w9gql"] Feb 17 13:49:31 crc kubenswrapper[4955]: I0217 13:49:31.567367 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-w9gql" podUID="9fb91c58-93f6-46f1-b4e7-2fed9c334e96" containerName="registry-server" containerID="cri-o://3194e22d49962cfbe20e0c69579523024be60c1fe9117cbbe6e7d68695d75c9f" gracePeriod=2 Feb 17 13:49:32 crc kubenswrapper[4955]: I0217 13:49:32.070729 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w9gql" Feb 17 13:49:32 crc kubenswrapper[4955]: I0217 13:49:32.199382 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9f8md\" (UniqueName: \"kubernetes.io/projected/9fb91c58-93f6-46f1-b4e7-2fed9c334e96-kube-api-access-9f8md\") pod \"9fb91c58-93f6-46f1-b4e7-2fed9c334e96\" (UID: \"9fb91c58-93f6-46f1-b4e7-2fed9c334e96\") " Feb 17 13:49:32 crc kubenswrapper[4955]: I0217 13:49:32.199465 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fb91c58-93f6-46f1-b4e7-2fed9c334e96-utilities\") pod \"9fb91c58-93f6-46f1-b4e7-2fed9c334e96\" (UID: \"9fb91c58-93f6-46f1-b4e7-2fed9c334e96\") " Feb 17 13:49:32 crc kubenswrapper[4955]: I0217 13:49:32.199521 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fb91c58-93f6-46f1-b4e7-2fed9c334e96-catalog-content\") pod \"9fb91c58-93f6-46f1-b4e7-2fed9c334e96\" (UID: \"9fb91c58-93f6-46f1-b4e7-2fed9c334e96\") " Feb 17 13:49:32 crc kubenswrapper[4955]: I0217 13:49:32.200515 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9fb91c58-93f6-46f1-b4e7-2fed9c334e96-utilities" (OuterVolumeSpecName: "utilities") pod "9fb91c58-93f6-46f1-b4e7-2fed9c334e96" (UID: "9fb91c58-93f6-46f1-b4e7-2fed9c334e96"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:49:32 crc kubenswrapper[4955]: I0217 13:49:32.208126 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fb91c58-93f6-46f1-b4e7-2fed9c334e96-kube-api-access-9f8md" (OuterVolumeSpecName: "kube-api-access-9f8md") pod "9fb91c58-93f6-46f1-b4e7-2fed9c334e96" (UID: "9fb91c58-93f6-46f1-b4e7-2fed9c334e96"). InnerVolumeSpecName "kube-api-access-9f8md". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:49:32 crc kubenswrapper[4955]: I0217 13:49:32.291703 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9fb91c58-93f6-46f1-b4e7-2fed9c334e96-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9fb91c58-93f6-46f1-b4e7-2fed9c334e96" (UID: "9fb91c58-93f6-46f1-b4e7-2fed9c334e96"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:49:32 crc kubenswrapper[4955]: I0217 13:49:32.302454 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9f8md\" (UniqueName: \"kubernetes.io/projected/9fb91c58-93f6-46f1-b4e7-2fed9c334e96-kube-api-access-9f8md\") on node \"crc\" DevicePath \"\"" Feb 17 13:49:32 crc kubenswrapper[4955]: I0217 13:49:32.302492 4955 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fb91c58-93f6-46f1-b4e7-2fed9c334e96-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 13:49:32 crc kubenswrapper[4955]: I0217 13:49:32.302504 4955 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fb91c58-93f6-46f1-b4e7-2fed9c334e96-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 13:49:32 crc kubenswrapper[4955]: I0217 13:49:32.577480 4955 generic.go:334] "Generic (PLEG): container finished" podID="9fb91c58-93f6-46f1-b4e7-2fed9c334e96" containerID="3194e22d49962cfbe20e0c69579523024be60c1fe9117cbbe6e7d68695d75c9f" exitCode=0 Feb 17 13:49:32 crc kubenswrapper[4955]: I0217 13:49:32.577534 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w9gql" event={"ID":"9fb91c58-93f6-46f1-b4e7-2fed9c334e96","Type":"ContainerDied","Data":"3194e22d49962cfbe20e0c69579523024be60c1fe9117cbbe6e7d68695d75c9f"} Feb 17 13:49:32 crc kubenswrapper[4955]: I0217 13:49:32.577553 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w9gql" Feb 17 13:49:32 crc kubenswrapper[4955]: I0217 13:49:32.577569 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w9gql" event={"ID":"9fb91c58-93f6-46f1-b4e7-2fed9c334e96","Type":"ContainerDied","Data":"d85a6ad15e6b6c9374cd32fd81c7ec8aa720ef04e1c31576e202a8d0efac1a3c"} Feb 17 13:49:32 crc kubenswrapper[4955]: I0217 13:49:32.577586 4955 scope.go:117] "RemoveContainer" containerID="3194e22d49962cfbe20e0c69579523024be60c1fe9117cbbe6e7d68695d75c9f" Feb 17 13:49:32 crc kubenswrapper[4955]: I0217 13:49:32.597940 4955 scope.go:117] "RemoveContainer" containerID="dba06894d0feac3b5fa3187ccd46d2b447cfb5254bd78eed841ec958f7612625" Feb 17 13:49:32 crc kubenswrapper[4955]: I0217 13:49:32.609321 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-w9gql"] Feb 17 13:49:32 crc kubenswrapper[4955]: I0217 13:49:32.617437 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-w9gql"] Feb 17 13:49:32 crc kubenswrapper[4955]: I0217 13:49:32.635152 4955 scope.go:117] "RemoveContainer" containerID="9d3b7ec8baeba3e74be585817fede85bd91632f7e53ad1a53f9754bd504c8019" Feb 17 13:49:32 crc kubenswrapper[4955]: I0217 13:49:32.666445 4955 scope.go:117] "RemoveContainer" containerID="3194e22d49962cfbe20e0c69579523024be60c1fe9117cbbe6e7d68695d75c9f" Feb 17 13:49:32 crc kubenswrapper[4955]: E0217 13:49:32.666960 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3194e22d49962cfbe20e0c69579523024be60c1fe9117cbbe6e7d68695d75c9f\": container with ID starting with 3194e22d49962cfbe20e0c69579523024be60c1fe9117cbbe6e7d68695d75c9f not found: ID does not exist" containerID="3194e22d49962cfbe20e0c69579523024be60c1fe9117cbbe6e7d68695d75c9f" Feb 17 13:49:32 crc kubenswrapper[4955]: I0217 13:49:32.667015 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3194e22d49962cfbe20e0c69579523024be60c1fe9117cbbe6e7d68695d75c9f"} err="failed to get container status \"3194e22d49962cfbe20e0c69579523024be60c1fe9117cbbe6e7d68695d75c9f\": rpc error: code = NotFound desc = could not find container \"3194e22d49962cfbe20e0c69579523024be60c1fe9117cbbe6e7d68695d75c9f\": container with ID starting with 3194e22d49962cfbe20e0c69579523024be60c1fe9117cbbe6e7d68695d75c9f not found: ID does not exist" Feb 17 13:49:32 crc kubenswrapper[4955]: I0217 13:49:32.667047 4955 scope.go:117] "RemoveContainer" containerID="dba06894d0feac3b5fa3187ccd46d2b447cfb5254bd78eed841ec958f7612625" Feb 17 13:49:32 crc kubenswrapper[4955]: E0217 13:49:32.667591 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dba06894d0feac3b5fa3187ccd46d2b447cfb5254bd78eed841ec958f7612625\": container with ID starting with dba06894d0feac3b5fa3187ccd46d2b447cfb5254bd78eed841ec958f7612625 not found: ID does not exist" containerID="dba06894d0feac3b5fa3187ccd46d2b447cfb5254bd78eed841ec958f7612625" Feb 17 13:49:32 crc kubenswrapper[4955]: I0217 13:49:32.667617 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dba06894d0feac3b5fa3187ccd46d2b447cfb5254bd78eed841ec958f7612625"} err="failed to get container status \"dba06894d0feac3b5fa3187ccd46d2b447cfb5254bd78eed841ec958f7612625\": rpc error: code = NotFound desc = could not find container \"dba06894d0feac3b5fa3187ccd46d2b447cfb5254bd78eed841ec958f7612625\": container with ID starting with dba06894d0feac3b5fa3187ccd46d2b447cfb5254bd78eed841ec958f7612625 not found: ID does not exist" Feb 17 13:49:32 crc kubenswrapper[4955]: I0217 13:49:32.667635 4955 scope.go:117] "RemoveContainer" containerID="9d3b7ec8baeba3e74be585817fede85bd91632f7e53ad1a53f9754bd504c8019" Feb 17 13:49:32 crc kubenswrapper[4955]: E0217 13:49:32.667907 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d3b7ec8baeba3e74be585817fede85bd91632f7e53ad1a53f9754bd504c8019\": container with ID starting with 9d3b7ec8baeba3e74be585817fede85bd91632f7e53ad1a53f9754bd504c8019 not found: ID does not exist" containerID="9d3b7ec8baeba3e74be585817fede85bd91632f7e53ad1a53f9754bd504c8019" Feb 17 13:49:32 crc kubenswrapper[4955]: I0217 13:49:32.667936 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d3b7ec8baeba3e74be585817fede85bd91632f7e53ad1a53f9754bd504c8019"} err="failed to get container status \"9d3b7ec8baeba3e74be585817fede85bd91632f7e53ad1a53f9754bd504c8019\": rpc error: code = NotFound desc = could not find container \"9d3b7ec8baeba3e74be585817fede85bd91632f7e53ad1a53f9754bd504c8019\": container with ID starting with 9d3b7ec8baeba3e74be585817fede85bd91632f7e53ad1a53f9754bd504c8019 not found: ID does not exist" Feb 17 13:49:34 crc kubenswrapper[4955]: I0217 13:49:34.232954 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fb91c58-93f6-46f1-b4e7-2fed9c334e96" path="/var/lib/kubelet/pods/9fb91c58-93f6-46f1-b4e7-2fed9c334e96/volumes" Feb 17 13:49:34 crc kubenswrapper[4955]: I0217 13:49:34.575309 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 13:49:34 crc kubenswrapper[4955]: I0217 13:49:34.575391 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 13:50:04 crc kubenswrapper[4955]: I0217 13:50:04.575752 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 13:50:04 crc kubenswrapper[4955]: I0217 13:50:04.576771 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 13:50:04 crc kubenswrapper[4955]: I0217 13:50:04.576900 4955 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" Feb 17 13:50:04 crc kubenswrapper[4955]: I0217 13:50:04.578254 4955 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e51d645860a15fd43cd83c6bac8255e568dfbde7c066e876b04046df9a6623fc"} pod="openshift-machine-config-operator/machine-config-daemon-29qxq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 17 13:50:04 crc kubenswrapper[4955]: I0217 13:50:04.578335 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" containerID="cri-o://e51d645860a15fd43cd83c6bac8255e568dfbde7c066e876b04046df9a6623fc" gracePeriod=600 Feb 17 13:50:04 crc kubenswrapper[4955]: I0217 13:50:04.861279 4955 generic.go:334] "Generic (PLEG): container finished" podID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerID="e51d645860a15fd43cd83c6bac8255e568dfbde7c066e876b04046df9a6623fc" exitCode=0 Feb 17 13:50:04 crc kubenswrapper[4955]: I0217 13:50:04.861355 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" event={"ID":"5bf36887-2a2b-4193-abdb-a15beaf46300","Type":"ContainerDied","Data":"e51d645860a15fd43cd83c6bac8255e568dfbde7c066e876b04046df9a6623fc"} Feb 17 13:50:04 crc kubenswrapper[4955]: I0217 13:50:04.861657 4955 scope.go:117] "RemoveContainer" containerID="e81d66db8b1e6a694c97af3c245fca0ec09bc5dbf6e24a77ebe044515c97ad4e" Feb 17 13:50:05 crc kubenswrapper[4955]: I0217 13:50:05.871610 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" event={"ID":"5bf36887-2a2b-4193-abdb-a15beaf46300","Type":"ContainerStarted","Data":"61c4421445b156f2b4a07625b9dae400b378009c3aef4994cad2a6ccfbc32b03"} Feb 17 13:52:04 crc kubenswrapper[4955]: I0217 13:52:04.574438 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 13:52:04 crc kubenswrapper[4955]: I0217 13:52:04.575044 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 13:52:34 crc kubenswrapper[4955]: I0217 13:52:34.574596 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 13:52:34 crc kubenswrapper[4955]: I0217 13:52:34.575141 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 13:53:04 crc kubenswrapper[4955]: I0217 13:53:04.575379 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 13:53:04 crc kubenswrapper[4955]: I0217 13:53:04.576026 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 13:53:04 crc kubenswrapper[4955]: I0217 13:53:04.576086 4955 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" Feb 17 13:53:04 crc kubenswrapper[4955]: I0217 13:53:04.577006 4955 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"61c4421445b156f2b4a07625b9dae400b378009c3aef4994cad2a6ccfbc32b03"} pod="openshift-machine-config-operator/machine-config-daemon-29qxq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 17 13:53:04 crc kubenswrapper[4955]: I0217 13:53:04.577072 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" containerID="cri-o://61c4421445b156f2b4a07625b9dae400b378009c3aef4994cad2a6ccfbc32b03" gracePeriod=600 Feb 17 13:53:04 crc kubenswrapper[4955]: E0217 13:53:04.714508 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:53:05 crc kubenswrapper[4955]: I0217 13:53:05.465984 4955 generic.go:334] "Generic (PLEG): container finished" podID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerID="61c4421445b156f2b4a07625b9dae400b378009c3aef4994cad2a6ccfbc32b03" exitCode=0 Feb 17 13:53:05 crc kubenswrapper[4955]: I0217 13:53:05.466080 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" event={"ID":"5bf36887-2a2b-4193-abdb-a15beaf46300","Type":"ContainerDied","Data":"61c4421445b156f2b4a07625b9dae400b378009c3aef4994cad2a6ccfbc32b03"} Feb 17 13:53:05 crc kubenswrapper[4955]: I0217 13:53:05.466539 4955 scope.go:117] "RemoveContainer" containerID="e51d645860a15fd43cd83c6bac8255e568dfbde7c066e876b04046df9a6623fc" Feb 17 13:53:05 crc kubenswrapper[4955]: I0217 13:53:05.467352 4955 scope.go:117] "RemoveContainer" containerID="61c4421445b156f2b4a07625b9dae400b378009c3aef4994cad2a6ccfbc32b03" Feb 17 13:53:05 crc kubenswrapper[4955]: E0217 13:53:05.467615 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:53:19 crc kubenswrapper[4955]: I0217 13:53:19.222636 4955 scope.go:117] "RemoveContainer" containerID="61c4421445b156f2b4a07625b9dae400b378009c3aef4994cad2a6ccfbc32b03" Feb 17 13:53:19 crc kubenswrapper[4955]: E0217 13:53:19.223447 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:53:31 crc kubenswrapper[4955]: I0217 13:53:31.223808 4955 scope.go:117] "RemoveContainer" containerID="61c4421445b156f2b4a07625b9dae400b378009c3aef4994cad2a6ccfbc32b03" Feb 17 13:53:31 crc kubenswrapper[4955]: E0217 13:53:31.224546 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:53:42 crc kubenswrapper[4955]: I0217 13:53:42.231049 4955 scope.go:117] "RemoveContainer" containerID="61c4421445b156f2b4a07625b9dae400b378009c3aef4994cad2a6ccfbc32b03" Feb 17 13:53:42 crc kubenswrapper[4955]: E0217 13:53:42.233148 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:53:53 crc kubenswrapper[4955]: I0217 13:53:53.224132 4955 scope.go:117] "RemoveContainer" containerID="61c4421445b156f2b4a07625b9dae400b378009c3aef4994cad2a6ccfbc32b03" Feb 17 13:53:53 crc kubenswrapper[4955]: E0217 13:53:53.227563 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:54:07 crc kubenswrapper[4955]: I0217 13:54:07.223502 4955 scope.go:117] "RemoveContainer" containerID="61c4421445b156f2b4a07625b9dae400b378009c3aef4994cad2a6ccfbc32b03" Feb 17 13:54:07 crc kubenswrapper[4955]: E0217 13:54:07.224304 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:54:18 crc kubenswrapper[4955]: I0217 13:54:18.223560 4955 scope.go:117] "RemoveContainer" containerID="61c4421445b156f2b4a07625b9dae400b378009c3aef4994cad2a6ccfbc32b03" Feb 17 13:54:18 crc kubenswrapper[4955]: E0217 13:54:18.224359 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:54:32 crc kubenswrapper[4955]: I0217 13:54:32.231365 4955 scope.go:117] "RemoveContainer" containerID="61c4421445b156f2b4a07625b9dae400b378009c3aef4994cad2a6ccfbc32b03" Feb 17 13:54:32 crc kubenswrapper[4955]: E0217 13:54:32.232177 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:54:44 crc kubenswrapper[4955]: I0217 13:54:44.222902 4955 scope.go:117] "RemoveContainer" containerID="61c4421445b156f2b4a07625b9dae400b378009c3aef4994cad2a6ccfbc32b03" Feb 17 13:54:44 crc kubenswrapper[4955]: E0217 13:54:44.223612 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:54:56 crc kubenswrapper[4955]: I0217 13:54:56.223214 4955 scope.go:117] "RemoveContainer" containerID="61c4421445b156f2b4a07625b9dae400b378009c3aef4994cad2a6ccfbc32b03" Feb 17 13:54:56 crc kubenswrapper[4955]: E0217 13:54:56.225127 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:55:11 crc kubenswrapper[4955]: I0217 13:55:11.223075 4955 scope.go:117] "RemoveContainer" containerID="61c4421445b156f2b4a07625b9dae400b378009c3aef4994cad2a6ccfbc32b03" Feb 17 13:55:11 crc kubenswrapper[4955]: E0217 13:55:11.223844 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:55:23 crc kubenswrapper[4955]: I0217 13:55:23.223415 4955 scope.go:117] "RemoveContainer" containerID="61c4421445b156f2b4a07625b9dae400b378009c3aef4994cad2a6ccfbc32b03" Feb 17 13:55:23 crc kubenswrapper[4955]: E0217 13:55:23.224379 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:55:34 crc kubenswrapper[4955]: I0217 13:55:34.223227 4955 scope.go:117] "RemoveContainer" containerID="61c4421445b156f2b4a07625b9dae400b378009c3aef4994cad2a6ccfbc32b03" Feb 17 13:55:34 crc kubenswrapper[4955]: E0217 13:55:34.224001 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:55:49 crc kubenswrapper[4955]: I0217 13:55:49.223812 4955 scope.go:117] "RemoveContainer" containerID="61c4421445b156f2b4a07625b9dae400b378009c3aef4994cad2a6ccfbc32b03" Feb 17 13:55:49 crc kubenswrapper[4955]: E0217 13:55:49.224529 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:55:51 crc kubenswrapper[4955]: I0217 13:55:51.434253 4955 scope.go:117] "RemoveContainer" containerID="ab24368181bb7eb177ee3d0e0e4ad9f554b15a78230283333c3a10e38aee27ff" Feb 17 13:55:51 crc kubenswrapper[4955]: I0217 13:55:51.460418 4955 scope.go:117] "RemoveContainer" containerID="f31630f9654f8f27f7eccbc6dd8e7e6f3cf6872bf64bae9ab8dca23a5e241b8e" Feb 17 13:55:51 crc kubenswrapper[4955]: I0217 13:55:51.483406 4955 scope.go:117] "RemoveContainer" containerID="3f5a2b945382e487bd6a432f00b616ecb4b94a7a6237d09de75cf03055fa0ea4" Feb 17 13:56:04 crc kubenswrapper[4955]: I0217 13:56:04.223432 4955 scope.go:117] "RemoveContainer" containerID="61c4421445b156f2b4a07625b9dae400b378009c3aef4994cad2a6ccfbc32b03" Feb 17 13:56:04 crc kubenswrapper[4955]: E0217 13:56:04.235348 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:56:18 crc kubenswrapper[4955]: I0217 13:56:18.223846 4955 scope.go:117] "RemoveContainer" containerID="61c4421445b156f2b4a07625b9dae400b378009c3aef4994cad2a6ccfbc32b03" Feb 17 13:56:18 crc kubenswrapper[4955]: E0217 13:56:18.225033 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:56:24 crc kubenswrapper[4955]: I0217 13:56:24.353644 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cwhq9"] Feb 17 13:56:24 crc kubenswrapper[4955]: E0217 13:56:24.354922 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fb91c58-93f6-46f1-b4e7-2fed9c334e96" containerName="extract-utilities" Feb 17 13:56:24 crc kubenswrapper[4955]: I0217 13:56:24.354941 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fb91c58-93f6-46f1-b4e7-2fed9c334e96" containerName="extract-utilities" Feb 17 13:56:24 crc kubenswrapper[4955]: E0217 13:56:24.354965 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="927236ac-2b15-4be4-8688-cf4494211b7e" containerName="extract-utilities" Feb 17 13:56:24 crc kubenswrapper[4955]: I0217 13:56:24.354972 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="927236ac-2b15-4be4-8688-cf4494211b7e" containerName="extract-utilities" Feb 17 13:56:24 crc kubenswrapper[4955]: E0217 13:56:24.354994 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="927236ac-2b15-4be4-8688-cf4494211b7e" containerName="extract-content" Feb 17 13:56:24 crc kubenswrapper[4955]: I0217 13:56:24.355004 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="927236ac-2b15-4be4-8688-cf4494211b7e" containerName="extract-content" Feb 17 13:56:24 crc kubenswrapper[4955]: E0217 13:56:24.355012 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="927236ac-2b15-4be4-8688-cf4494211b7e" containerName="registry-server" Feb 17 13:56:24 crc kubenswrapper[4955]: I0217 13:56:24.355019 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="927236ac-2b15-4be4-8688-cf4494211b7e" containerName="registry-server" Feb 17 13:56:24 crc kubenswrapper[4955]: E0217 13:56:24.355036 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fb91c58-93f6-46f1-b4e7-2fed9c334e96" containerName="extract-content" Feb 17 13:56:24 crc kubenswrapper[4955]: I0217 13:56:24.355043 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fb91c58-93f6-46f1-b4e7-2fed9c334e96" containerName="extract-content" Feb 17 13:56:24 crc kubenswrapper[4955]: E0217 13:56:24.355069 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fb91c58-93f6-46f1-b4e7-2fed9c334e96" containerName="registry-server" Feb 17 13:56:24 crc kubenswrapper[4955]: I0217 13:56:24.355076 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fb91c58-93f6-46f1-b4e7-2fed9c334e96" containerName="registry-server" Feb 17 13:56:24 crc kubenswrapper[4955]: I0217 13:56:24.355287 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fb91c58-93f6-46f1-b4e7-2fed9c334e96" containerName="registry-server" Feb 17 13:56:24 crc kubenswrapper[4955]: I0217 13:56:24.355323 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="927236ac-2b15-4be4-8688-cf4494211b7e" containerName="registry-server" Feb 17 13:56:24 crc kubenswrapper[4955]: I0217 13:56:24.356933 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cwhq9" Feb 17 13:56:24 crc kubenswrapper[4955]: I0217 13:56:24.367576 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cwhq9"] Feb 17 13:56:24 crc kubenswrapper[4955]: I0217 13:56:24.531491 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff405b07-9eff-432a-b709-55b3b2fb6180-catalog-content\") pod \"redhat-operators-cwhq9\" (UID: \"ff405b07-9eff-432a-b709-55b3b2fb6180\") " pod="openshift-marketplace/redhat-operators-cwhq9" Feb 17 13:56:24 crc kubenswrapper[4955]: I0217 13:56:24.531840 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff405b07-9eff-432a-b709-55b3b2fb6180-utilities\") pod \"redhat-operators-cwhq9\" (UID: \"ff405b07-9eff-432a-b709-55b3b2fb6180\") " pod="openshift-marketplace/redhat-operators-cwhq9" Feb 17 13:56:24 crc kubenswrapper[4955]: I0217 13:56:24.531924 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62cdw\" (UniqueName: \"kubernetes.io/projected/ff405b07-9eff-432a-b709-55b3b2fb6180-kube-api-access-62cdw\") pod \"redhat-operators-cwhq9\" (UID: \"ff405b07-9eff-432a-b709-55b3b2fb6180\") " pod="openshift-marketplace/redhat-operators-cwhq9" Feb 17 13:56:24 crc kubenswrapper[4955]: I0217 13:56:24.633656 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff405b07-9eff-432a-b709-55b3b2fb6180-catalog-content\") pod \"redhat-operators-cwhq9\" (UID: \"ff405b07-9eff-432a-b709-55b3b2fb6180\") " pod="openshift-marketplace/redhat-operators-cwhq9" Feb 17 13:56:24 crc kubenswrapper[4955]: I0217 13:56:24.633766 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff405b07-9eff-432a-b709-55b3b2fb6180-utilities\") pod \"redhat-operators-cwhq9\" (UID: \"ff405b07-9eff-432a-b709-55b3b2fb6180\") " pod="openshift-marketplace/redhat-operators-cwhq9" Feb 17 13:56:24 crc kubenswrapper[4955]: I0217 13:56:24.633815 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62cdw\" (UniqueName: \"kubernetes.io/projected/ff405b07-9eff-432a-b709-55b3b2fb6180-kube-api-access-62cdw\") pod \"redhat-operators-cwhq9\" (UID: \"ff405b07-9eff-432a-b709-55b3b2fb6180\") " pod="openshift-marketplace/redhat-operators-cwhq9" Feb 17 13:56:24 crc kubenswrapper[4955]: I0217 13:56:24.634226 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff405b07-9eff-432a-b709-55b3b2fb6180-catalog-content\") pod \"redhat-operators-cwhq9\" (UID: \"ff405b07-9eff-432a-b709-55b3b2fb6180\") " pod="openshift-marketplace/redhat-operators-cwhq9" Feb 17 13:56:24 crc kubenswrapper[4955]: I0217 13:56:24.634336 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff405b07-9eff-432a-b709-55b3b2fb6180-utilities\") pod \"redhat-operators-cwhq9\" (UID: \"ff405b07-9eff-432a-b709-55b3b2fb6180\") " pod="openshift-marketplace/redhat-operators-cwhq9" Feb 17 13:56:24 crc kubenswrapper[4955]: I0217 13:56:24.659340 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62cdw\" (UniqueName: \"kubernetes.io/projected/ff405b07-9eff-432a-b709-55b3b2fb6180-kube-api-access-62cdw\") pod \"redhat-operators-cwhq9\" (UID: \"ff405b07-9eff-432a-b709-55b3b2fb6180\") " pod="openshift-marketplace/redhat-operators-cwhq9" Feb 17 13:56:24 crc kubenswrapper[4955]: I0217 13:56:24.679418 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cwhq9" Feb 17 13:56:25 crc kubenswrapper[4955]: I0217 13:56:25.185409 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cwhq9"] Feb 17 13:56:26 crc kubenswrapper[4955]: I0217 13:56:26.156230 4955 generic.go:334] "Generic (PLEG): container finished" podID="ff405b07-9eff-432a-b709-55b3b2fb6180" containerID="3fddb0e3c4f6bbcc36bb5bc82ea8088ae8c482a3321f60c7833776d0d19b1aa0" exitCode=0 Feb 17 13:56:26 crc kubenswrapper[4955]: I0217 13:56:26.156339 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cwhq9" event={"ID":"ff405b07-9eff-432a-b709-55b3b2fb6180","Type":"ContainerDied","Data":"3fddb0e3c4f6bbcc36bb5bc82ea8088ae8c482a3321f60c7833776d0d19b1aa0"} Feb 17 13:56:26 crc kubenswrapper[4955]: I0217 13:56:26.156886 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cwhq9" event={"ID":"ff405b07-9eff-432a-b709-55b3b2fb6180","Type":"ContainerStarted","Data":"e7d5a4efa0bef4a040ea27f3797f101cf7bac7834a0d93bfc7a694da0a0397a7"} Feb 17 13:56:26 crc kubenswrapper[4955]: I0217 13:56:26.159481 4955 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 17 13:56:28 crc kubenswrapper[4955]: I0217 13:56:28.179808 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cwhq9" event={"ID":"ff405b07-9eff-432a-b709-55b3b2fb6180","Type":"ContainerStarted","Data":"5ce01c00315824c1f7753dc39a194c3d72ea32835298b391acf82885a4a0fdb2"} Feb 17 13:56:29 crc kubenswrapper[4955]: I0217 13:56:29.224383 4955 scope.go:117] "RemoveContainer" containerID="61c4421445b156f2b4a07625b9dae400b378009c3aef4994cad2a6ccfbc32b03" Feb 17 13:56:29 crc kubenswrapper[4955]: E0217 13:56:29.225113 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:56:31 crc kubenswrapper[4955]: I0217 13:56:31.205669 4955 generic.go:334] "Generic (PLEG): container finished" podID="ff405b07-9eff-432a-b709-55b3b2fb6180" containerID="5ce01c00315824c1f7753dc39a194c3d72ea32835298b391acf82885a4a0fdb2" exitCode=0 Feb 17 13:56:31 crc kubenswrapper[4955]: I0217 13:56:31.205887 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cwhq9" event={"ID":"ff405b07-9eff-432a-b709-55b3b2fb6180","Type":"ContainerDied","Data":"5ce01c00315824c1f7753dc39a194c3d72ea32835298b391acf82885a4a0fdb2"} Feb 17 13:56:32 crc kubenswrapper[4955]: I0217 13:56:32.217738 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cwhq9" event={"ID":"ff405b07-9eff-432a-b709-55b3b2fb6180","Type":"ContainerStarted","Data":"6a1d1750324a6ff97fade362a096e0efa70e0f1f97a6618a9be62302dd7a46c5"} Feb 17 13:56:32 crc kubenswrapper[4955]: I0217 13:56:32.240821 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cwhq9" podStartSLOduration=2.7609530380000002 podStartE2EDuration="8.240800873s" podCreationTimestamp="2026-02-17 13:56:24 +0000 UTC" firstStartedPulling="2026-02-17 13:56:26.15925471 +0000 UTC m=+3124.681984253" lastFinishedPulling="2026-02-17 13:56:31.639102545 +0000 UTC m=+3130.161832088" observedRunningTime="2026-02-17 13:56:32.239711332 +0000 UTC m=+3130.762440885" watchObservedRunningTime="2026-02-17 13:56:32.240800873 +0000 UTC m=+3130.763530416" Feb 17 13:56:34 crc kubenswrapper[4955]: I0217 13:56:34.680681 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cwhq9" Feb 17 13:56:34 crc kubenswrapper[4955]: I0217 13:56:34.682159 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cwhq9" Feb 17 13:56:35 crc kubenswrapper[4955]: I0217 13:56:35.730628 4955 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-cwhq9" podUID="ff405b07-9eff-432a-b709-55b3b2fb6180" containerName="registry-server" probeResult="failure" output=< Feb 17 13:56:35 crc kubenswrapper[4955]: timeout: failed to connect service ":50051" within 1s Feb 17 13:56:35 crc kubenswrapper[4955]: > Feb 17 13:56:41 crc kubenswrapper[4955]: I0217 13:56:41.223425 4955 scope.go:117] "RemoveContainer" containerID="61c4421445b156f2b4a07625b9dae400b378009c3aef4994cad2a6ccfbc32b03" Feb 17 13:56:41 crc kubenswrapper[4955]: E0217 13:56:41.224145 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:56:44 crc kubenswrapper[4955]: I0217 13:56:44.747747 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cwhq9" Feb 17 13:56:44 crc kubenswrapper[4955]: I0217 13:56:44.799203 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cwhq9" Feb 17 13:56:44 crc kubenswrapper[4955]: I0217 13:56:44.988440 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cwhq9"] Feb 17 13:56:46 crc kubenswrapper[4955]: I0217 13:56:46.329145 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-cwhq9" podUID="ff405b07-9eff-432a-b709-55b3b2fb6180" containerName="registry-server" containerID="cri-o://6a1d1750324a6ff97fade362a096e0efa70e0f1f97a6618a9be62302dd7a46c5" gracePeriod=2 Feb 17 13:56:46 crc kubenswrapper[4955]: I0217 13:56:46.798645 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cwhq9" Feb 17 13:56:46 crc kubenswrapper[4955]: I0217 13:56:46.953473 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff405b07-9eff-432a-b709-55b3b2fb6180-utilities\") pod \"ff405b07-9eff-432a-b709-55b3b2fb6180\" (UID: \"ff405b07-9eff-432a-b709-55b3b2fb6180\") " Feb 17 13:56:46 crc kubenswrapper[4955]: I0217 13:56:46.953611 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62cdw\" (UniqueName: \"kubernetes.io/projected/ff405b07-9eff-432a-b709-55b3b2fb6180-kube-api-access-62cdw\") pod \"ff405b07-9eff-432a-b709-55b3b2fb6180\" (UID: \"ff405b07-9eff-432a-b709-55b3b2fb6180\") " Feb 17 13:56:46 crc kubenswrapper[4955]: I0217 13:56:46.954589 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff405b07-9eff-432a-b709-55b3b2fb6180-catalog-content\") pod \"ff405b07-9eff-432a-b709-55b3b2fb6180\" (UID: \"ff405b07-9eff-432a-b709-55b3b2fb6180\") " Feb 17 13:56:46 crc kubenswrapper[4955]: I0217 13:56:46.954752 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff405b07-9eff-432a-b709-55b3b2fb6180-utilities" (OuterVolumeSpecName: "utilities") pod "ff405b07-9eff-432a-b709-55b3b2fb6180" (UID: "ff405b07-9eff-432a-b709-55b3b2fb6180"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:56:46 crc kubenswrapper[4955]: I0217 13:56:46.955761 4955 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff405b07-9eff-432a-b709-55b3b2fb6180-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 13:56:46 crc kubenswrapper[4955]: I0217 13:56:46.959126 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff405b07-9eff-432a-b709-55b3b2fb6180-kube-api-access-62cdw" (OuterVolumeSpecName: "kube-api-access-62cdw") pod "ff405b07-9eff-432a-b709-55b3b2fb6180" (UID: "ff405b07-9eff-432a-b709-55b3b2fb6180"). InnerVolumeSpecName "kube-api-access-62cdw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:56:47 crc kubenswrapper[4955]: I0217 13:56:47.057880 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62cdw\" (UniqueName: \"kubernetes.io/projected/ff405b07-9eff-432a-b709-55b3b2fb6180-kube-api-access-62cdw\") on node \"crc\" DevicePath \"\"" Feb 17 13:56:47 crc kubenswrapper[4955]: I0217 13:56:47.083935 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff405b07-9eff-432a-b709-55b3b2fb6180-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ff405b07-9eff-432a-b709-55b3b2fb6180" (UID: "ff405b07-9eff-432a-b709-55b3b2fb6180"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:56:47 crc kubenswrapper[4955]: I0217 13:56:47.159815 4955 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff405b07-9eff-432a-b709-55b3b2fb6180-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 13:56:47 crc kubenswrapper[4955]: I0217 13:56:47.340857 4955 generic.go:334] "Generic (PLEG): container finished" podID="ff405b07-9eff-432a-b709-55b3b2fb6180" containerID="6a1d1750324a6ff97fade362a096e0efa70e0f1f97a6618a9be62302dd7a46c5" exitCode=0 Feb 17 13:56:47 crc kubenswrapper[4955]: I0217 13:56:47.340910 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cwhq9" event={"ID":"ff405b07-9eff-432a-b709-55b3b2fb6180","Type":"ContainerDied","Data":"6a1d1750324a6ff97fade362a096e0efa70e0f1f97a6618a9be62302dd7a46c5"} Feb 17 13:56:47 crc kubenswrapper[4955]: I0217 13:56:47.340939 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cwhq9" event={"ID":"ff405b07-9eff-432a-b709-55b3b2fb6180","Type":"ContainerDied","Data":"e7d5a4efa0bef4a040ea27f3797f101cf7bac7834a0d93bfc7a694da0a0397a7"} Feb 17 13:56:47 crc kubenswrapper[4955]: I0217 13:56:47.340939 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cwhq9" Feb 17 13:56:47 crc kubenswrapper[4955]: I0217 13:56:47.340968 4955 scope.go:117] "RemoveContainer" containerID="6a1d1750324a6ff97fade362a096e0efa70e0f1f97a6618a9be62302dd7a46c5" Feb 17 13:56:47 crc kubenswrapper[4955]: I0217 13:56:47.360685 4955 scope.go:117] "RemoveContainer" containerID="5ce01c00315824c1f7753dc39a194c3d72ea32835298b391acf82885a4a0fdb2" Feb 17 13:56:47 crc kubenswrapper[4955]: I0217 13:56:47.379105 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cwhq9"] Feb 17 13:56:47 crc kubenswrapper[4955]: I0217 13:56:47.384161 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-cwhq9"] Feb 17 13:56:47 crc kubenswrapper[4955]: I0217 13:56:47.402554 4955 scope.go:117] "RemoveContainer" containerID="3fddb0e3c4f6bbcc36bb5bc82ea8088ae8c482a3321f60c7833776d0d19b1aa0" Feb 17 13:56:47 crc kubenswrapper[4955]: I0217 13:56:47.430581 4955 scope.go:117] "RemoveContainer" containerID="6a1d1750324a6ff97fade362a096e0efa70e0f1f97a6618a9be62302dd7a46c5" Feb 17 13:56:47 crc kubenswrapper[4955]: E0217 13:56:47.431146 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a1d1750324a6ff97fade362a096e0efa70e0f1f97a6618a9be62302dd7a46c5\": container with ID starting with 6a1d1750324a6ff97fade362a096e0efa70e0f1f97a6618a9be62302dd7a46c5 not found: ID does not exist" containerID="6a1d1750324a6ff97fade362a096e0efa70e0f1f97a6618a9be62302dd7a46c5" Feb 17 13:56:47 crc kubenswrapper[4955]: I0217 13:56:47.431196 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a1d1750324a6ff97fade362a096e0efa70e0f1f97a6618a9be62302dd7a46c5"} err="failed to get container status \"6a1d1750324a6ff97fade362a096e0efa70e0f1f97a6618a9be62302dd7a46c5\": rpc error: code = NotFound desc = could not find container \"6a1d1750324a6ff97fade362a096e0efa70e0f1f97a6618a9be62302dd7a46c5\": container with ID starting with 6a1d1750324a6ff97fade362a096e0efa70e0f1f97a6618a9be62302dd7a46c5 not found: ID does not exist" Feb 17 13:56:47 crc kubenswrapper[4955]: I0217 13:56:47.431226 4955 scope.go:117] "RemoveContainer" containerID="5ce01c00315824c1f7753dc39a194c3d72ea32835298b391acf82885a4a0fdb2" Feb 17 13:56:47 crc kubenswrapper[4955]: E0217 13:56:47.431588 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ce01c00315824c1f7753dc39a194c3d72ea32835298b391acf82885a4a0fdb2\": container with ID starting with 5ce01c00315824c1f7753dc39a194c3d72ea32835298b391acf82885a4a0fdb2 not found: ID does not exist" containerID="5ce01c00315824c1f7753dc39a194c3d72ea32835298b391acf82885a4a0fdb2" Feb 17 13:56:47 crc kubenswrapper[4955]: I0217 13:56:47.431627 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ce01c00315824c1f7753dc39a194c3d72ea32835298b391acf82885a4a0fdb2"} err="failed to get container status \"5ce01c00315824c1f7753dc39a194c3d72ea32835298b391acf82885a4a0fdb2\": rpc error: code = NotFound desc = could not find container \"5ce01c00315824c1f7753dc39a194c3d72ea32835298b391acf82885a4a0fdb2\": container with ID starting with 5ce01c00315824c1f7753dc39a194c3d72ea32835298b391acf82885a4a0fdb2 not found: ID does not exist" Feb 17 13:56:47 crc kubenswrapper[4955]: I0217 13:56:47.431640 4955 scope.go:117] "RemoveContainer" containerID="3fddb0e3c4f6bbcc36bb5bc82ea8088ae8c482a3321f60c7833776d0d19b1aa0" Feb 17 13:56:47 crc kubenswrapper[4955]: E0217 13:56:47.432050 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fddb0e3c4f6bbcc36bb5bc82ea8088ae8c482a3321f60c7833776d0d19b1aa0\": container with ID starting with 3fddb0e3c4f6bbcc36bb5bc82ea8088ae8c482a3321f60c7833776d0d19b1aa0 not found: ID does not exist" containerID="3fddb0e3c4f6bbcc36bb5bc82ea8088ae8c482a3321f60c7833776d0d19b1aa0" Feb 17 13:56:47 crc kubenswrapper[4955]: I0217 13:56:47.432082 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fddb0e3c4f6bbcc36bb5bc82ea8088ae8c482a3321f60c7833776d0d19b1aa0"} err="failed to get container status \"3fddb0e3c4f6bbcc36bb5bc82ea8088ae8c482a3321f60c7833776d0d19b1aa0\": rpc error: code = NotFound desc = could not find container \"3fddb0e3c4f6bbcc36bb5bc82ea8088ae8c482a3321f60c7833776d0d19b1aa0\": container with ID starting with 3fddb0e3c4f6bbcc36bb5bc82ea8088ae8c482a3321f60c7833776d0d19b1aa0 not found: ID does not exist" Feb 17 13:56:48 crc kubenswrapper[4955]: I0217 13:56:48.505046 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff405b07-9eff-432a-b709-55b3b2fb6180" path="/var/lib/kubelet/pods/ff405b07-9eff-432a-b709-55b3b2fb6180/volumes" Feb 17 13:56:49 crc kubenswrapper[4955]: I0217 13:56:49.478092 4955 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/redhat-operators-qz5rg" podUID="a5c8c29f-7770-4ba6-aad9-12604d10981f" containerName="registry-server" probeResult="failure" output=< Feb 17 13:56:49 crc kubenswrapper[4955]: timeout: failed to connect service ":50051" within 1s Feb 17 13:56:49 crc kubenswrapper[4955]: > Feb 17 13:56:49 crc kubenswrapper[4955]: I0217 13:56:49.488374 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/redhat-operators-qz5rg" podUID="a5c8c29f-7770-4ba6-aad9-12604d10981f" containerName="registry-server" probeResult="failure" output=< Feb 17 13:56:49 crc kubenswrapper[4955]: timeout: failed to connect service ":50051" within 1s Feb 17 13:56:49 crc kubenswrapper[4955]: > Feb 17 13:56:55 crc kubenswrapper[4955]: I0217 13:56:55.222964 4955 scope.go:117] "RemoveContainer" containerID="61c4421445b156f2b4a07625b9dae400b378009c3aef4994cad2a6ccfbc32b03" Feb 17 13:56:55 crc kubenswrapper[4955]: E0217 13:56:55.223815 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:57:07 crc kubenswrapper[4955]: I0217 13:57:07.223161 4955 scope.go:117] "RemoveContainer" containerID="61c4421445b156f2b4a07625b9dae400b378009c3aef4994cad2a6ccfbc32b03" Feb 17 13:57:07 crc kubenswrapper[4955]: E0217 13:57:07.223948 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:57:20 crc kubenswrapper[4955]: I0217 13:57:20.223642 4955 scope.go:117] "RemoveContainer" containerID="61c4421445b156f2b4a07625b9dae400b378009c3aef4994cad2a6ccfbc32b03" Feb 17 13:57:20 crc kubenswrapper[4955]: E0217 13:57:20.224452 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:57:32 crc kubenswrapper[4955]: I0217 13:57:32.230480 4955 scope.go:117] "RemoveContainer" containerID="61c4421445b156f2b4a07625b9dae400b378009c3aef4994cad2a6ccfbc32b03" Feb 17 13:57:32 crc kubenswrapper[4955]: E0217 13:57:32.231250 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:57:46 crc kubenswrapper[4955]: I0217 13:57:46.224196 4955 scope.go:117] "RemoveContainer" containerID="61c4421445b156f2b4a07625b9dae400b378009c3aef4994cad2a6ccfbc32b03" Feb 17 13:57:46 crc kubenswrapper[4955]: E0217 13:57:46.225073 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:57:57 crc kubenswrapper[4955]: I0217 13:57:57.122317 4955 generic.go:334] "Generic (PLEG): container finished" podID="de1d0057-d623-4185-b31a-572dfaabcce2" containerID="bdbd1017b2f134a5c55399b939afe437fbca6626b6654182d4ac13d5dde9f701" exitCode=0 Feb 17 13:57:57 crc kubenswrapper[4955]: I0217 13:57:57.122388 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"de1d0057-d623-4185-b31a-572dfaabcce2","Type":"ContainerDied","Data":"bdbd1017b2f134a5c55399b939afe437fbca6626b6654182d4ac13d5dde9f701"} Feb 17 13:57:58 crc kubenswrapper[4955]: I0217 13:57:58.223333 4955 scope.go:117] "RemoveContainer" containerID="61c4421445b156f2b4a07625b9dae400b378009c3aef4994cad2a6ccfbc32b03" Feb 17 13:57:58 crc kubenswrapper[4955]: E0217 13:57:58.224010 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 13:57:58 crc kubenswrapper[4955]: I0217 13:57:58.497766 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 17 13:57:58 crc kubenswrapper[4955]: I0217 13:57:58.596105 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/de1d0057-d623-4185-b31a-572dfaabcce2-config-data\") pod \"de1d0057-d623-4185-b31a-572dfaabcce2\" (UID: \"de1d0057-d623-4185-b31a-572dfaabcce2\") " Feb 17 13:57:58 crc kubenswrapper[4955]: I0217 13:57:58.596162 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/de1d0057-d623-4185-b31a-572dfaabcce2-ca-certs\") pod \"de1d0057-d623-4185-b31a-572dfaabcce2\" (UID: \"de1d0057-d623-4185-b31a-572dfaabcce2\") " Feb 17 13:57:58 crc kubenswrapper[4955]: I0217 13:57:58.596189 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/de1d0057-d623-4185-b31a-572dfaabcce2-openstack-config\") pod \"de1d0057-d623-4185-b31a-572dfaabcce2\" (UID: \"de1d0057-d623-4185-b31a-572dfaabcce2\") " Feb 17 13:57:58 crc kubenswrapper[4955]: I0217 13:57:58.596318 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/de1d0057-d623-4185-b31a-572dfaabcce2-test-operator-ephemeral-workdir\") pod \"de1d0057-d623-4185-b31a-572dfaabcce2\" (UID: \"de1d0057-d623-4185-b31a-572dfaabcce2\") " Feb 17 13:57:58 crc kubenswrapper[4955]: I0217 13:57:58.596393 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/de1d0057-d623-4185-b31a-572dfaabcce2-openstack-config-secret\") pod \"de1d0057-d623-4185-b31a-572dfaabcce2\" (UID: \"de1d0057-d623-4185-b31a-572dfaabcce2\") " Feb 17 13:57:58 crc kubenswrapper[4955]: I0217 13:57:58.596416 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2vt7w\" (UniqueName: \"kubernetes.io/projected/de1d0057-d623-4185-b31a-572dfaabcce2-kube-api-access-2vt7w\") pod \"de1d0057-d623-4185-b31a-572dfaabcce2\" (UID: \"de1d0057-d623-4185-b31a-572dfaabcce2\") " Feb 17 13:57:58 crc kubenswrapper[4955]: I0217 13:57:58.596439 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"de1d0057-d623-4185-b31a-572dfaabcce2\" (UID: \"de1d0057-d623-4185-b31a-572dfaabcce2\") " Feb 17 13:57:58 crc kubenswrapper[4955]: I0217 13:57:58.596470 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de1d0057-d623-4185-b31a-572dfaabcce2-ssh-key\") pod \"de1d0057-d623-4185-b31a-572dfaabcce2\" (UID: \"de1d0057-d623-4185-b31a-572dfaabcce2\") " Feb 17 13:57:58 crc kubenswrapper[4955]: I0217 13:57:58.596507 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/de1d0057-d623-4185-b31a-572dfaabcce2-test-operator-ephemeral-temporary\") pod \"de1d0057-d623-4185-b31a-572dfaabcce2\" (UID: \"de1d0057-d623-4185-b31a-572dfaabcce2\") " Feb 17 13:57:58 crc kubenswrapper[4955]: I0217 13:57:58.596983 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de1d0057-d623-4185-b31a-572dfaabcce2-config-data" (OuterVolumeSpecName: "config-data") pod "de1d0057-d623-4185-b31a-572dfaabcce2" (UID: "de1d0057-d623-4185-b31a-572dfaabcce2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:57:58 crc kubenswrapper[4955]: I0217 13:57:58.597222 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de1d0057-d623-4185-b31a-572dfaabcce2-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "de1d0057-d623-4185-b31a-572dfaabcce2" (UID: "de1d0057-d623-4185-b31a-572dfaabcce2"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:57:58 crc kubenswrapper[4955]: I0217 13:57:58.602826 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de1d0057-d623-4185-b31a-572dfaabcce2-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "de1d0057-d623-4185-b31a-572dfaabcce2" (UID: "de1d0057-d623-4185-b31a-572dfaabcce2"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:57:58 crc kubenswrapper[4955]: I0217 13:57:58.602955 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "test-operator-logs") pod "de1d0057-d623-4185-b31a-572dfaabcce2" (UID: "de1d0057-d623-4185-b31a-572dfaabcce2"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 17 13:57:58 crc kubenswrapper[4955]: I0217 13:57:58.615705 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de1d0057-d623-4185-b31a-572dfaabcce2-kube-api-access-2vt7w" (OuterVolumeSpecName: "kube-api-access-2vt7w") pod "de1d0057-d623-4185-b31a-572dfaabcce2" (UID: "de1d0057-d623-4185-b31a-572dfaabcce2"). InnerVolumeSpecName "kube-api-access-2vt7w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:57:58 crc kubenswrapper[4955]: I0217 13:57:58.625605 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de1d0057-d623-4185-b31a-572dfaabcce2-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "de1d0057-d623-4185-b31a-572dfaabcce2" (UID: "de1d0057-d623-4185-b31a-572dfaabcce2"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:57:58 crc kubenswrapper[4955]: I0217 13:57:58.628684 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de1d0057-d623-4185-b31a-572dfaabcce2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "de1d0057-d623-4185-b31a-572dfaabcce2" (UID: "de1d0057-d623-4185-b31a-572dfaabcce2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:57:58 crc kubenswrapper[4955]: I0217 13:57:58.633047 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de1d0057-d623-4185-b31a-572dfaabcce2-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "de1d0057-d623-4185-b31a-572dfaabcce2" (UID: "de1d0057-d623-4185-b31a-572dfaabcce2"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 13:57:58 crc kubenswrapper[4955]: I0217 13:57:58.663925 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de1d0057-d623-4185-b31a-572dfaabcce2-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "de1d0057-d623-4185-b31a-572dfaabcce2" (UID: "de1d0057-d623-4185-b31a-572dfaabcce2"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 13:57:58 crc kubenswrapper[4955]: I0217 13:57:58.699620 4955 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/de1d0057-d623-4185-b31a-572dfaabcce2-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Feb 17 13:57:58 crc kubenswrapper[4955]: I0217 13:57:58.699684 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2vt7w\" (UniqueName: \"kubernetes.io/projected/de1d0057-d623-4185-b31a-572dfaabcce2-kube-api-access-2vt7w\") on node \"crc\" DevicePath \"\"" Feb 17 13:57:58 crc kubenswrapper[4955]: I0217 13:57:58.699726 4955 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Feb 17 13:57:58 crc kubenswrapper[4955]: I0217 13:57:58.699739 4955 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de1d0057-d623-4185-b31a-572dfaabcce2-ssh-key\") on node \"crc\" DevicePath \"\"" Feb 17 13:57:58 crc kubenswrapper[4955]: I0217 13:57:58.699792 4955 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/de1d0057-d623-4185-b31a-572dfaabcce2-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Feb 17 13:57:58 crc kubenswrapper[4955]: I0217 13:57:58.699807 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/de1d0057-d623-4185-b31a-572dfaabcce2-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 13:57:58 crc kubenswrapper[4955]: I0217 13:57:58.699821 4955 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/de1d0057-d623-4185-b31a-572dfaabcce2-ca-certs\") on node \"crc\" DevicePath \"\"" Feb 17 13:57:58 crc kubenswrapper[4955]: I0217 13:57:58.699833 4955 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/de1d0057-d623-4185-b31a-572dfaabcce2-openstack-config\") on node \"crc\" DevicePath \"\"" Feb 17 13:57:58 crc kubenswrapper[4955]: I0217 13:57:58.699846 4955 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/de1d0057-d623-4185-b31a-572dfaabcce2-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Feb 17 13:57:58 crc kubenswrapper[4955]: I0217 13:57:58.724050 4955 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Feb 17 13:57:58 crc kubenswrapper[4955]: I0217 13:57:58.801649 4955 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Feb 17 13:57:59 crc kubenswrapper[4955]: I0217 13:57:59.142775 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"de1d0057-d623-4185-b31a-572dfaabcce2","Type":"ContainerDied","Data":"1b6db8750e4fdbf7eb868f485323ebcabc6e6a0f3301eae8715e83791a8df991"} Feb 17 13:57:59 crc kubenswrapper[4955]: I0217 13:57:59.142848 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b6db8750e4fdbf7eb868f485323ebcabc6e6a0f3301eae8715e83791a8df991" Feb 17 13:57:59 crc kubenswrapper[4955]: I0217 13:57:59.142951 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 17 13:58:01 crc kubenswrapper[4955]: I0217 13:58:01.521178 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 17 13:58:01 crc kubenswrapper[4955]: E0217 13:58:01.522182 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de1d0057-d623-4185-b31a-572dfaabcce2" containerName="tempest-tests-tempest-tests-runner" Feb 17 13:58:01 crc kubenswrapper[4955]: I0217 13:58:01.522196 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="de1d0057-d623-4185-b31a-572dfaabcce2" containerName="tempest-tests-tempest-tests-runner" Feb 17 13:58:01 crc kubenswrapper[4955]: E0217 13:58:01.522205 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff405b07-9eff-432a-b709-55b3b2fb6180" containerName="extract-utilities" Feb 17 13:58:01 crc kubenswrapper[4955]: I0217 13:58:01.522211 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff405b07-9eff-432a-b709-55b3b2fb6180" containerName="extract-utilities" Feb 17 13:58:01 crc kubenswrapper[4955]: E0217 13:58:01.522252 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff405b07-9eff-432a-b709-55b3b2fb6180" containerName="extract-content" Feb 17 13:58:01 crc kubenswrapper[4955]: I0217 13:58:01.522259 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff405b07-9eff-432a-b709-55b3b2fb6180" containerName="extract-content" Feb 17 13:58:01 crc kubenswrapper[4955]: E0217 13:58:01.522274 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff405b07-9eff-432a-b709-55b3b2fb6180" containerName="registry-server" Feb 17 13:58:01 crc kubenswrapper[4955]: I0217 13:58:01.522279 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff405b07-9eff-432a-b709-55b3b2fb6180" containerName="registry-server" Feb 17 13:58:01 crc kubenswrapper[4955]: I0217 13:58:01.522481 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="de1d0057-d623-4185-b31a-572dfaabcce2" containerName="tempest-tests-tempest-tests-runner" Feb 17 13:58:01 crc kubenswrapper[4955]: I0217 13:58:01.522496 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff405b07-9eff-432a-b709-55b3b2fb6180" containerName="registry-server" Feb 17 13:58:01 crc kubenswrapper[4955]: I0217 13:58:01.523144 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 17 13:58:01 crc kubenswrapper[4955]: I0217 13:58:01.525141 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-kwzmd" Feb 17 13:58:01 crc kubenswrapper[4955]: I0217 13:58:01.530071 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 17 13:58:01 crc kubenswrapper[4955]: I0217 13:58:01.657670 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"35a9a907-8d3b-49dd-9dd3-0f2e1a5aa7bc\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 17 13:58:01 crc kubenswrapper[4955]: I0217 13:58:01.657913 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbjff\" (UniqueName: \"kubernetes.io/projected/35a9a907-8d3b-49dd-9dd3-0f2e1a5aa7bc-kube-api-access-hbjff\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"35a9a907-8d3b-49dd-9dd3-0f2e1a5aa7bc\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 17 13:58:01 crc kubenswrapper[4955]: I0217 13:58:01.760048 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"35a9a907-8d3b-49dd-9dd3-0f2e1a5aa7bc\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 17 13:58:01 crc kubenswrapper[4955]: I0217 13:58:01.760450 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbjff\" (UniqueName: \"kubernetes.io/projected/35a9a907-8d3b-49dd-9dd3-0f2e1a5aa7bc-kube-api-access-hbjff\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"35a9a907-8d3b-49dd-9dd3-0f2e1a5aa7bc\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 17 13:58:01 crc kubenswrapper[4955]: I0217 13:58:01.760612 4955 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"35a9a907-8d3b-49dd-9dd3-0f2e1a5aa7bc\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 17 13:58:01 crc kubenswrapper[4955]: I0217 13:58:01.781811 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbjff\" (UniqueName: \"kubernetes.io/projected/35a9a907-8d3b-49dd-9dd3-0f2e1a5aa7bc-kube-api-access-hbjff\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"35a9a907-8d3b-49dd-9dd3-0f2e1a5aa7bc\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 17 13:58:01 crc kubenswrapper[4955]: I0217 13:58:01.792296 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"35a9a907-8d3b-49dd-9dd3-0f2e1a5aa7bc\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 17 13:58:01 crc kubenswrapper[4955]: I0217 13:58:01.851895 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 17 13:58:02 crc kubenswrapper[4955]: I0217 13:58:02.309012 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 17 13:58:03 crc kubenswrapper[4955]: I0217 13:58:03.172086 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"35a9a907-8d3b-49dd-9dd3-0f2e1a5aa7bc","Type":"ContainerStarted","Data":"55ea777c536e8c24ac7bf940c23db34b98663da81fbc22557d286a82b6aaa92a"} Feb 17 13:58:04 crc kubenswrapper[4955]: I0217 13:58:04.183174 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"35a9a907-8d3b-49dd-9dd3-0f2e1a5aa7bc","Type":"ContainerStarted","Data":"4a44c8131e7293c214a61bac6cc14dbcf0a84908372ad293641d7febbcffd6a6"} Feb 17 13:58:04 crc kubenswrapper[4955]: I0217 13:58:04.199910 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.206281671 podStartE2EDuration="3.199888204s" podCreationTimestamp="2026-02-17 13:58:01 +0000 UTC" firstStartedPulling="2026-02-17 13:58:02.317506758 +0000 UTC m=+3220.840236301" lastFinishedPulling="2026-02-17 13:58:03.311113291 +0000 UTC m=+3221.833842834" observedRunningTime="2026-02-17 13:58:04.195819897 +0000 UTC m=+3222.718549460" watchObservedRunningTime="2026-02-17 13:58:04.199888204 +0000 UTC m=+3222.722617747" Feb 17 13:58:11 crc kubenswrapper[4955]: I0217 13:58:11.223027 4955 scope.go:117] "RemoveContainer" containerID="61c4421445b156f2b4a07625b9dae400b378009c3aef4994cad2a6ccfbc32b03" Feb 17 13:58:12 crc kubenswrapper[4955]: I0217 13:58:12.250174 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" event={"ID":"5bf36887-2a2b-4193-abdb-a15beaf46300","Type":"ContainerStarted","Data":"f365e77198a4218f176dac0c14964d043e7082a6840a3a671b2914acc5870d29"} Feb 17 13:58:24 crc kubenswrapper[4955]: I0217 13:58:24.925810 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-b5tlj/must-gather-nmxsd"] Feb 17 13:58:24 crc kubenswrapper[4955]: I0217 13:58:24.928223 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b5tlj/must-gather-nmxsd" Feb 17 13:58:24 crc kubenswrapper[4955]: I0217 13:58:24.932150 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-b5tlj"/"openshift-service-ca.crt" Feb 17 13:58:24 crc kubenswrapper[4955]: I0217 13:58:24.932316 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-b5tlj"/"kube-root-ca.crt" Feb 17 13:58:24 crc kubenswrapper[4955]: I0217 13:58:24.955274 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-b5tlj/must-gather-nmxsd"] Feb 17 13:58:25 crc kubenswrapper[4955]: I0217 13:58:25.032643 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/131a5d3f-e605-4310-a2ff-2044b7afcdbd-must-gather-output\") pod \"must-gather-nmxsd\" (UID: \"131a5d3f-e605-4310-a2ff-2044b7afcdbd\") " pod="openshift-must-gather-b5tlj/must-gather-nmxsd" Feb 17 13:58:25 crc kubenswrapper[4955]: I0217 13:58:25.032807 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7fhs\" (UniqueName: \"kubernetes.io/projected/131a5d3f-e605-4310-a2ff-2044b7afcdbd-kube-api-access-h7fhs\") pod \"must-gather-nmxsd\" (UID: \"131a5d3f-e605-4310-a2ff-2044b7afcdbd\") " pod="openshift-must-gather-b5tlj/must-gather-nmxsd" Feb 17 13:58:25 crc kubenswrapper[4955]: I0217 13:58:25.133865 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7fhs\" (UniqueName: \"kubernetes.io/projected/131a5d3f-e605-4310-a2ff-2044b7afcdbd-kube-api-access-h7fhs\") pod \"must-gather-nmxsd\" (UID: \"131a5d3f-e605-4310-a2ff-2044b7afcdbd\") " pod="openshift-must-gather-b5tlj/must-gather-nmxsd" Feb 17 13:58:25 crc kubenswrapper[4955]: I0217 13:58:25.133977 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/131a5d3f-e605-4310-a2ff-2044b7afcdbd-must-gather-output\") pod \"must-gather-nmxsd\" (UID: \"131a5d3f-e605-4310-a2ff-2044b7afcdbd\") " pod="openshift-must-gather-b5tlj/must-gather-nmxsd" Feb 17 13:58:25 crc kubenswrapper[4955]: I0217 13:58:25.134388 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/131a5d3f-e605-4310-a2ff-2044b7afcdbd-must-gather-output\") pod \"must-gather-nmxsd\" (UID: \"131a5d3f-e605-4310-a2ff-2044b7afcdbd\") " pod="openshift-must-gather-b5tlj/must-gather-nmxsd" Feb 17 13:58:25 crc kubenswrapper[4955]: I0217 13:58:25.152656 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7fhs\" (UniqueName: \"kubernetes.io/projected/131a5d3f-e605-4310-a2ff-2044b7afcdbd-kube-api-access-h7fhs\") pod \"must-gather-nmxsd\" (UID: \"131a5d3f-e605-4310-a2ff-2044b7afcdbd\") " pod="openshift-must-gather-b5tlj/must-gather-nmxsd" Feb 17 13:58:25 crc kubenswrapper[4955]: I0217 13:58:25.249924 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b5tlj/must-gather-nmxsd" Feb 17 13:58:25 crc kubenswrapper[4955]: I0217 13:58:25.713343 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-b5tlj/must-gather-nmxsd"] Feb 17 13:58:26 crc kubenswrapper[4955]: I0217 13:58:26.365270 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b5tlj/must-gather-nmxsd" event={"ID":"131a5d3f-e605-4310-a2ff-2044b7afcdbd","Type":"ContainerStarted","Data":"e4cd4f85959e252b0a5bca51cd9613f92e95fa099e3c7285b841845b60b45c92"} Feb 17 13:58:32 crc kubenswrapper[4955]: I0217 13:58:32.414469 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b5tlj/must-gather-nmxsd" event={"ID":"131a5d3f-e605-4310-a2ff-2044b7afcdbd","Type":"ContainerStarted","Data":"251a0123e1e6c1cab048e9516aa909045ebc74811c69feaa5983297329d348f4"} Feb 17 13:58:32 crc kubenswrapper[4955]: I0217 13:58:32.415042 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b5tlj/must-gather-nmxsd" event={"ID":"131a5d3f-e605-4310-a2ff-2044b7afcdbd","Type":"ContainerStarted","Data":"5b074e5605e0d42f70cda17d7dbba758f1c0a68e29c8ad26ad1f174dee3db699"} Feb 17 13:58:32 crc kubenswrapper[4955]: I0217 13:58:32.431700 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-b5tlj/must-gather-nmxsd" podStartSLOduration=2.671437876 podStartE2EDuration="8.431677041s" podCreationTimestamp="2026-02-17 13:58:24 +0000 UTC" firstStartedPulling="2026-02-17 13:58:25.723216911 +0000 UTC m=+3244.245946454" lastFinishedPulling="2026-02-17 13:58:31.483456076 +0000 UTC m=+3250.006185619" observedRunningTime="2026-02-17 13:58:32.430294281 +0000 UTC m=+3250.953023814" watchObservedRunningTime="2026-02-17 13:58:32.431677041 +0000 UTC m=+3250.954406604" Feb 17 13:58:35 crc kubenswrapper[4955]: I0217 13:58:35.584953 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-b5tlj/crc-debug-98j4z"] Feb 17 13:58:35 crc kubenswrapper[4955]: I0217 13:58:35.586342 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b5tlj/crc-debug-98j4z" Feb 17 13:58:35 crc kubenswrapper[4955]: I0217 13:58:35.589350 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-b5tlj"/"default-dockercfg-56j4v" Feb 17 13:58:35 crc kubenswrapper[4955]: I0217 13:58:35.761368 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/325195a8-36d8-4ec8-9af3-ce87ad94af40-host\") pod \"crc-debug-98j4z\" (UID: \"325195a8-36d8-4ec8-9af3-ce87ad94af40\") " pod="openshift-must-gather-b5tlj/crc-debug-98j4z" Feb 17 13:58:35 crc kubenswrapper[4955]: I0217 13:58:35.761807 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svk74\" (UniqueName: \"kubernetes.io/projected/325195a8-36d8-4ec8-9af3-ce87ad94af40-kube-api-access-svk74\") pod \"crc-debug-98j4z\" (UID: \"325195a8-36d8-4ec8-9af3-ce87ad94af40\") " pod="openshift-must-gather-b5tlj/crc-debug-98j4z" Feb 17 13:58:35 crc kubenswrapper[4955]: I0217 13:58:35.863633 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/325195a8-36d8-4ec8-9af3-ce87ad94af40-host\") pod \"crc-debug-98j4z\" (UID: \"325195a8-36d8-4ec8-9af3-ce87ad94af40\") " pod="openshift-must-gather-b5tlj/crc-debug-98j4z" Feb 17 13:58:35 crc kubenswrapper[4955]: I0217 13:58:35.863901 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svk74\" (UniqueName: \"kubernetes.io/projected/325195a8-36d8-4ec8-9af3-ce87ad94af40-kube-api-access-svk74\") pod \"crc-debug-98j4z\" (UID: \"325195a8-36d8-4ec8-9af3-ce87ad94af40\") " pod="openshift-must-gather-b5tlj/crc-debug-98j4z" Feb 17 13:58:35 crc kubenswrapper[4955]: I0217 13:58:35.863734 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/325195a8-36d8-4ec8-9af3-ce87ad94af40-host\") pod \"crc-debug-98j4z\" (UID: \"325195a8-36d8-4ec8-9af3-ce87ad94af40\") " pod="openshift-must-gather-b5tlj/crc-debug-98j4z" Feb 17 13:58:35 crc kubenswrapper[4955]: I0217 13:58:35.886247 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svk74\" (UniqueName: \"kubernetes.io/projected/325195a8-36d8-4ec8-9af3-ce87ad94af40-kube-api-access-svk74\") pod \"crc-debug-98j4z\" (UID: \"325195a8-36d8-4ec8-9af3-ce87ad94af40\") " pod="openshift-must-gather-b5tlj/crc-debug-98j4z" Feb 17 13:58:35 crc kubenswrapper[4955]: I0217 13:58:35.906487 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b5tlj/crc-debug-98j4z" Feb 17 13:58:35 crc kubenswrapper[4955]: W0217 13:58:35.985875 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod325195a8_36d8_4ec8_9af3_ce87ad94af40.slice/crio-1acbf02e5ef827881b13b117c1ab89775ae25ee5416a47df61008f1b83a6667a WatchSource:0}: Error finding container 1acbf02e5ef827881b13b117c1ab89775ae25ee5416a47df61008f1b83a6667a: Status 404 returned error can't find the container with id 1acbf02e5ef827881b13b117c1ab89775ae25ee5416a47df61008f1b83a6667a Feb 17 13:58:36 crc kubenswrapper[4955]: I0217 13:58:36.344164 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kzfmk"] Feb 17 13:58:36 crc kubenswrapper[4955]: I0217 13:58:36.346905 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kzfmk" Feb 17 13:58:36 crc kubenswrapper[4955]: I0217 13:58:36.355683 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kzfmk"] Feb 17 13:58:36 crc kubenswrapper[4955]: I0217 13:58:36.448021 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b5tlj/crc-debug-98j4z" event={"ID":"325195a8-36d8-4ec8-9af3-ce87ad94af40","Type":"ContainerStarted","Data":"1acbf02e5ef827881b13b117c1ab89775ae25ee5416a47df61008f1b83a6667a"} Feb 17 13:58:36 crc kubenswrapper[4955]: I0217 13:58:36.478408 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5bzt\" (UniqueName: \"kubernetes.io/projected/5bdcb9ea-c5c7-4893-9067-b1a17072dbe8-kube-api-access-k5bzt\") pod \"redhat-marketplace-kzfmk\" (UID: \"5bdcb9ea-c5c7-4893-9067-b1a17072dbe8\") " pod="openshift-marketplace/redhat-marketplace-kzfmk" Feb 17 13:58:36 crc kubenswrapper[4955]: I0217 13:58:36.478463 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bdcb9ea-c5c7-4893-9067-b1a17072dbe8-utilities\") pod \"redhat-marketplace-kzfmk\" (UID: \"5bdcb9ea-c5c7-4893-9067-b1a17072dbe8\") " pod="openshift-marketplace/redhat-marketplace-kzfmk" Feb 17 13:58:36 crc kubenswrapper[4955]: I0217 13:58:36.478493 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bdcb9ea-c5c7-4893-9067-b1a17072dbe8-catalog-content\") pod \"redhat-marketplace-kzfmk\" (UID: \"5bdcb9ea-c5c7-4893-9067-b1a17072dbe8\") " pod="openshift-marketplace/redhat-marketplace-kzfmk" Feb 17 13:58:36 crc kubenswrapper[4955]: I0217 13:58:36.581013 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5bzt\" (UniqueName: \"kubernetes.io/projected/5bdcb9ea-c5c7-4893-9067-b1a17072dbe8-kube-api-access-k5bzt\") pod \"redhat-marketplace-kzfmk\" (UID: \"5bdcb9ea-c5c7-4893-9067-b1a17072dbe8\") " pod="openshift-marketplace/redhat-marketplace-kzfmk" Feb 17 13:58:36 crc kubenswrapper[4955]: I0217 13:58:36.581077 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bdcb9ea-c5c7-4893-9067-b1a17072dbe8-utilities\") pod \"redhat-marketplace-kzfmk\" (UID: \"5bdcb9ea-c5c7-4893-9067-b1a17072dbe8\") " pod="openshift-marketplace/redhat-marketplace-kzfmk" Feb 17 13:58:36 crc kubenswrapper[4955]: I0217 13:58:36.581128 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bdcb9ea-c5c7-4893-9067-b1a17072dbe8-catalog-content\") pod \"redhat-marketplace-kzfmk\" (UID: \"5bdcb9ea-c5c7-4893-9067-b1a17072dbe8\") " pod="openshift-marketplace/redhat-marketplace-kzfmk" Feb 17 13:58:36 crc kubenswrapper[4955]: I0217 13:58:36.581637 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bdcb9ea-c5c7-4893-9067-b1a17072dbe8-utilities\") pod \"redhat-marketplace-kzfmk\" (UID: \"5bdcb9ea-c5c7-4893-9067-b1a17072dbe8\") " pod="openshift-marketplace/redhat-marketplace-kzfmk" Feb 17 13:58:36 crc kubenswrapper[4955]: I0217 13:58:36.581679 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bdcb9ea-c5c7-4893-9067-b1a17072dbe8-catalog-content\") pod \"redhat-marketplace-kzfmk\" (UID: \"5bdcb9ea-c5c7-4893-9067-b1a17072dbe8\") " pod="openshift-marketplace/redhat-marketplace-kzfmk" Feb 17 13:58:36 crc kubenswrapper[4955]: I0217 13:58:36.601690 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5bzt\" (UniqueName: \"kubernetes.io/projected/5bdcb9ea-c5c7-4893-9067-b1a17072dbe8-kube-api-access-k5bzt\") pod \"redhat-marketplace-kzfmk\" (UID: \"5bdcb9ea-c5c7-4893-9067-b1a17072dbe8\") " pod="openshift-marketplace/redhat-marketplace-kzfmk" Feb 17 13:58:36 crc kubenswrapper[4955]: I0217 13:58:36.681315 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kzfmk" Feb 17 13:58:37 crc kubenswrapper[4955]: I0217 13:58:37.282586 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kzfmk"] Feb 17 13:58:37 crc kubenswrapper[4955]: I0217 13:58:37.471831 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kzfmk" event={"ID":"5bdcb9ea-c5c7-4893-9067-b1a17072dbe8","Type":"ContainerStarted","Data":"614f282f17a50832326c9d008e5594cd956b3b6fdbe5f3148fa3ce56c9b33696"} Feb 17 13:58:38 crc kubenswrapper[4955]: I0217 13:58:38.481814 4955 generic.go:334] "Generic (PLEG): container finished" podID="5bdcb9ea-c5c7-4893-9067-b1a17072dbe8" containerID="59410cf5393a924855abb129a83f4b3aca3c339a70233c2dad7476ca5d85fad8" exitCode=0 Feb 17 13:58:38 crc kubenswrapper[4955]: I0217 13:58:38.482120 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kzfmk" event={"ID":"5bdcb9ea-c5c7-4893-9067-b1a17072dbe8","Type":"ContainerDied","Data":"59410cf5393a924855abb129a83f4b3aca3c339a70233c2dad7476ca5d85fad8"} Feb 17 13:58:39 crc kubenswrapper[4955]: I0217 13:58:39.494408 4955 generic.go:334] "Generic (PLEG): container finished" podID="5bdcb9ea-c5c7-4893-9067-b1a17072dbe8" containerID="4a0e7d54c7368f13336ab0ea8fa18583001ba1768eebb7fb048d1d8d56d5d04f" exitCode=0 Feb 17 13:58:39 crc kubenswrapper[4955]: I0217 13:58:39.494491 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kzfmk" event={"ID":"5bdcb9ea-c5c7-4893-9067-b1a17072dbe8","Type":"ContainerDied","Data":"4a0e7d54c7368f13336ab0ea8fa18583001ba1768eebb7fb048d1d8d56d5d04f"} Feb 17 13:58:40 crc kubenswrapper[4955]: I0217 13:58:40.530130 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kzfmk" event={"ID":"5bdcb9ea-c5c7-4893-9067-b1a17072dbe8","Type":"ContainerStarted","Data":"955841fd3a9aaf41f8536e73b49561ed53741e28be4a2ed98f009b33e906b467"} Feb 17 13:58:40 crc kubenswrapper[4955]: I0217 13:58:40.560374 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kzfmk" podStartSLOduration=2.853205917 podStartE2EDuration="4.560348846s" podCreationTimestamp="2026-02-17 13:58:36 +0000 UTC" firstStartedPulling="2026-02-17 13:58:38.484260975 +0000 UTC m=+3257.006990518" lastFinishedPulling="2026-02-17 13:58:40.191403904 +0000 UTC m=+3258.714133447" observedRunningTime="2026-02-17 13:58:40.559317536 +0000 UTC m=+3259.082047099" watchObservedRunningTime="2026-02-17 13:58:40.560348846 +0000 UTC m=+3259.083078389" Feb 17 13:58:46 crc kubenswrapper[4955]: I0217 13:58:46.681730 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kzfmk" Feb 17 13:58:46 crc kubenswrapper[4955]: I0217 13:58:46.682396 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kzfmk" Feb 17 13:58:46 crc kubenswrapper[4955]: I0217 13:58:46.750353 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kzfmk" Feb 17 13:58:47 crc kubenswrapper[4955]: I0217 13:58:47.667657 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kzfmk" Feb 17 13:58:47 crc kubenswrapper[4955]: I0217 13:58:47.718953 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kzfmk"] Feb 17 13:58:49 crc kubenswrapper[4955]: I0217 13:58:49.635873 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b5tlj/crc-debug-98j4z" event={"ID":"325195a8-36d8-4ec8-9af3-ce87ad94af40","Type":"ContainerStarted","Data":"240783d5b2b351a103df8937664ab4a362aa27e4e7f0deeae2dc9febc23ed024"} Feb 17 13:58:49 crc kubenswrapper[4955]: I0217 13:58:49.636068 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kzfmk" podUID="5bdcb9ea-c5c7-4893-9067-b1a17072dbe8" containerName="registry-server" containerID="cri-o://955841fd3a9aaf41f8536e73b49561ed53741e28be4a2ed98f009b33e906b467" gracePeriod=2 Feb 17 13:58:50 crc kubenswrapper[4955]: I0217 13:58:50.181344 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kzfmk" Feb 17 13:58:50 crc kubenswrapper[4955]: I0217 13:58:50.228279 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-b5tlj/crc-debug-98j4z" podStartSLOduration=2.028644754 podStartE2EDuration="15.228263333s" podCreationTimestamp="2026-02-17 13:58:35 +0000 UTC" firstStartedPulling="2026-02-17 13:58:35.98886346 +0000 UTC m=+3254.511593013" lastFinishedPulling="2026-02-17 13:58:49.188482049 +0000 UTC m=+3267.711211592" observedRunningTime="2026-02-17 13:58:49.663247359 +0000 UTC m=+3268.185976902" watchObservedRunningTime="2026-02-17 13:58:50.228263333 +0000 UTC m=+3268.750992876" Feb 17 13:58:50 crc kubenswrapper[4955]: I0217 13:58:50.362044 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5bzt\" (UniqueName: \"kubernetes.io/projected/5bdcb9ea-c5c7-4893-9067-b1a17072dbe8-kube-api-access-k5bzt\") pod \"5bdcb9ea-c5c7-4893-9067-b1a17072dbe8\" (UID: \"5bdcb9ea-c5c7-4893-9067-b1a17072dbe8\") " Feb 17 13:58:50 crc kubenswrapper[4955]: I0217 13:58:50.362164 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bdcb9ea-c5c7-4893-9067-b1a17072dbe8-utilities\") pod \"5bdcb9ea-c5c7-4893-9067-b1a17072dbe8\" (UID: \"5bdcb9ea-c5c7-4893-9067-b1a17072dbe8\") " Feb 17 13:58:50 crc kubenswrapper[4955]: I0217 13:58:50.362201 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bdcb9ea-c5c7-4893-9067-b1a17072dbe8-catalog-content\") pod \"5bdcb9ea-c5c7-4893-9067-b1a17072dbe8\" (UID: \"5bdcb9ea-c5c7-4893-9067-b1a17072dbe8\") " Feb 17 13:58:50 crc kubenswrapper[4955]: I0217 13:58:50.362821 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5bdcb9ea-c5c7-4893-9067-b1a17072dbe8-utilities" (OuterVolumeSpecName: "utilities") pod "5bdcb9ea-c5c7-4893-9067-b1a17072dbe8" (UID: "5bdcb9ea-c5c7-4893-9067-b1a17072dbe8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:58:50 crc kubenswrapper[4955]: I0217 13:58:50.363668 4955 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bdcb9ea-c5c7-4893-9067-b1a17072dbe8-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 13:58:50 crc kubenswrapper[4955]: I0217 13:58:50.368086 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bdcb9ea-c5c7-4893-9067-b1a17072dbe8-kube-api-access-k5bzt" (OuterVolumeSpecName: "kube-api-access-k5bzt") pod "5bdcb9ea-c5c7-4893-9067-b1a17072dbe8" (UID: "5bdcb9ea-c5c7-4893-9067-b1a17072dbe8"). InnerVolumeSpecName "kube-api-access-k5bzt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:58:50 crc kubenswrapper[4955]: I0217 13:58:50.390013 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5bdcb9ea-c5c7-4893-9067-b1a17072dbe8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5bdcb9ea-c5c7-4893-9067-b1a17072dbe8" (UID: "5bdcb9ea-c5c7-4893-9067-b1a17072dbe8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:58:50 crc kubenswrapper[4955]: I0217 13:58:50.465403 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5bzt\" (UniqueName: \"kubernetes.io/projected/5bdcb9ea-c5c7-4893-9067-b1a17072dbe8-kube-api-access-k5bzt\") on node \"crc\" DevicePath \"\"" Feb 17 13:58:50 crc kubenswrapper[4955]: I0217 13:58:50.465448 4955 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bdcb9ea-c5c7-4893-9067-b1a17072dbe8-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 13:58:50 crc kubenswrapper[4955]: I0217 13:58:50.647822 4955 generic.go:334] "Generic (PLEG): container finished" podID="5bdcb9ea-c5c7-4893-9067-b1a17072dbe8" containerID="955841fd3a9aaf41f8536e73b49561ed53741e28be4a2ed98f009b33e906b467" exitCode=0 Feb 17 13:58:50 crc kubenswrapper[4955]: I0217 13:58:50.647925 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kzfmk" Feb 17 13:58:50 crc kubenswrapper[4955]: I0217 13:58:50.648830 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kzfmk" event={"ID":"5bdcb9ea-c5c7-4893-9067-b1a17072dbe8","Type":"ContainerDied","Data":"955841fd3a9aaf41f8536e73b49561ed53741e28be4a2ed98f009b33e906b467"} Feb 17 13:58:50 crc kubenswrapper[4955]: I0217 13:58:50.648860 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kzfmk" event={"ID":"5bdcb9ea-c5c7-4893-9067-b1a17072dbe8","Type":"ContainerDied","Data":"614f282f17a50832326c9d008e5594cd956b3b6fdbe5f3148fa3ce56c9b33696"} Feb 17 13:58:50 crc kubenswrapper[4955]: I0217 13:58:50.648877 4955 scope.go:117] "RemoveContainer" containerID="955841fd3a9aaf41f8536e73b49561ed53741e28be4a2ed98f009b33e906b467" Feb 17 13:58:50 crc kubenswrapper[4955]: I0217 13:58:50.677084 4955 scope.go:117] "RemoveContainer" containerID="4a0e7d54c7368f13336ab0ea8fa18583001ba1768eebb7fb048d1d8d56d5d04f" Feb 17 13:58:50 crc kubenswrapper[4955]: I0217 13:58:50.702590 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kzfmk"] Feb 17 13:58:50 crc kubenswrapper[4955]: I0217 13:58:50.714899 4955 scope.go:117] "RemoveContainer" containerID="59410cf5393a924855abb129a83f4b3aca3c339a70233c2dad7476ca5d85fad8" Feb 17 13:58:50 crc kubenswrapper[4955]: I0217 13:58:50.717737 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kzfmk"] Feb 17 13:58:50 crc kubenswrapper[4955]: I0217 13:58:50.753338 4955 scope.go:117] "RemoveContainer" containerID="955841fd3a9aaf41f8536e73b49561ed53741e28be4a2ed98f009b33e906b467" Feb 17 13:58:50 crc kubenswrapper[4955]: E0217 13:58:50.753991 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"955841fd3a9aaf41f8536e73b49561ed53741e28be4a2ed98f009b33e906b467\": container with ID starting with 955841fd3a9aaf41f8536e73b49561ed53741e28be4a2ed98f009b33e906b467 not found: ID does not exist" containerID="955841fd3a9aaf41f8536e73b49561ed53741e28be4a2ed98f009b33e906b467" Feb 17 13:58:50 crc kubenswrapper[4955]: I0217 13:58:50.754049 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"955841fd3a9aaf41f8536e73b49561ed53741e28be4a2ed98f009b33e906b467"} err="failed to get container status \"955841fd3a9aaf41f8536e73b49561ed53741e28be4a2ed98f009b33e906b467\": rpc error: code = NotFound desc = could not find container \"955841fd3a9aaf41f8536e73b49561ed53741e28be4a2ed98f009b33e906b467\": container with ID starting with 955841fd3a9aaf41f8536e73b49561ed53741e28be4a2ed98f009b33e906b467 not found: ID does not exist" Feb 17 13:58:50 crc kubenswrapper[4955]: I0217 13:58:50.754077 4955 scope.go:117] "RemoveContainer" containerID="4a0e7d54c7368f13336ab0ea8fa18583001ba1768eebb7fb048d1d8d56d5d04f" Feb 17 13:58:50 crc kubenswrapper[4955]: E0217 13:58:50.754762 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a0e7d54c7368f13336ab0ea8fa18583001ba1768eebb7fb048d1d8d56d5d04f\": container with ID starting with 4a0e7d54c7368f13336ab0ea8fa18583001ba1768eebb7fb048d1d8d56d5d04f not found: ID does not exist" containerID="4a0e7d54c7368f13336ab0ea8fa18583001ba1768eebb7fb048d1d8d56d5d04f" Feb 17 13:58:50 crc kubenswrapper[4955]: I0217 13:58:50.754956 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a0e7d54c7368f13336ab0ea8fa18583001ba1768eebb7fb048d1d8d56d5d04f"} err="failed to get container status \"4a0e7d54c7368f13336ab0ea8fa18583001ba1768eebb7fb048d1d8d56d5d04f\": rpc error: code = NotFound desc = could not find container \"4a0e7d54c7368f13336ab0ea8fa18583001ba1768eebb7fb048d1d8d56d5d04f\": container with ID starting with 4a0e7d54c7368f13336ab0ea8fa18583001ba1768eebb7fb048d1d8d56d5d04f not found: ID does not exist" Feb 17 13:58:50 crc kubenswrapper[4955]: I0217 13:58:50.754980 4955 scope.go:117] "RemoveContainer" containerID="59410cf5393a924855abb129a83f4b3aca3c339a70233c2dad7476ca5d85fad8" Feb 17 13:58:50 crc kubenswrapper[4955]: E0217 13:58:50.755319 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59410cf5393a924855abb129a83f4b3aca3c339a70233c2dad7476ca5d85fad8\": container with ID starting with 59410cf5393a924855abb129a83f4b3aca3c339a70233c2dad7476ca5d85fad8 not found: ID does not exist" containerID="59410cf5393a924855abb129a83f4b3aca3c339a70233c2dad7476ca5d85fad8" Feb 17 13:58:50 crc kubenswrapper[4955]: I0217 13:58:50.755347 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59410cf5393a924855abb129a83f4b3aca3c339a70233c2dad7476ca5d85fad8"} err="failed to get container status \"59410cf5393a924855abb129a83f4b3aca3c339a70233c2dad7476ca5d85fad8\": rpc error: code = NotFound desc = could not find container \"59410cf5393a924855abb129a83f4b3aca3c339a70233c2dad7476ca5d85fad8\": container with ID starting with 59410cf5393a924855abb129a83f4b3aca3c339a70233c2dad7476ca5d85fad8 not found: ID does not exist" Feb 17 13:58:52 crc kubenswrapper[4955]: I0217 13:58:52.234414 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bdcb9ea-c5c7-4893-9067-b1a17072dbe8" path="/var/lib/kubelet/pods/5bdcb9ea-c5c7-4893-9067-b1a17072dbe8/volumes" Feb 17 13:59:27 crc kubenswrapper[4955]: I0217 13:59:27.977914 4955 generic.go:334] "Generic (PLEG): container finished" podID="325195a8-36d8-4ec8-9af3-ce87ad94af40" containerID="240783d5b2b351a103df8937664ab4a362aa27e4e7f0deeae2dc9febc23ed024" exitCode=0 Feb 17 13:59:27 crc kubenswrapper[4955]: I0217 13:59:27.978009 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b5tlj/crc-debug-98j4z" event={"ID":"325195a8-36d8-4ec8-9af3-ce87ad94af40","Type":"ContainerDied","Data":"240783d5b2b351a103df8937664ab4a362aa27e4e7f0deeae2dc9febc23ed024"} Feb 17 13:59:28 crc kubenswrapper[4955]: I0217 13:59:28.852727 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-z65ht"] Feb 17 13:59:28 crc kubenswrapper[4955]: E0217 13:59:28.853547 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bdcb9ea-c5c7-4893-9067-b1a17072dbe8" containerName="extract-content" Feb 17 13:59:28 crc kubenswrapper[4955]: I0217 13:59:28.853571 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bdcb9ea-c5c7-4893-9067-b1a17072dbe8" containerName="extract-content" Feb 17 13:59:28 crc kubenswrapper[4955]: E0217 13:59:28.853589 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bdcb9ea-c5c7-4893-9067-b1a17072dbe8" containerName="registry-server" Feb 17 13:59:28 crc kubenswrapper[4955]: I0217 13:59:28.853598 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bdcb9ea-c5c7-4893-9067-b1a17072dbe8" containerName="registry-server" Feb 17 13:59:28 crc kubenswrapper[4955]: E0217 13:59:28.853612 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bdcb9ea-c5c7-4893-9067-b1a17072dbe8" containerName="extract-utilities" Feb 17 13:59:28 crc kubenswrapper[4955]: I0217 13:59:28.853620 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bdcb9ea-c5c7-4893-9067-b1a17072dbe8" containerName="extract-utilities" Feb 17 13:59:28 crc kubenswrapper[4955]: I0217 13:59:28.853882 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bdcb9ea-c5c7-4893-9067-b1a17072dbe8" containerName="registry-server" Feb 17 13:59:28 crc kubenswrapper[4955]: I0217 13:59:28.855504 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z65ht" Feb 17 13:59:28 crc kubenswrapper[4955]: I0217 13:59:28.920850 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-z65ht"] Feb 17 13:59:29 crc kubenswrapper[4955]: I0217 13:59:29.025657 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6e6847c-4d47-4ea9-9d4e-40b2815616d5-utilities\") pod \"community-operators-z65ht\" (UID: \"a6e6847c-4d47-4ea9-9d4e-40b2815616d5\") " pod="openshift-marketplace/community-operators-z65ht" Feb 17 13:59:29 crc kubenswrapper[4955]: I0217 13:59:29.025759 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6e6847c-4d47-4ea9-9d4e-40b2815616d5-catalog-content\") pod \"community-operators-z65ht\" (UID: \"a6e6847c-4d47-4ea9-9d4e-40b2815616d5\") " pod="openshift-marketplace/community-operators-z65ht" Feb 17 13:59:29 crc kubenswrapper[4955]: I0217 13:59:29.025819 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djdrd\" (UniqueName: \"kubernetes.io/projected/a6e6847c-4d47-4ea9-9d4e-40b2815616d5-kube-api-access-djdrd\") pod \"community-operators-z65ht\" (UID: \"a6e6847c-4d47-4ea9-9d4e-40b2815616d5\") " pod="openshift-marketplace/community-operators-z65ht" Feb 17 13:59:29 crc kubenswrapper[4955]: I0217 13:59:29.100227 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b5tlj/crc-debug-98j4z" Feb 17 13:59:29 crc kubenswrapper[4955]: I0217 13:59:29.127218 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/325195a8-36d8-4ec8-9af3-ce87ad94af40-host\") pod \"325195a8-36d8-4ec8-9af3-ce87ad94af40\" (UID: \"325195a8-36d8-4ec8-9af3-ce87ad94af40\") " Feb 17 13:59:29 crc kubenswrapper[4955]: I0217 13:59:29.127278 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svk74\" (UniqueName: \"kubernetes.io/projected/325195a8-36d8-4ec8-9af3-ce87ad94af40-kube-api-access-svk74\") pod \"325195a8-36d8-4ec8-9af3-ce87ad94af40\" (UID: \"325195a8-36d8-4ec8-9af3-ce87ad94af40\") " Feb 17 13:59:29 crc kubenswrapper[4955]: I0217 13:59:29.127437 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6e6847c-4d47-4ea9-9d4e-40b2815616d5-catalog-content\") pod \"community-operators-z65ht\" (UID: \"a6e6847c-4d47-4ea9-9d4e-40b2815616d5\") " pod="openshift-marketplace/community-operators-z65ht" Feb 17 13:59:29 crc kubenswrapper[4955]: I0217 13:59:29.127463 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djdrd\" (UniqueName: \"kubernetes.io/projected/a6e6847c-4d47-4ea9-9d4e-40b2815616d5-kube-api-access-djdrd\") pod \"community-operators-z65ht\" (UID: \"a6e6847c-4d47-4ea9-9d4e-40b2815616d5\") " pod="openshift-marketplace/community-operators-z65ht" Feb 17 13:59:29 crc kubenswrapper[4955]: I0217 13:59:29.127584 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6e6847c-4d47-4ea9-9d4e-40b2815616d5-utilities\") pod \"community-operators-z65ht\" (UID: \"a6e6847c-4d47-4ea9-9d4e-40b2815616d5\") " pod="openshift-marketplace/community-operators-z65ht" Feb 17 13:59:29 crc kubenswrapper[4955]: I0217 13:59:29.128150 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6e6847c-4d47-4ea9-9d4e-40b2815616d5-utilities\") pod \"community-operators-z65ht\" (UID: \"a6e6847c-4d47-4ea9-9d4e-40b2815616d5\") " pod="openshift-marketplace/community-operators-z65ht" Feb 17 13:59:29 crc kubenswrapper[4955]: I0217 13:59:29.128197 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/325195a8-36d8-4ec8-9af3-ce87ad94af40-host" (OuterVolumeSpecName: "host") pod "325195a8-36d8-4ec8-9af3-ce87ad94af40" (UID: "325195a8-36d8-4ec8-9af3-ce87ad94af40"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 13:59:29 crc kubenswrapper[4955]: I0217 13:59:29.129040 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6e6847c-4d47-4ea9-9d4e-40b2815616d5-catalog-content\") pod \"community-operators-z65ht\" (UID: \"a6e6847c-4d47-4ea9-9d4e-40b2815616d5\") " pod="openshift-marketplace/community-operators-z65ht" Feb 17 13:59:29 crc kubenswrapper[4955]: I0217 13:59:29.143003 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/325195a8-36d8-4ec8-9af3-ce87ad94af40-kube-api-access-svk74" (OuterVolumeSpecName: "kube-api-access-svk74") pod "325195a8-36d8-4ec8-9af3-ce87ad94af40" (UID: "325195a8-36d8-4ec8-9af3-ce87ad94af40"). InnerVolumeSpecName "kube-api-access-svk74". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:59:29 crc kubenswrapper[4955]: I0217 13:59:29.157870 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djdrd\" (UniqueName: \"kubernetes.io/projected/a6e6847c-4d47-4ea9-9d4e-40b2815616d5-kube-api-access-djdrd\") pod \"community-operators-z65ht\" (UID: \"a6e6847c-4d47-4ea9-9d4e-40b2815616d5\") " pod="openshift-marketplace/community-operators-z65ht" Feb 17 13:59:29 crc kubenswrapper[4955]: I0217 13:59:29.168034 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-b5tlj/crc-debug-98j4z"] Feb 17 13:59:29 crc kubenswrapper[4955]: I0217 13:59:29.176449 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-b5tlj/crc-debug-98j4z"] Feb 17 13:59:29 crc kubenswrapper[4955]: I0217 13:59:29.229017 4955 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/325195a8-36d8-4ec8-9af3-ce87ad94af40-host\") on node \"crc\" DevicePath \"\"" Feb 17 13:59:29 crc kubenswrapper[4955]: I0217 13:59:29.229055 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svk74\" (UniqueName: \"kubernetes.io/projected/325195a8-36d8-4ec8-9af3-ce87ad94af40-kube-api-access-svk74\") on node \"crc\" DevicePath \"\"" Feb 17 13:59:29 crc kubenswrapper[4955]: I0217 13:59:29.241746 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z65ht" Feb 17 13:59:29 crc kubenswrapper[4955]: I0217 13:59:29.769457 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-z65ht"] Feb 17 13:59:30 crc kubenswrapper[4955]: I0217 13:59:30.007520 4955 generic.go:334] "Generic (PLEG): container finished" podID="a6e6847c-4d47-4ea9-9d4e-40b2815616d5" containerID="be39ed15f2cbdfe55b9b8cdd9b0b22452fc73c5c45f05c45a02d454baf9844f6" exitCode=0 Feb 17 13:59:30 crc kubenswrapper[4955]: I0217 13:59:30.007602 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z65ht" event={"ID":"a6e6847c-4d47-4ea9-9d4e-40b2815616d5","Type":"ContainerDied","Data":"be39ed15f2cbdfe55b9b8cdd9b0b22452fc73c5c45f05c45a02d454baf9844f6"} Feb 17 13:59:30 crc kubenswrapper[4955]: I0217 13:59:30.008026 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z65ht" event={"ID":"a6e6847c-4d47-4ea9-9d4e-40b2815616d5","Type":"ContainerStarted","Data":"c359d3a1740b68ff65c4aac1a3f41ef0feeb6409724a9b391ad71f5608f0c36a"} Feb 17 13:59:30 crc kubenswrapper[4955]: I0217 13:59:30.009888 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1acbf02e5ef827881b13b117c1ab89775ae25ee5416a47df61008f1b83a6667a" Feb 17 13:59:30 crc kubenswrapper[4955]: I0217 13:59:30.009949 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b5tlj/crc-debug-98j4z" Feb 17 13:59:30 crc kubenswrapper[4955]: I0217 13:59:30.241432 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="325195a8-36d8-4ec8-9af3-ce87ad94af40" path="/var/lib/kubelet/pods/325195a8-36d8-4ec8-9af3-ce87ad94af40/volumes" Feb 17 13:59:30 crc kubenswrapper[4955]: I0217 13:59:30.362927 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-b5tlj/crc-debug-mtnvz"] Feb 17 13:59:30 crc kubenswrapper[4955]: E0217 13:59:30.363337 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="325195a8-36d8-4ec8-9af3-ce87ad94af40" containerName="container-00" Feb 17 13:59:30 crc kubenswrapper[4955]: I0217 13:59:30.363357 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="325195a8-36d8-4ec8-9af3-ce87ad94af40" containerName="container-00" Feb 17 13:59:30 crc kubenswrapper[4955]: I0217 13:59:30.363517 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="325195a8-36d8-4ec8-9af3-ce87ad94af40" containerName="container-00" Feb 17 13:59:30 crc kubenswrapper[4955]: I0217 13:59:30.364138 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b5tlj/crc-debug-mtnvz" Feb 17 13:59:30 crc kubenswrapper[4955]: I0217 13:59:30.366138 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-b5tlj"/"default-dockercfg-56j4v" Feb 17 13:59:30 crc kubenswrapper[4955]: I0217 13:59:30.452097 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fmj2\" (UniqueName: \"kubernetes.io/projected/d95fbb22-e76c-4f62-a19f-0dfcdff08376-kube-api-access-8fmj2\") pod \"crc-debug-mtnvz\" (UID: \"d95fbb22-e76c-4f62-a19f-0dfcdff08376\") " pod="openshift-must-gather-b5tlj/crc-debug-mtnvz" Feb 17 13:59:30 crc kubenswrapper[4955]: I0217 13:59:30.452218 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d95fbb22-e76c-4f62-a19f-0dfcdff08376-host\") pod \"crc-debug-mtnvz\" (UID: \"d95fbb22-e76c-4f62-a19f-0dfcdff08376\") " pod="openshift-must-gather-b5tlj/crc-debug-mtnvz" Feb 17 13:59:30 crc kubenswrapper[4955]: I0217 13:59:30.553563 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fmj2\" (UniqueName: \"kubernetes.io/projected/d95fbb22-e76c-4f62-a19f-0dfcdff08376-kube-api-access-8fmj2\") pod \"crc-debug-mtnvz\" (UID: \"d95fbb22-e76c-4f62-a19f-0dfcdff08376\") " pod="openshift-must-gather-b5tlj/crc-debug-mtnvz" Feb 17 13:59:30 crc kubenswrapper[4955]: I0217 13:59:30.553632 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d95fbb22-e76c-4f62-a19f-0dfcdff08376-host\") pod \"crc-debug-mtnvz\" (UID: \"d95fbb22-e76c-4f62-a19f-0dfcdff08376\") " pod="openshift-must-gather-b5tlj/crc-debug-mtnvz" Feb 17 13:59:30 crc kubenswrapper[4955]: I0217 13:59:30.553828 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d95fbb22-e76c-4f62-a19f-0dfcdff08376-host\") pod \"crc-debug-mtnvz\" (UID: \"d95fbb22-e76c-4f62-a19f-0dfcdff08376\") " pod="openshift-must-gather-b5tlj/crc-debug-mtnvz" Feb 17 13:59:30 crc kubenswrapper[4955]: I0217 13:59:30.587588 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fmj2\" (UniqueName: \"kubernetes.io/projected/d95fbb22-e76c-4f62-a19f-0dfcdff08376-kube-api-access-8fmj2\") pod \"crc-debug-mtnvz\" (UID: \"d95fbb22-e76c-4f62-a19f-0dfcdff08376\") " pod="openshift-must-gather-b5tlj/crc-debug-mtnvz" Feb 17 13:59:30 crc kubenswrapper[4955]: I0217 13:59:30.680853 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b5tlj/crc-debug-mtnvz" Feb 17 13:59:30 crc kubenswrapper[4955]: W0217 13:59:30.710659 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd95fbb22_e76c_4f62_a19f_0dfcdff08376.slice/crio-9050c84d983f16624e7db20bb588d99cc2ee60e1276e5f066e20235cd8a057cc WatchSource:0}: Error finding container 9050c84d983f16624e7db20bb588d99cc2ee60e1276e5f066e20235cd8a057cc: Status 404 returned error can't find the container with id 9050c84d983f16624e7db20bb588d99cc2ee60e1276e5f066e20235cd8a057cc Feb 17 13:59:31 crc kubenswrapper[4955]: I0217 13:59:31.021964 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z65ht" event={"ID":"a6e6847c-4d47-4ea9-9d4e-40b2815616d5","Type":"ContainerStarted","Data":"78a9c6e5b837e33dc30c257ce6b868340cd343bd12a7527c17fbf126d027ff4d"} Feb 17 13:59:31 crc kubenswrapper[4955]: I0217 13:59:31.024364 4955 generic.go:334] "Generic (PLEG): container finished" podID="d95fbb22-e76c-4f62-a19f-0dfcdff08376" containerID="fa7d2d77ca0542578eeaa0b311b0e50ae690984af7919c73eb7090df20ea0c78" exitCode=0 Feb 17 13:59:31 crc kubenswrapper[4955]: I0217 13:59:31.024399 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b5tlj/crc-debug-mtnvz" event={"ID":"d95fbb22-e76c-4f62-a19f-0dfcdff08376","Type":"ContainerDied","Data":"fa7d2d77ca0542578eeaa0b311b0e50ae690984af7919c73eb7090df20ea0c78"} Feb 17 13:59:31 crc kubenswrapper[4955]: I0217 13:59:31.024420 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b5tlj/crc-debug-mtnvz" event={"ID":"d95fbb22-e76c-4f62-a19f-0dfcdff08376","Type":"ContainerStarted","Data":"9050c84d983f16624e7db20bb588d99cc2ee60e1276e5f066e20235cd8a057cc"} Feb 17 13:59:31 crc kubenswrapper[4955]: I0217 13:59:31.488050 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-b5tlj/crc-debug-mtnvz"] Feb 17 13:59:31 crc kubenswrapper[4955]: I0217 13:59:31.506540 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-b5tlj/crc-debug-mtnvz"] Feb 17 13:59:32 crc kubenswrapper[4955]: I0217 13:59:32.036210 4955 generic.go:334] "Generic (PLEG): container finished" podID="a6e6847c-4d47-4ea9-9d4e-40b2815616d5" containerID="78a9c6e5b837e33dc30c257ce6b868340cd343bd12a7527c17fbf126d027ff4d" exitCode=0 Feb 17 13:59:32 crc kubenswrapper[4955]: I0217 13:59:32.037696 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z65ht" event={"ID":"a6e6847c-4d47-4ea9-9d4e-40b2815616d5","Type":"ContainerDied","Data":"78a9c6e5b837e33dc30c257ce6b868340cd343bd12a7527c17fbf126d027ff4d"} Feb 17 13:59:32 crc kubenswrapper[4955]: I0217 13:59:32.167858 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b5tlj/crc-debug-mtnvz" Feb 17 13:59:32 crc kubenswrapper[4955]: I0217 13:59:32.284388 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d95fbb22-e76c-4f62-a19f-0dfcdff08376-host\") pod \"d95fbb22-e76c-4f62-a19f-0dfcdff08376\" (UID: \"d95fbb22-e76c-4f62-a19f-0dfcdff08376\") " Feb 17 13:59:32 crc kubenswrapper[4955]: I0217 13:59:32.284478 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8fmj2\" (UniqueName: \"kubernetes.io/projected/d95fbb22-e76c-4f62-a19f-0dfcdff08376-kube-api-access-8fmj2\") pod \"d95fbb22-e76c-4f62-a19f-0dfcdff08376\" (UID: \"d95fbb22-e76c-4f62-a19f-0dfcdff08376\") " Feb 17 13:59:32 crc kubenswrapper[4955]: I0217 13:59:32.284530 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d95fbb22-e76c-4f62-a19f-0dfcdff08376-host" (OuterVolumeSpecName: "host") pod "d95fbb22-e76c-4f62-a19f-0dfcdff08376" (UID: "d95fbb22-e76c-4f62-a19f-0dfcdff08376"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 13:59:32 crc kubenswrapper[4955]: I0217 13:59:32.285216 4955 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d95fbb22-e76c-4f62-a19f-0dfcdff08376-host\") on node \"crc\" DevicePath \"\"" Feb 17 13:59:32 crc kubenswrapper[4955]: I0217 13:59:32.297134 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d95fbb22-e76c-4f62-a19f-0dfcdff08376-kube-api-access-8fmj2" (OuterVolumeSpecName: "kube-api-access-8fmj2") pod "d95fbb22-e76c-4f62-a19f-0dfcdff08376" (UID: "d95fbb22-e76c-4f62-a19f-0dfcdff08376"). InnerVolumeSpecName "kube-api-access-8fmj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:59:32 crc kubenswrapper[4955]: I0217 13:59:32.387348 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8fmj2\" (UniqueName: \"kubernetes.io/projected/d95fbb22-e76c-4f62-a19f-0dfcdff08376-kube-api-access-8fmj2\") on node \"crc\" DevicePath \"\"" Feb 17 13:59:32 crc kubenswrapper[4955]: I0217 13:59:32.700891 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-b5tlj/crc-debug-4vqcg"] Feb 17 13:59:32 crc kubenswrapper[4955]: E0217 13:59:32.701248 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d95fbb22-e76c-4f62-a19f-0dfcdff08376" containerName="container-00" Feb 17 13:59:32 crc kubenswrapper[4955]: I0217 13:59:32.701259 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="d95fbb22-e76c-4f62-a19f-0dfcdff08376" containerName="container-00" Feb 17 13:59:32 crc kubenswrapper[4955]: I0217 13:59:32.701477 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="d95fbb22-e76c-4f62-a19f-0dfcdff08376" containerName="container-00" Feb 17 13:59:32 crc kubenswrapper[4955]: I0217 13:59:32.702081 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b5tlj/crc-debug-4vqcg" Feb 17 13:59:32 crc kubenswrapper[4955]: I0217 13:59:32.896447 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2e0a7b49-8fab-4ad9-b4f1-3f3cd4a143f2-host\") pod \"crc-debug-4vqcg\" (UID: \"2e0a7b49-8fab-4ad9-b4f1-3f3cd4a143f2\") " pod="openshift-must-gather-b5tlj/crc-debug-4vqcg" Feb 17 13:59:32 crc kubenswrapper[4955]: I0217 13:59:32.896535 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9wsj\" (UniqueName: \"kubernetes.io/projected/2e0a7b49-8fab-4ad9-b4f1-3f3cd4a143f2-kube-api-access-q9wsj\") pod \"crc-debug-4vqcg\" (UID: \"2e0a7b49-8fab-4ad9-b4f1-3f3cd4a143f2\") " pod="openshift-must-gather-b5tlj/crc-debug-4vqcg" Feb 17 13:59:32 crc kubenswrapper[4955]: I0217 13:59:32.998361 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9wsj\" (UniqueName: \"kubernetes.io/projected/2e0a7b49-8fab-4ad9-b4f1-3f3cd4a143f2-kube-api-access-q9wsj\") pod \"crc-debug-4vqcg\" (UID: \"2e0a7b49-8fab-4ad9-b4f1-3f3cd4a143f2\") " pod="openshift-must-gather-b5tlj/crc-debug-4vqcg" Feb 17 13:59:32 crc kubenswrapper[4955]: I0217 13:59:32.998611 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2e0a7b49-8fab-4ad9-b4f1-3f3cd4a143f2-host\") pod \"crc-debug-4vqcg\" (UID: \"2e0a7b49-8fab-4ad9-b4f1-3f3cd4a143f2\") " pod="openshift-must-gather-b5tlj/crc-debug-4vqcg" Feb 17 13:59:32 crc kubenswrapper[4955]: I0217 13:59:32.998768 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2e0a7b49-8fab-4ad9-b4f1-3f3cd4a143f2-host\") pod \"crc-debug-4vqcg\" (UID: \"2e0a7b49-8fab-4ad9-b4f1-3f3cd4a143f2\") " pod="openshift-must-gather-b5tlj/crc-debug-4vqcg" Feb 17 13:59:33 crc kubenswrapper[4955]: I0217 13:59:33.027017 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9wsj\" (UniqueName: \"kubernetes.io/projected/2e0a7b49-8fab-4ad9-b4f1-3f3cd4a143f2-kube-api-access-q9wsj\") pod \"crc-debug-4vqcg\" (UID: \"2e0a7b49-8fab-4ad9-b4f1-3f3cd4a143f2\") " pod="openshift-must-gather-b5tlj/crc-debug-4vqcg" Feb 17 13:59:33 crc kubenswrapper[4955]: I0217 13:59:33.060302 4955 scope.go:117] "RemoveContainer" containerID="fa7d2d77ca0542578eeaa0b311b0e50ae690984af7919c73eb7090df20ea0c78" Feb 17 13:59:33 crc kubenswrapper[4955]: I0217 13:59:33.060350 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b5tlj/crc-debug-mtnvz" Feb 17 13:59:33 crc kubenswrapper[4955]: I0217 13:59:33.320103 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b5tlj/crc-debug-4vqcg" Feb 17 13:59:33 crc kubenswrapper[4955]: W0217 13:59:33.347703 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e0a7b49_8fab_4ad9_b4f1_3f3cd4a143f2.slice/crio-0c1f868f87fe7cad093f07c9314808bb775e7d2f2eed7332f112c15a7a5a03e3 WatchSource:0}: Error finding container 0c1f868f87fe7cad093f07c9314808bb775e7d2f2eed7332f112c15a7a5a03e3: Status 404 returned error can't find the container with id 0c1f868f87fe7cad093f07c9314808bb775e7d2f2eed7332f112c15a7a5a03e3 Feb 17 13:59:34 crc kubenswrapper[4955]: I0217 13:59:34.072418 4955 generic.go:334] "Generic (PLEG): container finished" podID="2e0a7b49-8fab-4ad9-b4f1-3f3cd4a143f2" containerID="caae994650cc144b8829f5df560f80a75771695d4367e4dbd3dfd6af0d894ee6" exitCode=0 Feb 17 13:59:34 crc kubenswrapper[4955]: I0217 13:59:34.072522 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b5tlj/crc-debug-4vqcg" event={"ID":"2e0a7b49-8fab-4ad9-b4f1-3f3cd4a143f2","Type":"ContainerDied","Data":"caae994650cc144b8829f5df560f80a75771695d4367e4dbd3dfd6af0d894ee6"} Feb 17 13:59:34 crc kubenswrapper[4955]: I0217 13:59:34.073495 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b5tlj/crc-debug-4vqcg" event={"ID":"2e0a7b49-8fab-4ad9-b4f1-3f3cd4a143f2","Type":"ContainerStarted","Data":"0c1f868f87fe7cad093f07c9314808bb775e7d2f2eed7332f112c15a7a5a03e3"} Feb 17 13:59:34 crc kubenswrapper[4955]: I0217 13:59:34.075200 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z65ht" event={"ID":"a6e6847c-4d47-4ea9-9d4e-40b2815616d5","Type":"ContainerStarted","Data":"cf3d3367709c8031630e8b650bbf3eae1f9a1aa7de339740de1a86f62609a645"} Feb 17 13:59:34 crc kubenswrapper[4955]: I0217 13:59:34.117832 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-b5tlj/crc-debug-4vqcg"] Feb 17 13:59:34 crc kubenswrapper[4955]: I0217 13:59:34.125061 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-z65ht" podStartSLOduration=3.123574212 podStartE2EDuration="6.125044773s" podCreationTimestamp="2026-02-17 13:59:28 +0000 UTC" firstStartedPulling="2026-02-17 13:59:30.009381906 +0000 UTC m=+3308.532111449" lastFinishedPulling="2026-02-17 13:59:33.010852467 +0000 UTC m=+3311.533582010" observedRunningTime="2026-02-17 13:59:34.109029215 +0000 UTC m=+3312.631758758" watchObservedRunningTime="2026-02-17 13:59:34.125044773 +0000 UTC m=+3312.647774316" Feb 17 13:59:34 crc kubenswrapper[4955]: I0217 13:59:34.139332 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-b5tlj/crc-debug-4vqcg"] Feb 17 13:59:34 crc kubenswrapper[4955]: I0217 13:59:34.233571 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d95fbb22-e76c-4f62-a19f-0dfcdff08376" path="/var/lib/kubelet/pods/d95fbb22-e76c-4f62-a19f-0dfcdff08376/volumes" Feb 17 13:59:35 crc kubenswrapper[4955]: I0217 13:59:35.185573 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b5tlj/crc-debug-4vqcg" Feb 17 13:59:35 crc kubenswrapper[4955]: I0217 13:59:35.341713 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9wsj\" (UniqueName: \"kubernetes.io/projected/2e0a7b49-8fab-4ad9-b4f1-3f3cd4a143f2-kube-api-access-q9wsj\") pod \"2e0a7b49-8fab-4ad9-b4f1-3f3cd4a143f2\" (UID: \"2e0a7b49-8fab-4ad9-b4f1-3f3cd4a143f2\") " Feb 17 13:59:35 crc kubenswrapper[4955]: I0217 13:59:35.341974 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2e0a7b49-8fab-4ad9-b4f1-3f3cd4a143f2-host\") pod \"2e0a7b49-8fab-4ad9-b4f1-3f3cd4a143f2\" (UID: \"2e0a7b49-8fab-4ad9-b4f1-3f3cd4a143f2\") " Feb 17 13:59:35 crc kubenswrapper[4955]: I0217 13:59:35.342592 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2e0a7b49-8fab-4ad9-b4f1-3f3cd4a143f2-host" (OuterVolumeSpecName: "host") pod "2e0a7b49-8fab-4ad9-b4f1-3f3cd4a143f2" (UID: "2e0a7b49-8fab-4ad9-b4f1-3f3cd4a143f2"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 13:59:35 crc kubenswrapper[4955]: I0217 13:59:35.344517 4955 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2e0a7b49-8fab-4ad9-b4f1-3f3cd4a143f2-host\") on node \"crc\" DevicePath \"\"" Feb 17 13:59:35 crc kubenswrapper[4955]: I0217 13:59:35.361684 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e0a7b49-8fab-4ad9-b4f1-3f3cd4a143f2-kube-api-access-q9wsj" (OuterVolumeSpecName: "kube-api-access-q9wsj") pod "2e0a7b49-8fab-4ad9-b4f1-3f3cd4a143f2" (UID: "2e0a7b49-8fab-4ad9-b4f1-3f3cd4a143f2"). InnerVolumeSpecName "kube-api-access-q9wsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:59:35 crc kubenswrapper[4955]: I0217 13:59:35.446090 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9wsj\" (UniqueName: \"kubernetes.io/projected/2e0a7b49-8fab-4ad9-b4f1-3f3cd4a143f2-kube-api-access-q9wsj\") on node \"crc\" DevicePath \"\"" Feb 17 13:59:36 crc kubenswrapper[4955]: I0217 13:59:36.104591 4955 scope.go:117] "RemoveContainer" containerID="caae994650cc144b8829f5df560f80a75771695d4367e4dbd3dfd6af0d894ee6" Feb 17 13:59:36 crc kubenswrapper[4955]: I0217 13:59:36.105228 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b5tlj/crc-debug-4vqcg" Feb 17 13:59:36 crc kubenswrapper[4955]: I0217 13:59:36.236307 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e0a7b49-8fab-4ad9-b4f1-3f3cd4a143f2" path="/var/lib/kubelet/pods/2e0a7b49-8fab-4ad9-b4f1-3f3cd4a143f2/volumes" Feb 17 13:59:39 crc kubenswrapper[4955]: I0217 13:59:39.242601 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-z65ht" Feb 17 13:59:39 crc kubenswrapper[4955]: I0217 13:59:39.243209 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-z65ht" Feb 17 13:59:39 crc kubenswrapper[4955]: I0217 13:59:39.286821 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-z65ht" Feb 17 13:59:40 crc kubenswrapper[4955]: I0217 13:59:40.187483 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-z65ht" Feb 17 13:59:41 crc kubenswrapper[4955]: I0217 13:59:41.449817 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-42t5x"] Feb 17 13:59:41 crc kubenswrapper[4955]: E0217 13:59:41.450650 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e0a7b49-8fab-4ad9-b4f1-3f3cd4a143f2" containerName="container-00" Feb 17 13:59:41 crc kubenswrapper[4955]: I0217 13:59:41.450669 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e0a7b49-8fab-4ad9-b4f1-3f3cd4a143f2" containerName="container-00" Feb 17 13:59:41 crc kubenswrapper[4955]: I0217 13:59:41.450936 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e0a7b49-8fab-4ad9-b4f1-3f3cd4a143f2" containerName="container-00" Feb 17 13:59:41 crc kubenswrapper[4955]: I0217 13:59:41.452607 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-42t5x" Feb 17 13:59:41 crc kubenswrapper[4955]: I0217 13:59:41.479187 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-42t5x"] Feb 17 13:59:41 crc kubenswrapper[4955]: I0217 13:59:41.569274 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48598b8d-cbb4-4999-859d-543aa8237c4b-utilities\") pod \"certified-operators-42t5x\" (UID: \"48598b8d-cbb4-4999-859d-543aa8237c4b\") " pod="openshift-marketplace/certified-operators-42t5x" Feb 17 13:59:41 crc kubenswrapper[4955]: I0217 13:59:41.569398 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6p7s\" (UniqueName: \"kubernetes.io/projected/48598b8d-cbb4-4999-859d-543aa8237c4b-kube-api-access-v6p7s\") pod \"certified-operators-42t5x\" (UID: \"48598b8d-cbb4-4999-859d-543aa8237c4b\") " pod="openshift-marketplace/certified-operators-42t5x" Feb 17 13:59:41 crc kubenswrapper[4955]: I0217 13:59:41.569426 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48598b8d-cbb4-4999-859d-543aa8237c4b-catalog-content\") pod \"certified-operators-42t5x\" (UID: \"48598b8d-cbb4-4999-859d-543aa8237c4b\") " pod="openshift-marketplace/certified-operators-42t5x" Feb 17 13:59:41 crc kubenswrapper[4955]: I0217 13:59:41.670964 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48598b8d-cbb4-4999-859d-543aa8237c4b-utilities\") pod \"certified-operators-42t5x\" (UID: \"48598b8d-cbb4-4999-859d-543aa8237c4b\") " pod="openshift-marketplace/certified-operators-42t5x" Feb 17 13:59:41 crc kubenswrapper[4955]: I0217 13:59:41.671054 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6p7s\" (UniqueName: \"kubernetes.io/projected/48598b8d-cbb4-4999-859d-543aa8237c4b-kube-api-access-v6p7s\") pod \"certified-operators-42t5x\" (UID: \"48598b8d-cbb4-4999-859d-543aa8237c4b\") " pod="openshift-marketplace/certified-operators-42t5x" Feb 17 13:59:41 crc kubenswrapper[4955]: I0217 13:59:41.671079 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48598b8d-cbb4-4999-859d-543aa8237c4b-catalog-content\") pod \"certified-operators-42t5x\" (UID: \"48598b8d-cbb4-4999-859d-543aa8237c4b\") " pod="openshift-marketplace/certified-operators-42t5x" Feb 17 13:59:41 crc kubenswrapper[4955]: I0217 13:59:41.671553 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48598b8d-cbb4-4999-859d-543aa8237c4b-catalog-content\") pod \"certified-operators-42t5x\" (UID: \"48598b8d-cbb4-4999-859d-543aa8237c4b\") " pod="openshift-marketplace/certified-operators-42t5x" Feb 17 13:59:41 crc kubenswrapper[4955]: I0217 13:59:41.671799 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48598b8d-cbb4-4999-859d-543aa8237c4b-utilities\") pod \"certified-operators-42t5x\" (UID: \"48598b8d-cbb4-4999-859d-543aa8237c4b\") " pod="openshift-marketplace/certified-operators-42t5x" Feb 17 13:59:41 crc kubenswrapper[4955]: I0217 13:59:41.698481 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6p7s\" (UniqueName: \"kubernetes.io/projected/48598b8d-cbb4-4999-859d-543aa8237c4b-kube-api-access-v6p7s\") pod \"certified-operators-42t5x\" (UID: \"48598b8d-cbb4-4999-859d-543aa8237c4b\") " pod="openshift-marketplace/certified-operators-42t5x" Feb 17 13:59:41 crc kubenswrapper[4955]: I0217 13:59:41.777370 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-42t5x" Feb 17 13:59:42 crc kubenswrapper[4955]: I0217 13:59:42.041481 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-z65ht"] Feb 17 13:59:42 crc kubenswrapper[4955]: I0217 13:59:42.187205 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-z65ht" podUID="a6e6847c-4d47-4ea9-9d4e-40b2815616d5" containerName="registry-server" containerID="cri-o://cf3d3367709c8031630e8b650bbf3eae1f9a1aa7de339740de1a86f62609a645" gracePeriod=2 Feb 17 13:59:42 crc kubenswrapper[4955]: I0217 13:59:42.346366 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-42t5x"] Feb 17 13:59:42 crc kubenswrapper[4955]: I0217 13:59:42.622828 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z65ht" Feb 17 13:59:42 crc kubenswrapper[4955]: I0217 13:59:42.793966 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-djdrd\" (UniqueName: \"kubernetes.io/projected/a6e6847c-4d47-4ea9-9d4e-40b2815616d5-kube-api-access-djdrd\") pod \"a6e6847c-4d47-4ea9-9d4e-40b2815616d5\" (UID: \"a6e6847c-4d47-4ea9-9d4e-40b2815616d5\") " Feb 17 13:59:42 crc kubenswrapper[4955]: I0217 13:59:42.794490 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6e6847c-4d47-4ea9-9d4e-40b2815616d5-catalog-content\") pod \"a6e6847c-4d47-4ea9-9d4e-40b2815616d5\" (UID: \"a6e6847c-4d47-4ea9-9d4e-40b2815616d5\") " Feb 17 13:59:42 crc kubenswrapper[4955]: I0217 13:59:42.794645 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6e6847c-4d47-4ea9-9d4e-40b2815616d5-utilities\") pod \"a6e6847c-4d47-4ea9-9d4e-40b2815616d5\" (UID: \"a6e6847c-4d47-4ea9-9d4e-40b2815616d5\") " Feb 17 13:59:42 crc kubenswrapper[4955]: I0217 13:59:42.795469 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6e6847c-4d47-4ea9-9d4e-40b2815616d5-utilities" (OuterVolumeSpecName: "utilities") pod "a6e6847c-4d47-4ea9-9d4e-40b2815616d5" (UID: "a6e6847c-4d47-4ea9-9d4e-40b2815616d5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:59:42 crc kubenswrapper[4955]: I0217 13:59:42.803132 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6e6847c-4d47-4ea9-9d4e-40b2815616d5-kube-api-access-djdrd" (OuterVolumeSpecName: "kube-api-access-djdrd") pod "a6e6847c-4d47-4ea9-9d4e-40b2815616d5" (UID: "a6e6847c-4d47-4ea9-9d4e-40b2815616d5"). InnerVolumeSpecName "kube-api-access-djdrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:59:42 crc kubenswrapper[4955]: I0217 13:59:42.849053 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6e6847c-4d47-4ea9-9d4e-40b2815616d5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a6e6847c-4d47-4ea9-9d4e-40b2815616d5" (UID: "a6e6847c-4d47-4ea9-9d4e-40b2815616d5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:59:42 crc kubenswrapper[4955]: I0217 13:59:42.896840 4955 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6e6847c-4d47-4ea9-9d4e-40b2815616d5-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 13:59:42 crc kubenswrapper[4955]: I0217 13:59:42.896900 4955 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6e6847c-4d47-4ea9-9d4e-40b2815616d5-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 13:59:42 crc kubenswrapper[4955]: I0217 13:59:42.896936 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-djdrd\" (UniqueName: \"kubernetes.io/projected/a6e6847c-4d47-4ea9-9d4e-40b2815616d5-kube-api-access-djdrd\") on node \"crc\" DevicePath \"\"" Feb 17 13:59:43 crc kubenswrapper[4955]: I0217 13:59:43.197421 4955 generic.go:334] "Generic (PLEG): container finished" podID="48598b8d-cbb4-4999-859d-543aa8237c4b" containerID="fb1ed7335ec31c3003123879dc062770ef1f1094eed4a5596b29841ad49580cf" exitCode=0 Feb 17 13:59:43 crc kubenswrapper[4955]: I0217 13:59:43.197534 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-42t5x" event={"ID":"48598b8d-cbb4-4999-859d-543aa8237c4b","Type":"ContainerDied","Data":"fb1ed7335ec31c3003123879dc062770ef1f1094eed4a5596b29841ad49580cf"} Feb 17 13:59:43 crc kubenswrapper[4955]: I0217 13:59:43.198050 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-42t5x" event={"ID":"48598b8d-cbb4-4999-859d-543aa8237c4b","Type":"ContainerStarted","Data":"e57b70cdc3188683a1eb41df9acb2abd351c5d454b7fdab04d8de191a2c9c60d"} Feb 17 13:59:43 crc kubenswrapper[4955]: I0217 13:59:43.200654 4955 generic.go:334] "Generic (PLEG): container finished" podID="a6e6847c-4d47-4ea9-9d4e-40b2815616d5" containerID="cf3d3367709c8031630e8b650bbf3eae1f9a1aa7de339740de1a86f62609a645" exitCode=0 Feb 17 13:59:43 crc kubenswrapper[4955]: I0217 13:59:43.200714 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z65ht" event={"ID":"a6e6847c-4d47-4ea9-9d4e-40b2815616d5","Type":"ContainerDied","Data":"cf3d3367709c8031630e8b650bbf3eae1f9a1aa7de339740de1a86f62609a645"} Feb 17 13:59:43 crc kubenswrapper[4955]: I0217 13:59:43.200731 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z65ht" Feb 17 13:59:43 crc kubenswrapper[4955]: I0217 13:59:43.200763 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z65ht" event={"ID":"a6e6847c-4d47-4ea9-9d4e-40b2815616d5","Type":"ContainerDied","Data":"c359d3a1740b68ff65c4aac1a3f41ef0feeb6409724a9b391ad71f5608f0c36a"} Feb 17 13:59:43 crc kubenswrapper[4955]: I0217 13:59:43.200804 4955 scope.go:117] "RemoveContainer" containerID="cf3d3367709c8031630e8b650bbf3eae1f9a1aa7de339740de1a86f62609a645" Feb 17 13:59:43 crc kubenswrapper[4955]: I0217 13:59:43.227412 4955 scope.go:117] "RemoveContainer" containerID="78a9c6e5b837e33dc30c257ce6b868340cd343bd12a7527c17fbf126d027ff4d" Feb 17 13:59:43 crc kubenswrapper[4955]: I0217 13:59:43.258847 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-z65ht"] Feb 17 13:59:43 crc kubenswrapper[4955]: I0217 13:59:43.268521 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-z65ht"] Feb 17 13:59:43 crc kubenswrapper[4955]: I0217 13:59:43.283192 4955 scope.go:117] "RemoveContainer" containerID="be39ed15f2cbdfe55b9b8cdd9b0b22452fc73c5c45f05c45a02d454baf9844f6" Feb 17 13:59:43 crc kubenswrapper[4955]: I0217 13:59:43.308335 4955 scope.go:117] "RemoveContainer" containerID="cf3d3367709c8031630e8b650bbf3eae1f9a1aa7de339740de1a86f62609a645" Feb 17 13:59:43 crc kubenswrapper[4955]: E0217 13:59:43.308962 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf3d3367709c8031630e8b650bbf3eae1f9a1aa7de339740de1a86f62609a645\": container with ID starting with cf3d3367709c8031630e8b650bbf3eae1f9a1aa7de339740de1a86f62609a645 not found: ID does not exist" containerID="cf3d3367709c8031630e8b650bbf3eae1f9a1aa7de339740de1a86f62609a645" Feb 17 13:59:43 crc kubenswrapper[4955]: I0217 13:59:43.309024 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf3d3367709c8031630e8b650bbf3eae1f9a1aa7de339740de1a86f62609a645"} err="failed to get container status \"cf3d3367709c8031630e8b650bbf3eae1f9a1aa7de339740de1a86f62609a645\": rpc error: code = NotFound desc = could not find container \"cf3d3367709c8031630e8b650bbf3eae1f9a1aa7de339740de1a86f62609a645\": container with ID starting with cf3d3367709c8031630e8b650bbf3eae1f9a1aa7de339740de1a86f62609a645 not found: ID does not exist" Feb 17 13:59:43 crc kubenswrapper[4955]: I0217 13:59:43.309057 4955 scope.go:117] "RemoveContainer" containerID="78a9c6e5b837e33dc30c257ce6b868340cd343bd12a7527c17fbf126d027ff4d" Feb 17 13:59:43 crc kubenswrapper[4955]: E0217 13:59:43.309478 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78a9c6e5b837e33dc30c257ce6b868340cd343bd12a7527c17fbf126d027ff4d\": container with ID starting with 78a9c6e5b837e33dc30c257ce6b868340cd343bd12a7527c17fbf126d027ff4d not found: ID does not exist" containerID="78a9c6e5b837e33dc30c257ce6b868340cd343bd12a7527c17fbf126d027ff4d" Feb 17 13:59:43 crc kubenswrapper[4955]: I0217 13:59:43.309518 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78a9c6e5b837e33dc30c257ce6b868340cd343bd12a7527c17fbf126d027ff4d"} err="failed to get container status \"78a9c6e5b837e33dc30c257ce6b868340cd343bd12a7527c17fbf126d027ff4d\": rpc error: code = NotFound desc = could not find container \"78a9c6e5b837e33dc30c257ce6b868340cd343bd12a7527c17fbf126d027ff4d\": container with ID starting with 78a9c6e5b837e33dc30c257ce6b868340cd343bd12a7527c17fbf126d027ff4d not found: ID does not exist" Feb 17 13:59:43 crc kubenswrapper[4955]: I0217 13:59:43.309540 4955 scope.go:117] "RemoveContainer" containerID="be39ed15f2cbdfe55b9b8cdd9b0b22452fc73c5c45f05c45a02d454baf9844f6" Feb 17 13:59:43 crc kubenswrapper[4955]: E0217 13:59:43.309947 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be39ed15f2cbdfe55b9b8cdd9b0b22452fc73c5c45f05c45a02d454baf9844f6\": container with ID starting with be39ed15f2cbdfe55b9b8cdd9b0b22452fc73c5c45f05c45a02d454baf9844f6 not found: ID does not exist" containerID="be39ed15f2cbdfe55b9b8cdd9b0b22452fc73c5c45f05c45a02d454baf9844f6" Feb 17 13:59:43 crc kubenswrapper[4955]: I0217 13:59:43.309971 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be39ed15f2cbdfe55b9b8cdd9b0b22452fc73c5c45f05c45a02d454baf9844f6"} err="failed to get container status \"be39ed15f2cbdfe55b9b8cdd9b0b22452fc73c5c45f05c45a02d454baf9844f6\": rpc error: code = NotFound desc = could not find container \"be39ed15f2cbdfe55b9b8cdd9b0b22452fc73c5c45f05c45a02d454baf9844f6\": container with ID starting with be39ed15f2cbdfe55b9b8cdd9b0b22452fc73c5c45f05c45a02d454baf9844f6 not found: ID does not exist" Feb 17 13:59:44 crc kubenswrapper[4955]: I0217 13:59:44.209603 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-42t5x" event={"ID":"48598b8d-cbb4-4999-859d-543aa8237c4b","Type":"ContainerStarted","Data":"14818f90de212fb61fa82ea34b3ec88244a7df9d80590a305b3750b8ee1419d8"} Feb 17 13:59:44 crc kubenswrapper[4955]: I0217 13:59:44.234275 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6e6847c-4d47-4ea9-9d4e-40b2815616d5" path="/var/lib/kubelet/pods/a6e6847c-4d47-4ea9-9d4e-40b2815616d5/volumes" Feb 17 13:59:45 crc kubenswrapper[4955]: I0217 13:59:45.222354 4955 generic.go:334] "Generic (PLEG): container finished" podID="48598b8d-cbb4-4999-859d-543aa8237c4b" containerID="14818f90de212fb61fa82ea34b3ec88244a7df9d80590a305b3750b8ee1419d8" exitCode=0 Feb 17 13:59:45 crc kubenswrapper[4955]: I0217 13:59:45.222423 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-42t5x" event={"ID":"48598b8d-cbb4-4999-859d-543aa8237c4b","Type":"ContainerDied","Data":"14818f90de212fb61fa82ea34b3ec88244a7df9d80590a305b3750b8ee1419d8"} Feb 17 13:59:46 crc kubenswrapper[4955]: I0217 13:59:46.246686 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-42t5x" event={"ID":"48598b8d-cbb4-4999-859d-543aa8237c4b","Type":"ContainerStarted","Data":"acfe7dfaeb8b96d62ba293f72bf04ae8df32a6bd18af0f114e64746c2ee0c897"} Feb 17 13:59:46 crc kubenswrapper[4955]: I0217 13:59:46.277019 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-42t5x" podStartSLOduration=2.8666857329999997 podStartE2EDuration="5.276994141s" podCreationTimestamp="2026-02-17 13:59:41 +0000 UTC" firstStartedPulling="2026-02-17 13:59:43.199956907 +0000 UTC m=+3321.722686450" lastFinishedPulling="2026-02-17 13:59:45.610265275 +0000 UTC m=+3324.132994858" observedRunningTime="2026-02-17 13:59:46.268057155 +0000 UTC m=+3324.790786698" watchObservedRunningTime="2026-02-17 13:59:46.276994141 +0000 UTC m=+3324.799723684" Feb 17 13:59:51 crc kubenswrapper[4955]: I0217 13:59:51.487259 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-57f5fb6b66-7vp2j_87c5d914-d03a-41e9-9b1c-8c34a5e0758b/barbican-api/0.log" Feb 17 13:59:51 crc kubenswrapper[4955]: I0217 13:59:51.690570 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-57f5fb6b66-7vp2j_87c5d914-d03a-41e9-9b1c-8c34a5e0758b/barbican-api-log/0.log" Feb 17 13:59:51 crc kubenswrapper[4955]: I0217 13:59:51.770144 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7789497c74-pgff5_8468e621-9a50-49f7-ba03-996bc8564f43/barbican-keystone-listener/0.log" Feb 17 13:59:51 crc kubenswrapper[4955]: I0217 13:59:51.771472 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7789497c74-pgff5_8468e621-9a50-49f7-ba03-996bc8564f43/barbican-keystone-listener-log/0.log" Feb 17 13:59:51 crc kubenswrapper[4955]: I0217 13:59:51.778481 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-42t5x" Feb 17 13:59:51 crc kubenswrapper[4955]: I0217 13:59:51.778535 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-42t5x" Feb 17 13:59:51 crc kubenswrapper[4955]: I0217 13:59:51.833081 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-42t5x" Feb 17 13:59:51 crc kubenswrapper[4955]: I0217 13:59:51.937587 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6ff67b7c6c-9dcnd_20f0b6cf-c788-45bd-bdb2-cb2e01608a08/barbican-worker/0.log" Feb 17 13:59:51 crc kubenswrapper[4955]: I0217 13:59:51.974676 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6ff67b7c6c-9dcnd_20f0b6cf-c788-45bd-bdb2-cb2e01608a08/barbican-worker-log/0.log" Feb 17 13:59:52 crc kubenswrapper[4955]: I0217 13:59:52.172993 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c46f44da-c85d-4a6c-a5db-28e900c054da/ceilometer-central-agent/0.log" Feb 17 13:59:52 crc kubenswrapper[4955]: I0217 13:59:52.185641 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-zn5l6_ea41b3c5-2c55-4294-b2dc-a25ab083c780/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Feb 17 13:59:52 crc kubenswrapper[4955]: I0217 13:59:52.302130 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c46f44da-c85d-4a6c-a5db-28e900c054da/ceilometer-notification-agent/0.log" Feb 17 13:59:52 crc kubenswrapper[4955]: I0217 13:59:52.337307 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-42t5x" Feb 17 13:59:52 crc kubenswrapper[4955]: I0217 13:59:52.405648 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c46f44da-c85d-4a6c-a5db-28e900c054da/proxy-httpd/0.log" Feb 17 13:59:52 crc kubenswrapper[4955]: I0217 13:59:52.450345 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c46f44da-c85d-4a6c-a5db-28e900c054da/sg-core/0.log" Feb 17 13:59:52 crc kubenswrapper[4955]: I0217 13:59:52.543171 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_b1148d9f-bf24-449c-83b4-9a64d0572df0/cinder-api/0.log" Feb 17 13:59:52 crc kubenswrapper[4955]: I0217 13:59:52.582045 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_b1148d9f-bf24-449c-83b4-9a64d0572df0/cinder-api-log/0.log" Feb 17 13:59:52 crc kubenswrapper[4955]: I0217 13:59:52.707444 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7/cinder-scheduler/0.log" Feb 17 13:59:52 crc kubenswrapper[4955]: I0217 13:59:52.792274 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7/probe/0.log" Feb 17 13:59:52 crc kubenswrapper[4955]: I0217 13:59:52.925545 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-ncz59_c74ecb0c-c07e-4196-a0a8-7c107404b60f/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 17 13:59:52 crc kubenswrapper[4955]: I0217 13:59:52.987239 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-sj68p_5e268a8e-d845-4915-b6ff-f10bc386cfc5/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 17 13:59:53 crc kubenswrapper[4955]: I0217 13:59:53.112803 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-hktkl_d64b1ace-43dd-409e-a0ca-f7573676d9b3/init/0.log" Feb 17 13:59:53 crc kubenswrapper[4955]: I0217 13:59:53.266448 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-hktkl_d64b1ace-43dd-409e-a0ca-f7573676d9b3/init/0.log" Feb 17 13:59:53 crc kubenswrapper[4955]: I0217 13:59:53.325923 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-hktkl_d64b1ace-43dd-409e-a0ca-f7573676d9b3/dnsmasq-dns/0.log" Feb 17 13:59:53 crc kubenswrapper[4955]: I0217 13:59:53.374607 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-bpbzk_682aa1ae-e0dc-476d-862d-45071ee0609a/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Feb 17 13:59:53 crc kubenswrapper[4955]: I0217 13:59:53.507461 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_72354e86-fe70-4224-b6ea-c7fed8a90f5d/glance-httpd/0.log" Feb 17 13:59:53 crc kubenswrapper[4955]: I0217 13:59:53.587820 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_72354e86-fe70-4224-b6ea-c7fed8a90f5d/glance-log/0.log" Feb 17 13:59:53 crc kubenswrapper[4955]: I0217 13:59:53.709977 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_539c0e46-aba7-4b5b-a176-3e9ac9089560/glance-log/0.log" Feb 17 13:59:53 crc kubenswrapper[4955]: I0217 13:59:53.722973 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_539c0e46-aba7-4b5b-a176-3e9ac9089560/glance-httpd/0.log" Feb 17 13:59:53 crc kubenswrapper[4955]: I0217 13:59:53.862541 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7fffbcf7cb-wr74b_d6cb09d2-ad7d-4486-a41d-89023789fd87/horizon/0.log" Feb 17 13:59:54 crc kubenswrapper[4955]: I0217 13:59:54.181375 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx_f3e71e93-00f0-46f0-b87c-df0ddde2af3d/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Feb 17 13:59:54 crc kubenswrapper[4955]: I0217 13:59:54.272384 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-5h2ng_86978b4a-2cb7-4f99-bec3-0afeacdeb7b6/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 17 13:59:54 crc kubenswrapper[4955]: I0217 13:59:54.290807 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7fffbcf7cb-wr74b_d6cb09d2-ad7d-4486-a41d-89023789fd87/horizon-log/0.log" Feb 17 13:59:54 crc kubenswrapper[4955]: I0217 13:59:54.510988 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_3e1af961-6b57-4a77-ae1c-dcfacea5c8c6/kube-state-metrics/0.log" Feb 17 13:59:54 crc kubenswrapper[4955]: I0217 13:59:54.568431 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-c5b8c48f6-zr5fn_fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75/keystone-api/0.log" Feb 17 13:59:54 crc kubenswrapper[4955]: I0217 13:59:54.726890 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr_e3944196-3712-4506-bde5-758adf20aa63/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Feb 17 13:59:54 crc kubenswrapper[4955]: I0217 13:59:54.832585 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-42t5x"] Feb 17 13:59:55 crc kubenswrapper[4955]: I0217 13:59:55.233029 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7c5ddc6895-552ts_25c07d2a-4f61-48fc-88e3-0ed671ad9cff/neutron-api/0.log" Feb 17 13:59:55 crc kubenswrapper[4955]: I0217 13:59:55.253207 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7c5ddc6895-552ts_25c07d2a-4f61-48fc-88e3-0ed671ad9cff/neutron-httpd/0.log" Feb 17 13:59:55 crc kubenswrapper[4955]: I0217 13:59:55.344856 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-42t5x" podUID="48598b8d-cbb4-4999-859d-543aa8237c4b" containerName="registry-server" containerID="cri-o://acfe7dfaeb8b96d62ba293f72bf04ae8df32a6bd18af0f114e64746c2ee0c897" gracePeriod=2 Feb 17 13:59:55 crc kubenswrapper[4955]: I0217 13:59:55.498347 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf_5ff9b191-a542-41b9-a5c4-479d29e04fc3/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Feb 17 13:59:55 crc kubenswrapper[4955]: I0217 13:59:55.914305 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-42t5x" Feb 17 13:59:55 crc kubenswrapper[4955]: I0217 13:59:55.987920 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48598b8d-cbb4-4999-859d-543aa8237c4b-catalog-content\") pod \"48598b8d-cbb4-4999-859d-543aa8237c4b\" (UID: \"48598b8d-cbb4-4999-859d-543aa8237c4b\") " Feb 17 13:59:55 crc kubenswrapper[4955]: I0217 13:59:55.988084 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6p7s\" (UniqueName: \"kubernetes.io/projected/48598b8d-cbb4-4999-859d-543aa8237c4b-kube-api-access-v6p7s\") pod \"48598b8d-cbb4-4999-859d-543aa8237c4b\" (UID: \"48598b8d-cbb4-4999-859d-543aa8237c4b\") " Feb 17 13:59:55 crc kubenswrapper[4955]: I0217 13:59:55.988153 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48598b8d-cbb4-4999-859d-543aa8237c4b-utilities\") pod \"48598b8d-cbb4-4999-859d-543aa8237c4b\" (UID: \"48598b8d-cbb4-4999-859d-543aa8237c4b\") " Feb 17 13:59:55 crc kubenswrapper[4955]: I0217 13:59:55.989039 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48598b8d-cbb4-4999-859d-543aa8237c4b-utilities" (OuterVolumeSpecName: "utilities") pod "48598b8d-cbb4-4999-859d-543aa8237c4b" (UID: "48598b8d-cbb4-4999-859d-543aa8237c4b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:59:55 crc kubenswrapper[4955]: I0217 13:59:55.995272 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48598b8d-cbb4-4999-859d-543aa8237c4b-kube-api-access-v6p7s" (OuterVolumeSpecName: "kube-api-access-v6p7s") pod "48598b8d-cbb4-4999-859d-543aa8237c4b" (UID: "48598b8d-cbb4-4999-859d-543aa8237c4b"). InnerVolumeSpecName "kube-api-access-v6p7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 13:59:56 crc kubenswrapper[4955]: I0217 13:59:56.051382 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48598b8d-cbb4-4999-859d-543aa8237c4b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "48598b8d-cbb4-4999-859d-543aa8237c4b" (UID: "48598b8d-cbb4-4999-859d-543aa8237c4b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 13:59:56 crc kubenswrapper[4955]: I0217 13:59:56.089093 4955 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48598b8d-cbb4-4999-859d-543aa8237c4b-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 13:59:56 crc kubenswrapper[4955]: I0217 13:59:56.089127 4955 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48598b8d-cbb4-4999-859d-543aa8237c4b-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 13:59:56 crc kubenswrapper[4955]: I0217 13:59:56.089140 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6p7s\" (UniqueName: \"kubernetes.io/projected/48598b8d-cbb4-4999-859d-543aa8237c4b-kube-api-access-v6p7s\") on node \"crc\" DevicePath \"\"" Feb 17 13:59:56 crc kubenswrapper[4955]: I0217 13:59:56.273412 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_a1a37e38-9f12-49ad-8464-f744f7f5fc2f/nova-api-log/0.log" Feb 17 13:59:56 crc kubenswrapper[4955]: I0217 13:59:56.275741 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_115dedcd-9ffe-408e-9b6b-878f53c6c252/nova-cell0-conductor-conductor/0.log" Feb 17 13:59:56 crc kubenswrapper[4955]: I0217 13:59:56.324574 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_a1a37e38-9f12-49ad-8464-f744f7f5fc2f/nova-api-api/0.log" Feb 17 13:59:56 crc kubenswrapper[4955]: I0217 13:59:56.367417 4955 generic.go:334] "Generic (PLEG): container finished" podID="48598b8d-cbb4-4999-859d-543aa8237c4b" containerID="acfe7dfaeb8b96d62ba293f72bf04ae8df32a6bd18af0f114e64746c2ee0c897" exitCode=0 Feb 17 13:59:56 crc kubenswrapper[4955]: I0217 13:59:56.367478 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-42t5x" event={"ID":"48598b8d-cbb4-4999-859d-543aa8237c4b","Type":"ContainerDied","Data":"acfe7dfaeb8b96d62ba293f72bf04ae8df32a6bd18af0f114e64746c2ee0c897"} Feb 17 13:59:56 crc kubenswrapper[4955]: I0217 13:59:56.367505 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-42t5x" Feb 17 13:59:56 crc kubenswrapper[4955]: I0217 13:59:56.367512 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-42t5x" event={"ID":"48598b8d-cbb4-4999-859d-543aa8237c4b","Type":"ContainerDied","Data":"e57b70cdc3188683a1eb41df9acb2abd351c5d454b7fdab04d8de191a2c9c60d"} Feb 17 13:59:56 crc kubenswrapper[4955]: I0217 13:59:56.367545 4955 scope.go:117] "RemoveContainer" containerID="acfe7dfaeb8b96d62ba293f72bf04ae8df32a6bd18af0f114e64746c2ee0c897" Feb 17 13:59:56 crc kubenswrapper[4955]: I0217 13:59:56.391222 4955 scope.go:117] "RemoveContainer" containerID="14818f90de212fb61fa82ea34b3ec88244a7df9d80590a305b3750b8ee1419d8" Feb 17 13:59:56 crc kubenswrapper[4955]: I0217 13:59:56.395485 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-42t5x"] Feb 17 13:59:56 crc kubenswrapper[4955]: I0217 13:59:56.410069 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-42t5x"] Feb 17 13:59:56 crc kubenswrapper[4955]: I0217 13:59:56.423994 4955 scope.go:117] "RemoveContainer" containerID="fb1ed7335ec31c3003123879dc062770ef1f1094eed4a5596b29841ad49580cf" Feb 17 13:59:56 crc kubenswrapper[4955]: I0217 13:59:56.476141 4955 scope.go:117] "RemoveContainer" containerID="acfe7dfaeb8b96d62ba293f72bf04ae8df32a6bd18af0f114e64746c2ee0c897" Feb 17 13:59:56 crc kubenswrapper[4955]: E0217 13:59:56.476971 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"acfe7dfaeb8b96d62ba293f72bf04ae8df32a6bd18af0f114e64746c2ee0c897\": container with ID starting with acfe7dfaeb8b96d62ba293f72bf04ae8df32a6bd18af0f114e64746c2ee0c897 not found: ID does not exist" containerID="acfe7dfaeb8b96d62ba293f72bf04ae8df32a6bd18af0f114e64746c2ee0c897" Feb 17 13:59:56 crc kubenswrapper[4955]: I0217 13:59:56.477058 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"acfe7dfaeb8b96d62ba293f72bf04ae8df32a6bd18af0f114e64746c2ee0c897"} err="failed to get container status \"acfe7dfaeb8b96d62ba293f72bf04ae8df32a6bd18af0f114e64746c2ee0c897\": rpc error: code = NotFound desc = could not find container \"acfe7dfaeb8b96d62ba293f72bf04ae8df32a6bd18af0f114e64746c2ee0c897\": container with ID starting with acfe7dfaeb8b96d62ba293f72bf04ae8df32a6bd18af0f114e64746c2ee0c897 not found: ID does not exist" Feb 17 13:59:56 crc kubenswrapper[4955]: I0217 13:59:56.477109 4955 scope.go:117] "RemoveContainer" containerID="14818f90de212fb61fa82ea34b3ec88244a7df9d80590a305b3750b8ee1419d8" Feb 17 13:59:56 crc kubenswrapper[4955]: E0217 13:59:56.478195 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14818f90de212fb61fa82ea34b3ec88244a7df9d80590a305b3750b8ee1419d8\": container with ID starting with 14818f90de212fb61fa82ea34b3ec88244a7df9d80590a305b3750b8ee1419d8 not found: ID does not exist" containerID="14818f90de212fb61fa82ea34b3ec88244a7df9d80590a305b3750b8ee1419d8" Feb 17 13:59:56 crc kubenswrapper[4955]: I0217 13:59:56.478255 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14818f90de212fb61fa82ea34b3ec88244a7df9d80590a305b3750b8ee1419d8"} err="failed to get container status \"14818f90de212fb61fa82ea34b3ec88244a7df9d80590a305b3750b8ee1419d8\": rpc error: code = NotFound desc = could not find container \"14818f90de212fb61fa82ea34b3ec88244a7df9d80590a305b3750b8ee1419d8\": container with ID starting with 14818f90de212fb61fa82ea34b3ec88244a7df9d80590a305b3750b8ee1419d8 not found: ID does not exist" Feb 17 13:59:56 crc kubenswrapper[4955]: I0217 13:59:56.478291 4955 scope.go:117] "RemoveContainer" containerID="fb1ed7335ec31c3003123879dc062770ef1f1094eed4a5596b29841ad49580cf" Feb 17 13:59:56 crc kubenswrapper[4955]: E0217 13:59:56.478750 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb1ed7335ec31c3003123879dc062770ef1f1094eed4a5596b29841ad49580cf\": container with ID starting with fb1ed7335ec31c3003123879dc062770ef1f1094eed4a5596b29841ad49580cf not found: ID does not exist" containerID="fb1ed7335ec31c3003123879dc062770ef1f1094eed4a5596b29841ad49580cf" Feb 17 13:59:56 crc kubenswrapper[4955]: I0217 13:59:56.478847 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb1ed7335ec31c3003123879dc062770ef1f1094eed4a5596b29841ad49580cf"} err="failed to get container status \"fb1ed7335ec31c3003123879dc062770ef1f1094eed4a5596b29841ad49580cf\": rpc error: code = NotFound desc = could not find container \"fb1ed7335ec31c3003123879dc062770ef1f1094eed4a5596b29841ad49580cf\": container with ID starting with fb1ed7335ec31c3003123879dc062770ef1f1094eed4a5596b29841ad49580cf not found: ID does not exist" Feb 17 13:59:56 crc kubenswrapper[4955]: I0217 13:59:56.528109 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_2cc025ff-ae43-4877-940e-dab395505735/nova-cell1-conductor-conductor/0.log" Feb 17 13:59:56 crc kubenswrapper[4955]: I0217 13:59:56.595739 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_3614da91-13c7-4d09-b508-933322e35f08/nova-cell1-novncproxy-novncproxy/0.log" Feb 17 13:59:56 crc kubenswrapper[4955]: I0217 13:59:56.762424 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-v478l_1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f/nova-edpm-deployment-openstack-edpm-ipam/0.log" Feb 17 13:59:56 crc kubenswrapper[4955]: I0217 13:59:56.949815 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_ff800864-d970-46b0-bb2e-eb213d423da4/nova-metadata-log/0.log" Feb 17 13:59:57 crc kubenswrapper[4955]: I0217 13:59:57.207999 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1c93012c-0538-49a7-b414-6b01bb6c6dce/mysql-bootstrap/0.log" Feb 17 13:59:57 crc kubenswrapper[4955]: I0217 13:59:57.211292 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_ca665950-2b74-4ac5-8d48-eacdac1f65ae/nova-scheduler-scheduler/0.log" Feb 17 13:59:57 crc kubenswrapper[4955]: I0217 13:59:57.452102 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1c93012c-0538-49a7-b414-6b01bb6c6dce/mysql-bootstrap/0.log" Feb 17 13:59:57 crc kubenswrapper[4955]: I0217 13:59:57.567867 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1c93012c-0538-49a7-b414-6b01bb6c6dce/galera/0.log" Feb 17 13:59:57 crc kubenswrapper[4955]: I0217 13:59:57.652356 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_def1308a-11ee-4465-89f0-e78ffc4b8b51/mysql-bootstrap/0.log" Feb 17 13:59:57 crc kubenswrapper[4955]: I0217 13:59:57.819886 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_def1308a-11ee-4465-89f0-e78ffc4b8b51/mysql-bootstrap/0.log" Feb 17 13:59:57 crc kubenswrapper[4955]: I0217 13:59:57.871203 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_def1308a-11ee-4465-89f0-e78ffc4b8b51/galera/0.log" Feb 17 13:59:58 crc kubenswrapper[4955]: I0217 13:59:58.014850 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_ff800864-d970-46b0-bb2e-eb213d423da4/nova-metadata-metadata/0.log" Feb 17 13:59:58 crc kubenswrapper[4955]: I0217 13:59:58.074044 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-f8cjz_452bb98a-815b-4154-84c5-6b31338fca29/ovn-controller/0.log" Feb 17 13:59:58 crc kubenswrapper[4955]: I0217 13:59:58.123140 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_c2c452a7-9ff5-4f92-a5a9-87925a22302f/openstackclient/0.log" Feb 17 13:59:58 crc kubenswrapper[4955]: I0217 13:59:58.234744 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48598b8d-cbb4-4999-859d-543aa8237c4b" path="/var/lib/kubelet/pods/48598b8d-cbb4-4999-859d-543aa8237c4b/volumes" Feb 17 13:59:58 crc kubenswrapper[4955]: I0217 13:59:58.298379 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-2s2tl_c2a9d7a6-de50-4a56-b18a-ed69f82ea656/openstack-network-exporter/0.log" Feb 17 13:59:58 crc kubenswrapper[4955]: I0217 13:59:58.395351 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-m27b4_9534cd01-b88e-48a4-a2ca-dd968e045dbc/ovsdb-server-init/0.log" Feb 17 13:59:58 crc kubenswrapper[4955]: I0217 13:59:58.589757 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-m27b4_9534cd01-b88e-48a4-a2ca-dd968e045dbc/ovsdb-server-init/0.log" Feb 17 13:59:58 crc kubenswrapper[4955]: I0217 13:59:58.590529 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-m27b4_9534cd01-b88e-48a4-a2ca-dd968e045dbc/ovsdb-server/0.log" Feb 17 13:59:58 crc kubenswrapper[4955]: I0217 13:59:58.602089 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-m27b4_9534cd01-b88e-48a4-a2ca-dd968e045dbc/ovs-vswitchd/0.log" Feb 17 13:59:58 crc kubenswrapper[4955]: I0217 13:59:58.813374 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_94e88c0f-4478-4de1-9340-99904aac77e9/openstack-network-exporter/0.log" Feb 17 13:59:58 crc kubenswrapper[4955]: I0217 13:59:58.845651 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-5zsq2_27e2834f-9629-4902-b1b9-e613e29164c0/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Feb 17 13:59:58 crc kubenswrapper[4955]: I0217 13:59:58.992302 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_b3384848-891d-46ff-a869-61d657304b74/openstack-network-exporter/0.log" Feb 17 13:59:59 crc kubenswrapper[4955]: I0217 13:59:59.015093 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_94e88c0f-4478-4de1-9340-99904aac77e9/ovn-northd/0.log" Feb 17 13:59:59 crc kubenswrapper[4955]: I0217 13:59:59.149433 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_b3384848-891d-46ff-a869-61d657304b74/ovsdbserver-nb/0.log" Feb 17 13:59:59 crc kubenswrapper[4955]: I0217 13:59:59.219302 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_aa892f09-0663-4fb6-b3ad-a15c300b5b65/openstack-network-exporter/0.log" Feb 17 13:59:59 crc kubenswrapper[4955]: I0217 13:59:59.310301 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_aa892f09-0663-4fb6-b3ad-a15c300b5b65/ovsdbserver-sb/0.log" Feb 17 13:59:59 crc kubenswrapper[4955]: I0217 13:59:59.470440 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5d56c994-tx7w7_54f84d50-4d2b-4d71-91e0-e6a7a00f568e/placement-api/0.log" Feb 17 13:59:59 crc kubenswrapper[4955]: I0217 13:59:59.617243 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5d56c994-tx7w7_54f84d50-4d2b-4d71-91e0-e6a7a00f568e/placement-log/0.log" Feb 17 13:59:59 crc kubenswrapper[4955]: I0217 13:59:59.692241 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_cd70e7eb-fb12-4b99-ab83-67d4abb9392b/setup-container/0.log" Feb 17 13:59:59 crc kubenswrapper[4955]: I0217 13:59:59.920154 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_cd70e7eb-fb12-4b99-ab83-67d4abb9392b/rabbitmq/0.log" Feb 17 13:59:59 crc kubenswrapper[4955]: I0217 13:59:59.948819 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_c8c91914-fefa-479e-b497-de0f1f2978a2/setup-container/0.log" Feb 17 13:59:59 crc kubenswrapper[4955]: I0217 13:59:59.994385 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_cd70e7eb-fb12-4b99-ab83-67d4abb9392b/setup-container/0.log" Feb 17 14:00:00 crc kubenswrapper[4955]: I0217 14:00:00.153772 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29522280-74ml7"] Feb 17 14:00:00 crc kubenswrapper[4955]: E0217 14:00:00.154283 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6e6847c-4d47-4ea9-9d4e-40b2815616d5" containerName="extract-content" Feb 17 14:00:00 crc kubenswrapper[4955]: I0217 14:00:00.154298 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6e6847c-4d47-4ea9-9d4e-40b2815616d5" containerName="extract-content" Feb 17 14:00:00 crc kubenswrapper[4955]: E0217 14:00:00.154320 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48598b8d-cbb4-4999-859d-543aa8237c4b" containerName="extract-content" Feb 17 14:00:00 crc kubenswrapper[4955]: I0217 14:00:00.154328 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="48598b8d-cbb4-4999-859d-543aa8237c4b" containerName="extract-content" Feb 17 14:00:00 crc kubenswrapper[4955]: E0217 14:00:00.154349 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6e6847c-4d47-4ea9-9d4e-40b2815616d5" containerName="extract-utilities" Feb 17 14:00:00 crc kubenswrapper[4955]: I0217 14:00:00.154357 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6e6847c-4d47-4ea9-9d4e-40b2815616d5" containerName="extract-utilities" Feb 17 14:00:00 crc kubenswrapper[4955]: E0217 14:00:00.154378 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48598b8d-cbb4-4999-859d-543aa8237c4b" containerName="extract-utilities" Feb 17 14:00:00 crc kubenswrapper[4955]: I0217 14:00:00.154386 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="48598b8d-cbb4-4999-859d-543aa8237c4b" containerName="extract-utilities" Feb 17 14:00:00 crc kubenswrapper[4955]: E0217 14:00:00.154416 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6e6847c-4d47-4ea9-9d4e-40b2815616d5" containerName="registry-server" Feb 17 14:00:00 crc kubenswrapper[4955]: I0217 14:00:00.154423 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6e6847c-4d47-4ea9-9d4e-40b2815616d5" containerName="registry-server" Feb 17 14:00:00 crc kubenswrapper[4955]: E0217 14:00:00.154440 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48598b8d-cbb4-4999-859d-543aa8237c4b" containerName="registry-server" Feb 17 14:00:00 crc kubenswrapper[4955]: I0217 14:00:00.154447 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="48598b8d-cbb4-4999-859d-543aa8237c4b" containerName="registry-server" Feb 17 14:00:00 crc kubenswrapper[4955]: I0217 14:00:00.154666 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6e6847c-4d47-4ea9-9d4e-40b2815616d5" containerName="registry-server" Feb 17 14:00:00 crc kubenswrapper[4955]: I0217 14:00:00.154686 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="48598b8d-cbb4-4999-859d-543aa8237c4b" containerName="registry-server" Feb 17 14:00:00 crc kubenswrapper[4955]: I0217 14:00:00.155465 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29522280-74ml7" Feb 17 14:00:00 crc kubenswrapper[4955]: I0217 14:00:00.161079 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 17 14:00:00 crc kubenswrapper[4955]: I0217 14:00:00.161608 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 17 14:00:00 crc kubenswrapper[4955]: I0217 14:00:00.166517 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29522280-74ml7"] Feb 17 14:00:00 crc kubenswrapper[4955]: I0217 14:00:00.251980 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_c8c91914-fefa-479e-b497-de0f1f2978a2/setup-container/0.log" Feb 17 14:00:00 crc kubenswrapper[4955]: I0217 14:00:00.255126 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a42cf226-741b-4e00-ad49-33f676e8d9be-config-volume\") pod \"collect-profiles-29522280-74ml7\" (UID: \"a42cf226-741b-4e00-ad49-33f676e8d9be\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522280-74ml7" Feb 17 14:00:00 crc kubenswrapper[4955]: I0217 14:00:00.255228 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xftn4\" (UniqueName: \"kubernetes.io/projected/a42cf226-741b-4e00-ad49-33f676e8d9be-kube-api-access-xftn4\") pod \"collect-profiles-29522280-74ml7\" (UID: \"a42cf226-741b-4e00-ad49-33f676e8d9be\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522280-74ml7" Feb 17 14:00:00 crc kubenswrapper[4955]: I0217 14:00:00.255299 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a42cf226-741b-4e00-ad49-33f676e8d9be-secret-volume\") pod \"collect-profiles-29522280-74ml7\" (UID: \"a42cf226-741b-4e00-ad49-33f676e8d9be\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522280-74ml7" Feb 17 14:00:00 crc kubenswrapper[4955]: I0217 14:00:00.274321 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_c8c91914-fefa-479e-b497-de0f1f2978a2/rabbitmq/0.log" Feb 17 14:00:00 crc kubenswrapper[4955]: I0217 14:00:00.278994 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-mqnxx_5fe6a78c-e6a0-441f-b374-ca6603778eab/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 17 14:00:00 crc kubenswrapper[4955]: I0217 14:00:00.356693 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a42cf226-741b-4e00-ad49-33f676e8d9be-config-volume\") pod \"collect-profiles-29522280-74ml7\" (UID: \"a42cf226-741b-4e00-ad49-33f676e8d9be\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522280-74ml7" Feb 17 14:00:00 crc kubenswrapper[4955]: I0217 14:00:00.356770 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xftn4\" (UniqueName: \"kubernetes.io/projected/a42cf226-741b-4e00-ad49-33f676e8d9be-kube-api-access-xftn4\") pod \"collect-profiles-29522280-74ml7\" (UID: \"a42cf226-741b-4e00-ad49-33f676e8d9be\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522280-74ml7" Feb 17 14:00:00 crc kubenswrapper[4955]: I0217 14:00:00.356850 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a42cf226-741b-4e00-ad49-33f676e8d9be-secret-volume\") pod \"collect-profiles-29522280-74ml7\" (UID: \"a42cf226-741b-4e00-ad49-33f676e8d9be\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522280-74ml7" Feb 17 14:00:00 crc kubenswrapper[4955]: I0217 14:00:00.360124 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a42cf226-741b-4e00-ad49-33f676e8d9be-config-volume\") pod \"collect-profiles-29522280-74ml7\" (UID: \"a42cf226-741b-4e00-ad49-33f676e8d9be\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522280-74ml7" Feb 17 14:00:00 crc kubenswrapper[4955]: I0217 14:00:00.378697 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xftn4\" (UniqueName: \"kubernetes.io/projected/a42cf226-741b-4e00-ad49-33f676e8d9be-kube-api-access-xftn4\") pod \"collect-profiles-29522280-74ml7\" (UID: \"a42cf226-741b-4e00-ad49-33f676e8d9be\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522280-74ml7" Feb 17 14:00:00 crc kubenswrapper[4955]: I0217 14:00:00.391830 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a42cf226-741b-4e00-ad49-33f676e8d9be-secret-volume\") pod \"collect-profiles-29522280-74ml7\" (UID: \"a42cf226-741b-4e00-ad49-33f676e8d9be\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522280-74ml7" Feb 17 14:00:00 crc kubenswrapper[4955]: I0217 14:00:00.465755 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-cv8cz_8e826189-c723-4c9e-8ab8-3e60c06b7e42/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Feb 17 14:00:00 crc kubenswrapper[4955]: I0217 14:00:00.489654 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29522280-74ml7" Feb 17 14:00:00 crc kubenswrapper[4955]: I0217 14:00:00.533421 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-g2nhl_e1bffb7a-a33a-466d-b382-a6896e83d8d9/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Feb 17 14:00:00 crc kubenswrapper[4955]: I0217 14:00:00.770935 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-qplrv_15be064a-4dca-49b2-aac9-6891c1afb237/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 17 14:00:00 crc kubenswrapper[4955]: I0217 14:00:00.868584 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-t5v8s_e6728d3d-3842-41be-a973-efb3d3fe67b9/ssh-known-hosts-edpm-deployment/0.log" Feb 17 14:00:01 crc kubenswrapper[4955]: I0217 14:00:01.015498 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29522280-74ml7"] Feb 17 14:00:01 crc kubenswrapper[4955]: I0217 14:00:01.114964 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5bcc76b6cf-4r2g9_0fa0efd9-940a-4580-ab68-c5954c0e732e/proxy-server/0.log" Feb 17 14:00:01 crc kubenswrapper[4955]: I0217 14:00:01.197182 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5bcc76b6cf-4r2g9_0fa0efd9-940a-4580-ab68-c5954c0e732e/proxy-httpd/0.log" Feb 17 14:00:01 crc kubenswrapper[4955]: I0217 14:00:01.289567 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-56p7b_46beefa1-21de-4818-a70f-e9f840a920ec/swift-ring-rebalance/0.log" Feb 17 14:00:01 crc kubenswrapper[4955]: I0217 14:00:01.401192 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fb6d9434-4476-438b-90e4-4781bd85b0bb/account-auditor/0.log" Feb 17 14:00:01 crc kubenswrapper[4955]: I0217 14:00:01.419409 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29522280-74ml7" event={"ID":"a42cf226-741b-4e00-ad49-33f676e8d9be","Type":"ContainerStarted","Data":"fbe170de473593bd1dfe8f501c256d9d1dc5e16db920ad94c66697480f6d0f18"} Feb 17 14:00:01 crc kubenswrapper[4955]: I0217 14:00:01.419466 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29522280-74ml7" event={"ID":"a42cf226-741b-4e00-ad49-33f676e8d9be","Type":"ContainerStarted","Data":"e4d8f3d589873e2d18218440ee0e34d6ea686f79e73c959f4b0aba5709fbe1dd"} Feb 17 14:00:01 crc kubenswrapper[4955]: I0217 14:00:01.441669 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29522280-74ml7" podStartSLOduration=1.441647271 podStartE2EDuration="1.441647271s" podCreationTimestamp="2026-02-17 14:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 14:00:01.438227683 +0000 UTC m=+3339.960957246" watchObservedRunningTime="2026-02-17 14:00:01.441647271 +0000 UTC m=+3339.964376814" Feb 17 14:00:01 crc kubenswrapper[4955]: I0217 14:00:01.522098 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fb6d9434-4476-438b-90e4-4781bd85b0bb/account-reaper/0.log" Feb 17 14:00:01 crc kubenswrapper[4955]: I0217 14:00:01.550219 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fb6d9434-4476-438b-90e4-4781bd85b0bb/account-replicator/0.log" Feb 17 14:00:01 crc kubenswrapper[4955]: I0217 14:00:01.610454 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fb6d9434-4476-438b-90e4-4781bd85b0bb/account-server/0.log" Feb 17 14:00:01 crc kubenswrapper[4955]: I0217 14:00:01.712574 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fb6d9434-4476-438b-90e4-4781bd85b0bb/container-auditor/0.log" Feb 17 14:00:01 crc kubenswrapper[4955]: I0217 14:00:01.775871 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fb6d9434-4476-438b-90e4-4781bd85b0bb/container-server/0.log" Feb 17 14:00:01 crc kubenswrapper[4955]: I0217 14:00:01.804651 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fb6d9434-4476-438b-90e4-4781bd85b0bb/container-updater/0.log" Feb 17 14:00:01 crc kubenswrapper[4955]: I0217 14:00:01.829524 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fb6d9434-4476-438b-90e4-4781bd85b0bb/container-replicator/0.log" Feb 17 14:00:01 crc kubenswrapper[4955]: I0217 14:00:01.944042 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fb6d9434-4476-438b-90e4-4781bd85b0bb/object-auditor/0.log" Feb 17 14:00:01 crc kubenswrapper[4955]: I0217 14:00:01.958866 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fb6d9434-4476-438b-90e4-4781bd85b0bb/object-expirer/0.log" Feb 17 14:00:02 crc kubenswrapper[4955]: I0217 14:00:02.010007 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fb6d9434-4476-438b-90e4-4781bd85b0bb/object-replicator/0.log" Feb 17 14:00:02 crc kubenswrapper[4955]: I0217 14:00:02.064905 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fb6d9434-4476-438b-90e4-4781bd85b0bb/object-server/0.log" Feb 17 14:00:02 crc kubenswrapper[4955]: I0217 14:00:02.193721 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fb6d9434-4476-438b-90e4-4781bd85b0bb/object-updater/0.log" Feb 17 14:00:02 crc kubenswrapper[4955]: I0217 14:00:02.257351 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fb6d9434-4476-438b-90e4-4781bd85b0bb/rsync/0.log" Feb 17 14:00:02 crc kubenswrapper[4955]: I0217 14:00:02.264119 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fb6d9434-4476-438b-90e4-4781bd85b0bb/swift-recon-cron/0.log" Feb 17 14:00:02 crc kubenswrapper[4955]: I0217 14:00:02.428811 4955 generic.go:334] "Generic (PLEG): container finished" podID="a42cf226-741b-4e00-ad49-33f676e8d9be" containerID="fbe170de473593bd1dfe8f501c256d9d1dc5e16db920ad94c66697480f6d0f18" exitCode=0 Feb 17 14:00:02 crc kubenswrapper[4955]: I0217 14:00:02.428879 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29522280-74ml7" event={"ID":"a42cf226-741b-4e00-ad49-33f676e8d9be","Type":"ContainerDied","Data":"fbe170de473593bd1dfe8f501c256d9d1dc5e16db920ad94c66697480f6d0f18"} Feb 17 14:00:02 crc kubenswrapper[4955]: I0217 14:00:02.540153 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-npz67_0acb7fb8-671d-4e90-8be4-f5f507308f11/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Feb 17 14:00:02 crc kubenswrapper[4955]: I0217 14:00:02.579267 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_de1d0057-d623-4185-b31a-572dfaabcce2/tempest-tests-tempest-tests-runner/0.log" Feb 17 14:00:02 crc kubenswrapper[4955]: I0217 14:00:02.756371 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-jr9rx_b743858f-111b-4d79-aeaf-6b9608fecbac/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 17 14:00:02 crc kubenswrapper[4955]: I0217 14:00:02.818239 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_35a9a907-8d3b-49dd-9dd3-0f2e1a5aa7bc/test-operator-logs-container/0.log" Feb 17 14:00:03 crc kubenswrapper[4955]: I0217 14:00:03.821267 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29522280-74ml7" Feb 17 14:00:03 crc kubenswrapper[4955]: I0217 14:00:03.958831 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xftn4\" (UniqueName: \"kubernetes.io/projected/a42cf226-741b-4e00-ad49-33f676e8d9be-kube-api-access-xftn4\") pod \"a42cf226-741b-4e00-ad49-33f676e8d9be\" (UID: \"a42cf226-741b-4e00-ad49-33f676e8d9be\") " Feb 17 14:00:03 crc kubenswrapper[4955]: I0217 14:00:03.958967 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a42cf226-741b-4e00-ad49-33f676e8d9be-config-volume\") pod \"a42cf226-741b-4e00-ad49-33f676e8d9be\" (UID: \"a42cf226-741b-4e00-ad49-33f676e8d9be\") " Feb 17 14:00:03 crc kubenswrapper[4955]: I0217 14:00:03.959144 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a42cf226-741b-4e00-ad49-33f676e8d9be-secret-volume\") pod \"a42cf226-741b-4e00-ad49-33f676e8d9be\" (UID: \"a42cf226-741b-4e00-ad49-33f676e8d9be\") " Feb 17 14:00:03 crc kubenswrapper[4955]: I0217 14:00:03.959746 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a42cf226-741b-4e00-ad49-33f676e8d9be-config-volume" (OuterVolumeSpecName: "config-volume") pod "a42cf226-741b-4e00-ad49-33f676e8d9be" (UID: "a42cf226-741b-4e00-ad49-33f676e8d9be"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 14:00:03 crc kubenswrapper[4955]: I0217 14:00:03.967069 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a42cf226-741b-4e00-ad49-33f676e8d9be-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a42cf226-741b-4e00-ad49-33f676e8d9be" (UID: "a42cf226-741b-4e00-ad49-33f676e8d9be"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 14:00:03 crc kubenswrapper[4955]: I0217 14:00:03.971061 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a42cf226-741b-4e00-ad49-33f676e8d9be-kube-api-access-xftn4" (OuterVolumeSpecName: "kube-api-access-xftn4") pod "a42cf226-741b-4e00-ad49-33f676e8d9be" (UID: "a42cf226-741b-4e00-ad49-33f676e8d9be"). InnerVolumeSpecName "kube-api-access-xftn4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 14:00:04 crc kubenswrapper[4955]: I0217 14:00:04.061247 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xftn4\" (UniqueName: \"kubernetes.io/projected/a42cf226-741b-4e00-ad49-33f676e8d9be-kube-api-access-xftn4\") on node \"crc\" DevicePath \"\"" Feb 17 14:00:04 crc kubenswrapper[4955]: I0217 14:00:04.061618 4955 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a42cf226-741b-4e00-ad49-33f676e8d9be-config-volume\") on node \"crc\" DevicePath \"\"" Feb 17 14:00:04 crc kubenswrapper[4955]: I0217 14:00:04.061632 4955 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a42cf226-741b-4e00-ad49-33f676e8d9be-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 17 14:00:04 crc kubenswrapper[4955]: I0217 14:00:04.446557 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29522280-74ml7" event={"ID":"a42cf226-741b-4e00-ad49-33f676e8d9be","Type":"ContainerDied","Data":"e4d8f3d589873e2d18218440ee0e34d6ea686f79e73c959f4b0aba5709fbe1dd"} Feb 17 14:00:04 crc kubenswrapper[4955]: I0217 14:00:04.446634 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29522280-74ml7" Feb 17 14:00:04 crc kubenswrapper[4955]: I0217 14:00:04.446634 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e4d8f3d589873e2d18218440ee0e34d6ea686f79e73c959f4b0aba5709fbe1dd" Feb 17 14:00:04 crc kubenswrapper[4955]: I0217 14:00:04.522703 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29522235-qssjp"] Feb 17 14:00:04 crc kubenswrapper[4955]: I0217 14:00:04.533693 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29522235-qssjp"] Feb 17 14:00:06 crc kubenswrapper[4955]: I0217 14:00:06.239435 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45631608-0d2f-4144-a96b-1a921c8f014b" path="/var/lib/kubelet/pods/45631608-0d2f-4144-a96b-1a921c8f014b/volumes" Feb 17 14:00:10 crc kubenswrapper[4955]: I0217 14:00:10.799901 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_7dab7b98-7d1a-4940-a4c3-8bbbd57fe121/memcached/0.log" Feb 17 14:00:27 crc kubenswrapper[4955]: I0217 14:00:27.342982 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p_f328d9c1-907b-4705-8580-0b5da54053c6/util/0.log" Feb 17 14:00:27 crc kubenswrapper[4955]: I0217 14:00:27.533915 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p_f328d9c1-907b-4705-8580-0b5da54053c6/util/0.log" Feb 17 14:00:27 crc kubenswrapper[4955]: I0217 14:00:27.595030 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p_f328d9c1-907b-4705-8580-0b5da54053c6/pull/0.log" Feb 17 14:00:27 crc kubenswrapper[4955]: I0217 14:00:27.773510 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p_f328d9c1-907b-4705-8580-0b5da54053c6/pull/0.log" Feb 17 14:00:27 crc kubenswrapper[4955]: I0217 14:00:27.963849 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p_f328d9c1-907b-4705-8580-0b5da54053c6/util/0.log" Feb 17 14:00:28 crc kubenswrapper[4955]: I0217 14:00:28.009439 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p_f328d9c1-907b-4705-8580-0b5da54053c6/pull/0.log" Feb 17 14:00:28 crc kubenswrapper[4955]: I0217 14:00:28.178356 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p_f328d9c1-907b-4705-8580-0b5da54053c6/extract/0.log" Feb 17 14:00:28 crc kubenswrapper[4955]: I0217 14:00:28.428194 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-6d8bf5c495-2c8ww_1e4fd408-9f09-4289-92ce-7c5ae18f235a/manager/0.log" Feb 17 14:00:28 crc kubenswrapper[4955]: I0217 14:00:28.785122 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987464f4-g967h_5c339089-ba43-481b-aaa8-f8bf09fd106d/manager/0.log" Feb 17 14:00:28 crc kubenswrapper[4955]: I0217 14:00:28.785361 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-5d946d989d-lpxnk_6d1e254f-7d68-45c5-bda0-e3b33155ae56/manager/0.log" Feb 17 14:00:28 crc kubenswrapper[4955]: I0217 14:00:28.873558 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-69f49c598c-75hpg_071c3b13-9304-48b6-bf22-d89dd62825a0/manager/0.log" Feb 17 14:00:29 crc kubenswrapper[4955]: I0217 14:00:29.042207 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5b9b8895d5-k2x44_6ade752d-a3d0-4449-8028-67405763ccc8/manager/0.log" Feb 17 14:00:29 crc kubenswrapper[4955]: I0217 14:00:29.388659 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-554564d7fc-h2rm6_9019acce-c356-421a-bd47-a163a8857f30/manager/0.log" Feb 17 14:00:29 crc kubenswrapper[4955]: I0217 14:00:29.602367 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-79d975b745-gpgxj_cc84cdf7-92aa-4134-a784-d821c076cde3/manager/0.log" Feb 17 14:00:29 crc kubenswrapper[4955]: I0217 14:00:29.704638 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-b4d948c87-mkqrg_d6decb51-f3f9-4183-a20a-6aa4491dba61/manager/0.log" Feb 17 14:00:29 crc kubenswrapper[4955]: I0217 14:00:29.814769 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-54f6768c69-4f8kb_01eb7df7-9e0a-4e78-85a3-0d456dec139f/manager/0.log" Feb 17 14:00:30 crc kubenswrapper[4955]: I0217 14:00:30.118333 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6994f66f48-nzn4k_308b7713-e37c-4890-8b56-11379bad3f8d/manager/0.log" Feb 17 14:00:30 crc kubenswrapper[4955]: I0217 14:00:30.346024 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-64ddbf8bb-xkvlp_037562f9-c04f-4e9d-8fa5-3e98ac7c8fc5/manager/0.log" Feb 17 14:00:30 crc kubenswrapper[4955]: I0217 14:00:30.418873 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-567668f5cf-5fxtz_744f25f9-1641-4864-8a55-ea14ead19cae/manager/0.log" Feb 17 14:00:30 crc kubenswrapper[4955]: I0217 14:00:30.792710 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-7c6767dc9cxkmq8_cd2ef1c9-23d1-4671-bc71-2c077afdfe08/manager/0.log" Feb 17 14:00:31 crc kubenswrapper[4955]: I0217 14:00:31.308322 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-57f9c85f89-8zxwf_5b01c583-8db4-4939-ac7b-e0d16b9f72bf/operator/0.log" Feb 17 14:00:31 crc kubenswrapper[4955]: I0217 14:00:31.505732 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-5qzcl_38d53049-7d41-4938-b2d6-c7c7568ba144/registry-server/0.log" Feb 17 14:00:31 crc kubenswrapper[4955]: I0217 14:00:31.810474 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-d44cf6b75-vq5wq_17cecde4-2eb0-445e-bcf4-f73f719627df/manager/0.log" Feb 17 14:00:32 crc kubenswrapper[4955]: I0217 14:00:32.055854 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-8497b45c89-fqrz9_788f9fac-dd10-49c4-8a25-4578d69aeb2b/manager/0.log" Feb 17 14:00:32 crc kubenswrapper[4955]: I0217 14:00:32.079866 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69f8888797-wkdhh_3f752291-5ba5-4074-a126-e99f6f0242aa/manager/0.log" Feb 17 14:00:32 crc kubenswrapper[4955]: I0217 14:00:32.331020 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-nmxwn_afaac24a-0894-4001-9b50-d3d7c2de90bc/operator/0.log" Feb 17 14:00:32 crc kubenswrapper[4955]: I0217 14:00:32.364627 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-68f46476f-5qz9n_60b19c88-d91f-422e-9012-ad97daf661bb/manager/0.log" Feb 17 14:00:32 crc kubenswrapper[4955]: I0217 14:00:32.666106 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7f45b4ff68-q68lr_72fe0652-eee2-4e49-83b9-3c1790e8ca9e/manager/0.log" Feb 17 14:00:32 crc kubenswrapper[4955]: I0217 14:00:32.688082 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-7866795846-xqxjt_90e44f80-9802-4134-8aa3-519bc93f4bd3/manager/0.log" Feb 17 14:00:32 crc kubenswrapper[4955]: I0217 14:00:32.913106 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-5db88f68c-v6jkn_a1f4d718-9535-4290-94eb-93d6288d6047/manager/0.log" Feb 17 14:00:33 crc kubenswrapper[4955]: I0217 14:00:33.125234 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5699594bc4-nrc2h_c93c2a50-6396-4a32-9b88-dcdc341ec6c5/manager/0.log" Feb 17 14:00:34 crc kubenswrapper[4955]: I0217 14:00:34.464383 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-868647ff47-ffkxx_602419fb-5033-4889-b587-e1eb97001910/manager/0.log" Feb 17 14:00:34 crc kubenswrapper[4955]: I0217 14:00:34.574695 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 14:00:34 crc kubenswrapper[4955]: I0217 14:00:34.574761 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 14:00:51 crc kubenswrapper[4955]: I0217 14:00:51.692893 4955 scope.go:117] "RemoveContainer" containerID="ea3c883435d1bb5d78114d7bd4d40ad96e51f3a077b2012a807279fed60386c0" Feb 17 14:00:52 crc kubenswrapper[4955]: I0217 14:00:52.200601 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-vtd5z_8f5feb72-96ca-4e4d-9566-0757c7dd93f9/control-plane-machine-set-operator/0.log" Feb 17 14:00:52 crc kubenswrapper[4955]: I0217 14:00:52.391901 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-d5sv6_847ef69d-8cd1-4831-8250-05f2cf789d88/kube-rbac-proxy/0.log" Feb 17 14:00:52 crc kubenswrapper[4955]: I0217 14:00:52.413988 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-d5sv6_847ef69d-8cd1-4831-8250-05f2cf789d88/machine-api-operator/0.log" Feb 17 14:01:00 crc kubenswrapper[4955]: I0217 14:01:00.145166 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29522281-wgswt"] Feb 17 14:01:00 crc kubenswrapper[4955]: E0217 14:01:00.146129 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a42cf226-741b-4e00-ad49-33f676e8d9be" containerName="collect-profiles" Feb 17 14:01:00 crc kubenswrapper[4955]: I0217 14:01:00.146148 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="a42cf226-741b-4e00-ad49-33f676e8d9be" containerName="collect-profiles" Feb 17 14:01:00 crc kubenswrapper[4955]: I0217 14:01:00.146342 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="a42cf226-741b-4e00-ad49-33f676e8d9be" containerName="collect-profiles" Feb 17 14:01:00 crc kubenswrapper[4955]: I0217 14:01:00.146975 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29522281-wgswt" Feb 17 14:01:00 crc kubenswrapper[4955]: I0217 14:01:00.170195 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29522281-wgswt"] Feb 17 14:01:00 crc kubenswrapper[4955]: I0217 14:01:00.307860 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0ab4d69-eeed-461b-bc7d-1005eacc732f-combined-ca-bundle\") pod \"keystone-cron-29522281-wgswt\" (UID: \"f0ab4d69-eeed-461b-bc7d-1005eacc732f\") " pod="openstack/keystone-cron-29522281-wgswt" Feb 17 14:01:00 crc kubenswrapper[4955]: I0217 14:01:00.307948 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67zbq\" (UniqueName: \"kubernetes.io/projected/f0ab4d69-eeed-461b-bc7d-1005eacc732f-kube-api-access-67zbq\") pod \"keystone-cron-29522281-wgswt\" (UID: \"f0ab4d69-eeed-461b-bc7d-1005eacc732f\") " pod="openstack/keystone-cron-29522281-wgswt" Feb 17 14:01:00 crc kubenswrapper[4955]: I0217 14:01:00.307984 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0ab4d69-eeed-461b-bc7d-1005eacc732f-config-data\") pod \"keystone-cron-29522281-wgswt\" (UID: \"f0ab4d69-eeed-461b-bc7d-1005eacc732f\") " pod="openstack/keystone-cron-29522281-wgswt" Feb 17 14:01:00 crc kubenswrapper[4955]: I0217 14:01:00.308011 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f0ab4d69-eeed-461b-bc7d-1005eacc732f-fernet-keys\") pod \"keystone-cron-29522281-wgswt\" (UID: \"f0ab4d69-eeed-461b-bc7d-1005eacc732f\") " pod="openstack/keystone-cron-29522281-wgswt" Feb 17 14:01:00 crc kubenswrapper[4955]: I0217 14:01:00.409166 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67zbq\" (UniqueName: \"kubernetes.io/projected/f0ab4d69-eeed-461b-bc7d-1005eacc732f-kube-api-access-67zbq\") pod \"keystone-cron-29522281-wgswt\" (UID: \"f0ab4d69-eeed-461b-bc7d-1005eacc732f\") " pod="openstack/keystone-cron-29522281-wgswt" Feb 17 14:01:00 crc kubenswrapper[4955]: I0217 14:01:00.409238 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0ab4d69-eeed-461b-bc7d-1005eacc732f-config-data\") pod \"keystone-cron-29522281-wgswt\" (UID: \"f0ab4d69-eeed-461b-bc7d-1005eacc732f\") " pod="openstack/keystone-cron-29522281-wgswt" Feb 17 14:01:00 crc kubenswrapper[4955]: I0217 14:01:00.409319 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f0ab4d69-eeed-461b-bc7d-1005eacc732f-fernet-keys\") pod \"keystone-cron-29522281-wgswt\" (UID: \"f0ab4d69-eeed-461b-bc7d-1005eacc732f\") " pod="openstack/keystone-cron-29522281-wgswt" Feb 17 14:01:00 crc kubenswrapper[4955]: I0217 14:01:00.409436 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0ab4d69-eeed-461b-bc7d-1005eacc732f-combined-ca-bundle\") pod \"keystone-cron-29522281-wgswt\" (UID: \"f0ab4d69-eeed-461b-bc7d-1005eacc732f\") " pod="openstack/keystone-cron-29522281-wgswt" Feb 17 14:01:00 crc kubenswrapper[4955]: I0217 14:01:00.415552 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0ab4d69-eeed-461b-bc7d-1005eacc732f-combined-ca-bundle\") pod \"keystone-cron-29522281-wgswt\" (UID: \"f0ab4d69-eeed-461b-bc7d-1005eacc732f\") " pod="openstack/keystone-cron-29522281-wgswt" Feb 17 14:01:00 crc kubenswrapper[4955]: I0217 14:01:00.417090 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f0ab4d69-eeed-461b-bc7d-1005eacc732f-fernet-keys\") pod \"keystone-cron-29522281-wgswt\" (UID: \"f0ab4d69-eeed-461b-bc7d-1005eacc732f\") " pod="openstack/keystone-cron-29522281-wgswt" Feb 17 14:01:00 crc kubenswrapper[4955]: I0217 14:01:00.417575 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0ab4d69-eeed-461b-bc7d-1005eacc732f-config-data\") pod \"keystone-cron-29522281-wgswt\" (UID: \"f0ab4d69-eeed-461b-bc7d-1005eacc732f\") " pod="openstack/keystone-cron-29522281-wgswt" Feb 17 14:01:00 crc kubenswrapper[4955]: I0217 14:01:00.430482 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67zbq\" (UniqueName: \"kubernetes.io/projected/f0ab4d69-eeed-461b-bc7d-1005eacc732f-kube-api-access-67zbq\") pod \"keystone-cron-29522281-wgswt\" (UID: \"f0ab4d69-eeed-461b-bc7d-1005eacc732f\") " pod="openstack/keystone-cron-29522281-wgswt" Feb 17 14:01:00 crc kubenswrapper[4955]: I0217 14:01:00.466539 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29522281-wgswt" Feb 17 14:01:00 crc kubenswrapper[4955]: I0217 14:01:00.907579 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29522281-wgswt"] Feb 17 14:01:01 crc kubenswrapper[4955]: I0217 14:01:01.471774 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29522281-wgswt" event={"ID":"f0ab4d69-eeed-461b-bc7d-1005eacc732f","Type":"ContainerStarted","Data":"913a04f96e49977c4ca1df5db96c0cd6b7e8242f0e7a1117e257eb870aab4f37"} Feb 17 14:01:01 crc kubenswrapper[4955]: I0217 14:01:01.472125 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29522281-wgswt" event={"ID":"f0ab4d69-eeed-461b-bc7d-1005eacc732f","Type":"ContainerStarted","Data":"ad016262af1b461e0971228accb35b3938529e9f418b8efb66f6881899d2b12e"} Feb 17 14:01:01 crc kubenswrapper[4955]: I0217 14:01:01.496834 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29522281-wgswt" podStartSLOduration=1.496781639 podStartE2EDuration="1.496781639s" podCreationTimestamp="2026-02-17 14:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 14:01:01.488190963 +0000 UTC m=+3400.010920506" watchObservedRunningTime="2026-02-17 14:01:01.496781639 +0000 UTC m=+3400.019511202" Feb 17 14:01:04 crc kubenswrapper[4955]: I0217 14:01:04.296218 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-lg6zf_54536ade-766d-4cf8-b1b5-057ccb961b13/cert-manager-controller/0.log" Feb 17 14:01:04 crc kubenswrapper[4955]: I0217 14:01:04.489874 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-lmhj4_de5cd487-58d9-4c86-a4e2-f6b000a351f5/cert-manager-cainjector/0.log" Feb 17 14:01:04 crc kubenswrapper[4955]: I0217 14:01:04.499946 4955 generic.go:334] "Generic (PLEG): container finished" podID="f0ab4d69-eeed-461b-bc7d-1005eacc732f" containerID="913a04f96e49977c4ca1df5db96c0cd6b7e8242f0e7a1117e257eb870aab4f37" exitCode=0 Feb 17 14:01:04 crc kubenswrapper[4955]: I0217 14:01:04.499992 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29522281-wgswt" event={"ID":"f0ab4d69-eeed-461b-bc7d-1005eacc732f","Type":"ContainerDied","Data":"913a04f96e49977c4ca1df5db96c0cd6b7e8242f0e7a1117e257eb870aab4f37"} Feb 17 14:01:04 crc kubenswrapper[4955]: I0217 14:01:04.561739 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-69dts_d4a28595-c677-4966-8d3a-1ad5d299fd51/cert-manager-webhook/0.log" Feb 17 14:01:04 crc kubenswrapper[4955]: I0217 14:01:04.574312 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 14:01:04 crc kubenswrapper[4955]: I0217 14:01:04.574373 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 14:01:05 crc kubenswrapper[4955]: I0217 14:01:05.868024 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29522281-wgswt" Feb 17 14:01:06 crc kubenswrapper[4955]: I0217 14:01:06.009671 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67zbq\" (UniqueName: \"kubernetes.io/projected/f0ab4d69-eeed-461b-bc7d-1005eacc732f-kube-api-access-67zbq\") pod \"f0ab4d69-eeed-461b-bc7d-1005eacc732f\" (UID: \"f0ab4d69-eeed-461b-bc7d-1005eacc732f\") " Feb 17 14:01:06 crc kubenswrapper[4955]: I0217 14:01:06.009940 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f0ab4d69-eeed-461b-bc7d-1005eacc732f-fernet-keys\") pod \"f0ab4d69-eeed-461b-bc7d-1005eacc732f\" (UID: \"f0ab4d69-eeed-461b-bc7d-1005eacc732f\") " Feb 17 14:01:06 crc kubenswrapper[4955]: I0217 14:01:06.010708 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0ab4d69-eeed-461b-bc7d-1005eacc732f-combined-ca-bundle\") pod \"f0ab4d69-eeed-461b-bc7d-1005eacc732f\" (UID: \"f0ab4d69-eeed-461b-bc7d-1005eacc732f\") " Feb 17 14:01:06 crc kubenswrapper[4955]: I0217 14:01:06.010833 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0ab4d69-eeed-461b-bc7d-1005eacc732f-config-data\") pod \"f0ab4d69-eeed-461b-bc7d-1005eacc732f\" (UID: \"f0ab4d69-eeed-461b-bc7d-1005eacc732f\") " Feb 17 14:01:06 crc kubenswrapper[4955]: I0217 14:01:06.016446 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0ab4d69-eeed-461b-bc7d-1005eacc732f-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "f0ab4d69-eeed-461b-bc7d-1005eacc732f" (UID: "f0ab4d69-eeed-461b-bc7d-1005eacc732f"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 14:01:06 crc kubenswrapper[4955]: I0217 14:01:06.029601 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0ab4d69-eeed-461b-bc7d-1005eacc732f-kube-api-access-67zbq" (OuterVolumeSpecName: "kube-api-access-67zbq") pod "f0ab4d69-eeed-461b-bc7d-1005eacc732f" (UID: "f0ab4d69-eeed-461b-bc7d-1005eacc732f"). InnerVolumeSpecName "kube-api-access-67zbq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 14:01:06 crc kubenswrapper[4955]: I0217 14:01:06.041763 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0ab4d69-eeed-461b-bc7d-1005eacc732f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f0ab4d69-eeed-461b-bc7d-1005eacc732f" (UID: "f0ab4d69-eeed-461b-bc7d-1005eacc732f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 14:01:06 crc kubenswrapper[4955]: I0217 14:01:06.073686 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0ab4d69-eeed-461b-bc7d-1005eacc732f-config-data" (OuterVolumeSpecName: "config-data") pod "f0ab4d69-eeed-461b-bc7d-1005eacc732f" (UID: "f0ab4d69-eeed-461b-bc7d-1005eacc732f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 14:01:06 crc kubenswrapper[4955]: I0217 14:01:06.113091 4955 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0ab4d69-eeed-461b-bc7d-1005eacc732f-config-data\") on node \"crc\" DevicePath \"\"" Feb 17 14:01:06 crc kubenswrapper[4955]: I0217 14:01:06.113124 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67zbq\" (UniqueName: \"kubernetes.io/projected/f0ab4d69-eeed-461b-bc7d-1005eacc732f-kube-api-access-67zbq\") on node \"crc\" DevicePath \"\"" Feb 17 14:01:06 crc kubenswrapper[4955]: I0217 14:01:06.113135 4955 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f0ab4d69-eeed-461b-bc7d-1005eacc732f-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 17 14:01:06 crc kubenswrapper[4955]: I0217 14:01:06.113143 4955 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0ab4d69-eeed-461b-bc7d-1005eacc732f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 17 14:01:06 crc kubenswrapper[4955]: I0217 14:01:06.518989 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29522281-wgswt" event={"ID":"f0ab4d69-eeed-461b-bc7d-1005eacc732f","Type":"ContainerDied","Data":"ad016262af1b461e0971228accb35b3938529e9f418b8efb66f6881899d2b12e"} Feb 17 14:01:06 crc kubenswrapper[4955]: I0217 14:01:06.519039 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad016262af1b461e0971228accb35b3938529e9f418b8efb66f6881899d2b12e" Feb 17 14:01:06 crc kubenswrapper[4955]: I0217 14:01:06.519319 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29522281-wgswt" Feb 17 14:01:16 crc kubenswrapper[4955]: I0217 14:01:16.575624 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5c78fc5d65-bdkvr_0afd3165-a15b-406f-88fb-4ce82ab17355/nmstate-console-plugin/0.log" Feb 17 14:01:16 crc kubenswrapper[4955]: I0217 14:01:16.782637 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-9vdw4_bb2572e6-ed65-4a6c-a3ca-f72c6cb4b5c0/nmstate-handler/0.log" Feb 17 14:01:16 crc kubenswrapper[4955]: I0217 14:01:16.802790 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58c85c668d-b4ngx_bcf404eb-4186-40c9-ba70-768b78c8a60d/kube-rbac-proxy/0.log" Feb 17 14:01:16 crc kubenswrapper[4955]: I0217 14:01:16.917441 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58c85c668d-b4ngx_bcf404eb-4186-40c9-ba70-768b78c8a60d/nmstate-metrics/0.log" Feb 17 14:01:17 crc kubenswrapper[4955]: I0217 14:01:17.015258 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-694c9596b7-j9h2x_4bd50965-e917-4e34-80b2-4e54e1d67feb/nmstate-operator/0.log" Feb 17 14:01:17 crc kubenswrapper[4955]: I0217 14:01:17.123459 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-866bcb46dc-d7hqh_3aca323f-af5c-452c-8a4e-8c2b55e39dd5/nmstate-webhook/0.log" Feb 17 14:01:34 crc kubenswrapper[4955]: I0217 14:01:34.574985 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 14:01:34 crc kubenswrapper[4955]: I0217 14:01:34.575610 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 14:01:34 crc kubenswrapper[4955]: I0217 14:01:34.575661 4955 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" Feb 17 14:01:34 crc kubenswrapper[4955]: I0217 14:01:34.576497 4955 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f365e77198a4218f176dac0c14964d043e7082a6840a3a671b2914acc5870d29"} pod="openshift-machine-config-operator/machine-config-daemon-29qxq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 17 14:01:34 crc kubenswrapper[4955]: I0217 14:01:34.576548 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" containerID="cri-o://f365e77198a4218f176dac0c14964d043e7082a6840a3a671b2914acc5870d29" gracePeriod=600 Feb 17 14:01:34 crc kubenswrapper[4955]: I0217 14:01:34.747408 4955 generic.go:334] "Generic (PLEG): container finished" podID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerID="f365e77198a4218f176dac0c14964d043e7082a6840a3a671b2914acc5870d29" exitCode=0 Feb 17 14:01:34 crc kubenswrapper[4955]: I0217 14:01:34.747451 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" event={"ID":"5bf36887-2a2b-4193-abdb-a15beaf46300","Type":"ContainerDied","Data":"f365e77198a4218f176dac0c14964d043e7082a6840a3a671b2914acc5870d29"} Feb 17 14:01:34 crc kubenswrapper[4955]: I0217 14:01:34.747487 4955 scope.go:117] "RemoveContainer" containerID="61c4421445b156f2b4a07625b9dae400b378009c3aef4994cad2a6ccfbc32b03" Feb 17 14:01:35 crc kubenswrapper[4955]: I0217 14:01:35.759648 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" event={"ID":"5bf36887-2a2b-4193-abdb-a15beaf46300","Type":"ContainerStarted","Data":"fa86f3109bf152769bd9369dabc17d54290d99858877992d6ce0702550e664a2"} Feb 17 14:01:42 crc kubenswrapper[4955]: I0217 14:01:42.481253 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-69bbfbf88f-pvzn9_8c11830d-6699-432a-a2d3-14b135bc1665/kube-rbac-proxy/0.log" Feb 17 14:01:42 crc kubenswrapper[4955]: I0217 14:01:42.703352 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-69bbfbf88f-pvzn9_8c11830d-6699-432a-a2d3-14b135bc1665/controller/0.log" Feb 17 14:01:42 crc kubenswrapper[4955]: I0217 14:01:42.852815 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vzhr7_720ac59e-55ea-44be-b7af-9b57fec4901b/cp-frr-files/0.log" Feb 17 14:01:43 crc kubenswrapper[4955]: I0217 14:01:43.003025 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vzhr7_720ac59e-55ea-44be-b7af-9b57fec4901b/cp-frr-files/0.log" Feb 17 14:01:43 crc kubenswrapper[4955]: I0217 14:01:43.050555 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vzhr7_720ac59e-55ea-44be-b7af-9b57fec4901b/cp-metrics/0.log" Feb 17 14:01:43 crc kubenswrapper[4955]: I0217 14:01:43.075246 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vzhr7_720ac59e-55ea-44be-b7af-9b57fec4901b/cp-reloader/0.log" Feb 17 14:01:43 crc kubenswrapper[4955]: I0217 14:01:43.111623 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vzhr7_720ac59e-55ea-44be-b7af-9b57fec4901b/cp-reloader/0.log" Feb 17 14:01:43 crc kubenswrapper[4955]: I0217 14:01:43.274951 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vzhr7_720ac59e-55ea-44be-b7af-9b57fec4901b/cp-metrics/0.log" Feb 17 14:01:43 crc kubenswrapper[4955]: I0217 14:01:43.300728 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vzhr7_720ac59e-55ea-44be-b7af-9b57fec4901b/cp-reloader/0.log" Feb 17 14:01:43 crc kubenswrapper[4955]: I0217 14:01:43.326544 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vzhr7_720ac59e-55ea-44be-b7af-9b57fec4901b/cp-frr-files/0.log" Feb 17 14:01:43 crc kubenswrapper[4955]: I0217 14:01:43.382053 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vzhr7_720ac59e-55ea-44be-b7af-9b57fec4901b/cp-metrics/0.log" Feb 17 14:01:43 crc kubenswrapper[4955]: I0217 14:01:43.474021 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vzhr7_720ac59e-55ea-44be-b7af-9b57fec4901b/cp-reloader/0.log" Feb 17 14:01:43 crc kubenswrapper[4955]: I0217 14:01:43.500937 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vzhr7_720ac59e-55ea-44be-b7af-9b57fec4901b/cp-frr-files/0.log" Feb 17 14:01:43 crc kubenswrapper[4955]: I0217 14:01:43.544349 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vzhr7_720ac59e-55ea-44be-b7af-9b57fec4901b/cp-metrics/0.log" Feb 17 14:01:43 crc kubenswrapper[4955]: I0217 14:01:43.579775 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vzhr7_720ac59e-55ea-44be-b7af-9b57fec4901b/controller/0.log" Feb 17 14:01:43 crc kubenswrapper[4955]: I0217 14:01:43.722067 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vzhr7_720ac59e-55ea-44be-b7af-9b57fec4901b/kube-rbac-proxy/0.log" Feb 17 14:01:43 crc kubenswrapper[4955]: I0217 14:01:43.779797 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vzhr7_720ac59e-55ea-44be-b7af-9b57fec4901b/frr-metrics/0.log" Feb 17 14:01:43 crc kubenswrapper[4955]: I0217 14:01:43.804752 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vzhr7_720ac59e-55ea-44be-b7af-9b57fec4901b/kube-rbac-proxy-frr/0.log" Feb 17 14:01:43 crc kubenswrapper[4955]: I0217 14:01:43.997148 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vzhr7_720ac59e-55ea-44be-b7af-9b57fec4901b/reloader/0.log" Feb 17 14:01:44 crc kubenswrapper[4955]: I0217 14:01:44.053193 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-78b44bf5bb-rcv4v_92393c39-fc3d-490c-93a3-c11ddef89c02/frr-k8s-webhook-server/0.log" Feb 17 14:01:44 crc kubenswrapper[4955]: I0217 14:01:44.301320 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-754bdc4867-j8r95_6f6cf1f8-d54b-4fad-8313-c5f235a9df88/manager/0.log" Feb 17 14:01:44 crc kubenswrapper[4955]: I0217 14:01:44.503609 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-dl5vp_618eb288-1aa4-43a9-b26a-9e2f5378411d/kube-rbac-proxy/0.log" Feb 17 14:01:44 crc kubenswrapper[4955]: I0217 14:01:44.540394 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-579d9646f7-zpz6n_4b384d59-9dc8-4d70-8b12-bd97252d1cfa/webhook-server/0.log" Feb 17 14:01:45 crc kubenswrapper[4955]: I0217 14:01:45.095939 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vzhr7_720ac59e-55ea-44be-b7af-9b57fec4901b/frr/0.log" Feb 17 14:01:45 crc kubenswrapper[4955]: I0217 14:01:45.232992 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-dl5vp_618eb288-1aa4-43a9-b26a-9e2f5378411d/speaker/0.log" Feb 17 14:01:57 crc kubenswrapper[4955]: I0217 14:01:57.732400 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l_102e3d48-f56c-4e7c-956e-23484bf45776/util/0.log" Feb 17 14:01:57 crc kubenswrapper[4955]: I0217 14:01:57.944224 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l_102e3d48-f56c-4e7c-956e-23484bf45776/util/0.log" Feb 17 14:01:57 crc kubenswrapper[4955]: I0217 14:01:57.955803 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l_102e3d48-f56c-4e7c-956e-23484bf45776/pull/0.log" Feb 17 14:01:57 crc kubenswrapper[4955]: I0217 14:01:57.977498 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l_102e3d48-f56c-4e7c-956e-23484bf45776/pull/0.log" Feb 17 14:01:58 crc kubenswrapper[4955]: I0217 14:01:58.148887 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l_102e3d48-f56c-4e7c-956e-23484bf45776/pull/0.log" Feb 17 14:01:58 crc kubenswrapper[4955]: I0217 14:01:58.178361 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l_102e3d48-f56c-4e7c-956e-23484bf45776/extract/0.log" Feb 17 14:01:58 crc kubenswrapper[4955]: I0217 14:01:58.203987 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l_102e3d48-f56c-4e7c-956e-23484bf45776/util/0.log" Feb 17 14:01:58 crc kubenswrapper[4955]: I0217 14:01:58.309401 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-krpv5_47a043d1-03b2-4a46-8c24-f1d1fef7efa6/extract-utilities/0.log" Feb 17 14:01:58 crc kubenswrapper[4955]: I0217 14:01:58.489332 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-krpv5_47a043d1-03b2-4a46-8c24-f1d1fef7efa6/extract-utilities/0.log" Feb 17 14:01:58 crc kubenswrapper[4955]: I0217 14:01:58.492317 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-krpv5_47a043d1-03b2-4a46-8c24-f1d1fef7efa6/extract-content/0.log" Feb 17 14:01:58 crc kubenswrapper[4955]: I0217 14:01:58.581584 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-krpv5_47a043d1-03b2-4a46-8c24-f1d1fef7efa6/extract-content/0.log" Feb 17 14:01:58 crc kubenswrapper[4955]: I0217 14:01:58.852671 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-krpv5_47a043d1-03b2-4a46-8c24-f1d1fef7efa6/extract-utilities/0.log" Feb 17 14:01:58 crc kubenswrapper[4955]: I0217 14:01:58.908960 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-krpv5_47a043d1-03b2-4a46-8c24-f1d1fef7efa6/extract-content/0.log" Feb 17 14:01:59 crc kubenswrapper[4955]: I0217 14:01:59.130175 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2jxwm_59516804-ad0e-428f-bf9a-1b5561f7ec1a/extract-utilities/0.log" Feb 17 14:01:59 crc kubenswrapper[4955]: I0217 14:01:59.352065 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-krpv5_47a043d1-03b2-4a46-8c24-f1d1fef7efa6/registry-server/0.log" Feb 17 14:01:59 crc kubenswrapper[4955]: I0217 14:01:59.357419 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2jxwm_59516804-ad0e-428f-bf9a-1b5561f7ec1a/extract-utilities/0.log" Feb 17 14:01:59 crc kubenswrapper[4955]: I0217 14:01:59.371560 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2jxwm_59516804-ad0e-428f-bf9a-1b5561f7ec1a/extract-content/0.log" Feb 17 14:01:59 crc kubenswrapper[4955]: I0217 14:01:59.430847 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2jxwm_59516804-ad0e-428f-bf9a-1b5561f7ec1a/extract-content/0.log" Feb 17 14:01:59 crc kubenswrapper[4955]: I0217 14:01:59.574350 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2jxwm_59516804-ad0e-428f-bf9a-1b5561f7ec1a/extract-utilities/0.log" Feb 17 14:01:59 crc kubenswrapper[4955]: I0217 14:01:59.604065 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2jxwm_59516804-ad0e-428f-bf9a-1b5561f7ec1a/extract-content/0.log" Feb 17 14:01:59 crc kubenswrapper[4955]: I0217 14:01:59.830150 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j_c5bff73f-3f9e-4313-93ac-6276ceaf16d7/util/0.log" Feb 17 14:02:00 crc kubenswrapper[4955]: I0217 14:02:00.078149 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j_c5bff73f-3f9e-4313-93ac-6276ceaf16d7/pull/0.log" Feb 17 14:02:00 crc kubenswrapper[4955]: I0217 14:02:00.122865 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j_c5bff73f-3f9e-4313-93ac-6276ceaf16d7/util/0.log" Feb 17 14:02:00 crc kubenswrapper[4955]: I0217 14:02:00.159529 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j_c5bff73f-3f9e-4313-93ac-6276ceaf16d7/pull/0.log" Feb 17 14:02:00 crc kubenswrapper[4955]: I0217 14:02:00.348382 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2jxwm_59516804-ad0e-428f-bf9a-1b5561f7ec1a/registry-server/0.log" Feb 17 14:02:00 crc kubenswrapper[4955]: I0217 14:02:00.352957 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j_c5bff73f-3f9e-4313-93ac-6276ceaf16d7/util/0.log" Feb 17 14:02:00 crc kubenswrapper[4955]: I0217 14:02:00.362691 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j_c5bff73f-3f9e-4313-93ac-6276ceaf16d7/pull/0.log" Feb 17 14:02:00 crc kubenswrapper[4955]: I0217 14:02:00.394769 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j_c5bff73f-3f9e-4313-93ac-6276ceaf16d7/extract/0.log" Feb 17 14:02:00 crc kubenswrapper[4955]: I0217 14:02:00.590087 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-spj8f_f7954d5c-c985-4421-96e9-24ae232c8ba1/marketplace-operator/0.log" Feb 17 14:02:00 crc kubenswrapper[4955]: I0217 14:02:00.691188 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jmrnj_1cf150d2-8598-45d3-b193-4d27e90ca648/extract-utilities/0.log" Feb 17 14:02:00 crc kubenswrapper[4955]: I0217 14:02:00.886086 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jmrnj_1cf150d2-8598-45d3-b193-4d27e90ca648/extract-utilities/0.log" Feb 17 14:02:00 crc kubenswrapper[4955]: I0217 14:02:00.896429 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jmrnj_1cf150d2-8598-45d3-b193-4d27e90ca648/extract-content/0.log" Feb 17 14:02:00 crc kubenswrapper[4955]: I0217 14:02:00.902251 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jmrnj_1cf150d2-8598-45d3-b193-4d27e90ca648/extract-content/0.log" Feb 17 14:02:01 crc kubenswrapper[4955]: I0217 14:02:01.078674 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jmrnj_1cf150d2-8598-45d3-b193-4d27e90ca648/extract-utilities/0.log" Feb 17 14:02:01 crc kubenswrapper[4955]: I0217 14:02:01.162949 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jmrnj_1cf150d2-8598-45d3-b193-4d27e90ca648/extract-content/0.log" Feb 17 14:02:01 crc kubenswrapper[4955]: I0217 14:02:01.233968 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jmrnj_1cf150d2-8598-45d3-b193-4d27e90ca648/registry-server/0.log" Feb 17 14:02:01 crc kubenswrapper[4955]: I0217 14:02:01.315446 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qz5rg_a5c8c29f-7770-4ba6-aad9-12604d10981f/extract-utilities/0.log" Feb 17 14:02:01 crc kubenswrapper[4955]: I0217 14:02:01.515237 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qz5rg_a5c8c29f-7770-4ba6-aad9-12604d10981f/extract-content/0.log" Feb 17 14:02:01 crc kubenswrapper[4955]: I0217 14:02:01.515810 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qz5rg_a5c8c29f-7770-4ba6-aad9-12604d10981f/extract-utilities/0.log" Feb 17 14:02:01 crc kubenswrapper[4955]: I0217 14:02:01.531123 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qz5rg_a5c8c29f-7770-4ba6-aad9-12604d10981f/extract-content/0.log" Feb 17 14:02:01 crc kubenswrapper[4955]: I0217 14:02:01.689763 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qz5rg_a5c8c29f-7770-4ba6-aad9-12604d10981f/extract-utilities/0.log" Feb 17 14:02:01 crc kubenswrapper[4955]: I0217 14:02:01.793142 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qz5rg_a5c8c29f-7770-4ba6-aad9-12604d10981f/extract-content/0.log" Feb 17 14:02:02 crc kubenswrapper[4955]: I0217 14:02:02.312875 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qz5rg_a5c8c29f-7770-4ba6-aad9-12604d10981f/registry-server/0.log" Feb 17 14:02:17 crc kubenswrapper[4955]: E0217 14:02:17.523091 4955 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.46:34172->38.102.83.46:34749: write tcp 38.102.83.46:34172->38.102.83.46:34749: write: broken pipe Feb 17 14:03:34 crc kubenswrapper[4955]: I0217 14:03:34.575859 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 14:03:34 crc kubenswrapper[4955]: I0217 14:03:34.576909 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 14:03:53 crc kubenswrapper[4955]: E0217 14:03:53.481798 4955 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod131a5d3f_e605_4310_a2ff_2044b7afcdbd.slice/crio-5b074e5605e0d42f70cda17d7dbba758f1c0a68e29c8ad26ad1f174dee3db699.scope\": RecentStats: unable to find data in memory cache]" Feb 17 14:03:54 crc kubenswrapper[4955]: I0217 14:03:54.041259 4955 generic.go:334] "Generic (PLEG): container finished" podID="131a5d3f-e605-4310-a2ff-2044b7afcdbd" containerID="5b074e5605e0d42f70cda17d7dbba758f1c0a68e29c8ad26ad1f174dee3db699" exitCode=0 Feb 17 14:03:54 crc kubenswrapper[4955]: I0217 14:03:54.041338 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b5tlj/must-gather-nmxsd" event={"ID":"131a5d3f-e605-4310-a2ff-2044b7afcdbd","Type":"ContainerDied","Data":"5b074e5605e0d42f70cda17d7dbba758f1c0a68e29c8ad26ad1f174dee3db699"} Feb 17 14:03:54 crc kubenswrapper[4955]: I0217 14:03:54.042377 4955 scope.go:117] "RemoveContainer" containerID="5b074e5605e0d42f70cda17d7dbba758f1c0a68e29c8ad26ad1f174dee3db699" Feb 17 14:03:54 crc kubenswrapper[4955]: I0217 14:03:54.520607 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-b5tlj_must-gather-nmxsd_131a5d3f-e605-4310-a2ff-2044b7afcdbd/gather/0.log" Feb 17 14:04:01 crc kubenswrapper[4955]: I0217 14:04:01.980755 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-b5tlj/must-gather-nmxsd"] Feb 17 14:04:01 crc kubenswrapper[4955]: I0217 14:04:01.981501 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-b5tlj/must-gather-nmxsd" podUID="131a5d3f-e605-4310-a2ff-2044b7afcdbd" containerName="copy" containerID="cri-o://251a0123e1e6c1cab048e9516aa909045ebc74811c69feaa5983297329d348f4" gracePeriod=2 Feb 17 14:04:01 crc kubenswrapper[4955]: I0217 14:04:01.989655 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-b5tlj/must-gather-nmxsd"] Feb 17 14:04:02 crc kubenswrapper[4955]: I0217 14:04:02.124587 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-b5tlj_must-gather-nmxsd_131a5d3f-e605-4310-a2ff-2044b7afcdbd/copy/0.log" Feb 17 14:04:02 crc kubenswrapper[4955]: I0217 14:04:02.125408 4955 generic.go:334] "Generic (PLEG): container finished" podID="131a5d3f-e605-4310-a2ff-2044b7afcdbd" containerID="251a0123e1e6c1cab048e9516aa909045ebc74811c69feaa5983297329d348f4" exitCode=143 Feb 17 14:04:02 crc kubenswrapper[4955]: I0217 14:04:02.440842 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-b5tlj_must-gather-nmxsd_131a5d3f-e605-4310-a2ff-2044b7afcdbd/copy/0.log" Feb 17 14:04:02 crc kubenswrapper[4955]: I0217 14:04:02.441546 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b5tlj/must-gather-nmxsd" Feb 17 14:04:02 crc kubenswrapper[4955]: I0217 14:04:02.542222 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7fhs\" (UniqueName: \"kubernetes.io/projected/131a5d3f-e605-4310-a2ff-2044b7afcdbd-kube-api-access-h7fhs\") pod \"131a5d3f-e605-4310-a2ff-2044b7afcdbd\" (UID: \"131a5d3f-e605-4310-a2ff-2044b7afcdbd\") " Feb 17 14:04:02 crc kubenswrapper[4955]: I0217 14:04:02.542418 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/131a5d3f-e605-4310-a2ff-2044b7afcdbd-must-gather-output\") pod \"131a5d3f-e605-4310-a2ff-2044b7afcdbd\" (UID: \"131a5d3f-e605-4310-a2ff-2044b7afcdbd\") " Feb 17 14:04:02 crc kubenswrapper[4955]: I0217 14:04:02.547764 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/131a5d3f-e605-4310-a2ff-2044b7afcdbd-kube-api-access-h7fhs" (OuterVolumeSpecName: "kube-api-access-h7fhs") pod "131a5d3f-e605-4310-a2ff-2044b7afcdbd" (UID: "131a5d3f-e605-4310-a2ff-2044b7afcdbd"). InnerVolumeSpecName "kube-api-access-h7fhs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 14:04:02 crc kubenswrapper[4955]: I0217 14:04:02.645240 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7fhs\" (UniqueName: \"kubernetes.io/projected/131a5d3f-e605-4310-a2ff-2044b7afcdbd-kube-api-access-h7fhs\") on node \"crc\" DevicePath \"\"" Feb 17 14:04:02 crc kubenswrapper[4955]: I0217 14:04:02.701985 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/131a5d3f-e605-4310-a2ff-2044b7afcdbd-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "131a5d3f-e605-4310-a2ff-2044b7afcdbd" (UID: "131a5d3f-e605-4310-a2ff-2044b7afcdbd"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 14:04:02 crc kubenswrapper[4955]: I0217 14:04:02.747077 4955 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/131a5d3f-e605-4310-a2ff-2044b7afcdbd-must-gather-output\") on node \"crc\" DevicePath \"\"" Feb 17 14:04:03 crc kubenswrapper[4955]: I0217 14:04:03.135540 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-b5tlj_must-gather-nmxsd_131a5d3f-e605-4310-a2ff-2044b7afcdbd/copy/0.log" Feb 17 14:04:03 crc kubenswrapper[4955]: I0217 14:04:03.136085 4955 scope.go:117] "RemoveContainer" containerID="251a0123e1e6c1cab048e9516aa909045ebc74811c69feaa5983297329d348f4" Feb 17 14:04:03 crc kubenswrapper[4955]: I0217 14:04:03.136154 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b5tlj/must-gather-nmxsd" Feb 17 14:04:03 crc kubenswrapper[4955]: I0217 14:04:03.158836 4955 scope.go:117] "RemoveContainer" containerID="5b074e5605e0d42f70cda17d7dbba758f1c0a68e29c8ad26ad1f174dee3db699" Feb 17 14:04:04 crc kubenswrapper[4955]: I0217 14:04:04.241259 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="131a5d3f-e605-4310-a2ff-2044b7afcdbd" path="/var/lib/kubelet/pods/131a5d3f-e605-4310-a2ff-2044b7afcdbd/volumes" Feb 17 14:04:04 crc kubenswrapper[4955]: I0217 14:04:04.575358 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 14:04:04 crc kubenswrapper[4955]: I0217 14:04:04.575693 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 14:04:34 crc kubenswrapper[4955]: I0217 14:04:34.575203 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 14:04:34 crc kubenswrapper[4955]: I0217 14:04:34.575895 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 14:04:34 crc kubenswrapper[4955]: I0217 14:04:34.575940 4955 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" Feb 17 14:04:34 crc kubenswrapper[4955]: I0217 14:04:34.576608 4955 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fa86f3109bf152769bd9369dabc17d54290d99858877992d6ce0702550e664a2"} pod="openshift-machine-config-operator/machine-config-daemon-29qxq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 17 14:04:34 crc kubenswrapper[4955]: I0217 14:04:34.576658 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" containerID="cri-o://fa86f3109bf152769bd9369dabc17d54290d99858877992d6ce0702550e664a2" gracePeriod=600 Feb 17 14:04:34 crc kubenswrapper[4955]: E0217 14:04:34.695736 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 14:04:35 crc kubenswrapper[4955]: I0217 14:04:35.451254 4955 generic.go:334] "Generic (PLEG): container finished" podID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerID="fa86f3109bf152769bd9369dabc17d54290d99858877992d6ce0702550e664a2" exitCode=0 Feb 17 14:04:35 crc kubenswrapper[4955]: I0217 14:04:35.451315 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" event={"ID":"5bf36887-2a2b-4193-abdb-a15beaf46300","Type":"ContainerDied","Data":"fa86f3109bf152769bd9369dabc17d54290d99858877992d6ce0702550e664a2"} Feb 17 14:04:35 crc kubenswrapper[4955]: I0217 14:04:35.451608 4955 scope.go:117] "RemoveContainer" containerID="f365e77198a4218f176dac0c14964d043e7082a6840a3a671b2914acc5870d29" Feb 17 14:04:35 crc kubenswrapper[4955]: I0217 14:04:35.452567 4955 scope.go:117] "RemoveContainer" containerID="fa86f3109bf152769bd9369dabc17d54290d99858877992d6ce0702550e664a2" Feb 17 14:04:35 crc kubenswrapper[4955]: E0217 14:04:35.453170 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 14:04:51 crc kubenswrapper[4955]: I0217 14:04:51.223527 4955 scope.go:117] "RemoveContainer" containerID="fa86f3109bf152769bd9369dabc17d54290d99858877992d6ce0702550e664a2" Feb 17 14:04:51 crc kubenswrapper[4955]: E0217 14:04:51.224252 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 14:04:51 crc kubenswrapper[4955]: I0217 14:04:51.832097 4955 scope.go:117] "RemoveContainer" containerID="240783d5b2b351a103df8937664ab4a362aa27e4e7f0deeae2dc9febc23ed024" Feb 17 14:05:03 crc kubenswrapper[4955]: I0217 14:05:03.223045 4955 scope.go:117] "RemoveContainer" containerID="fa86f3109bf152769bd9369dabc17d54290d99858877992d6ce0702550e664a2" Feb 17 14:05:03 crc kubenswrapper[4955]: E0217 14:05:03.223726 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 14:05:17 crc kubenswrapper[4955]: I0217 14:05:17.222520 4955 scope.go:117] "RemoveContainer" containerID="fa86f3109bf152769bd9369dabc17d54290d99858877992d6ce0702550e664a2" Feb 17 14:05:17 crc kubenswrapper[4955]: E0217 14:05:17.223375 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 14:05:28 crc kubenswrapper[4955]: I0217 14:05:28.223420 4955 scope.go:117] "RemoveContainer" containerID="fa86f3109bf152769bd9369dabc17d54290d99858877992d6ce0702550e664a2" Feb 17 14:05:28 crc kubenswrapper[4955]: E0217 14:05:28.224103 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 14:05:40 crc kubenswrapper[4955]: I0217 14:05:40.224139 4955 scope.go:117] "RemoveContainer" containerID="fa86f3109bf152769bd9369dabc17d54290d99858877992d6ce0702550e664a2" Feb 17 14:05:40 crc kubenswrapper[4955]: E0217 14:05:40.225121 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 14:05:51 crc kubenswrapper[4955]: I0217 14:05:51.223550 4955 scope.go:117] "RemoveContainer" containerID="fa86f3109bf152769bd9369dabc17d54290d99858877992d6ce0702550e664a2" Feb 17 14:05:51 crc kubenswrapper[4955]: E0217 14:05:51.224729 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 14:06:02 crc kubenswrapper[4955]: I0217 14:06:02.231625 4955 scope.go:117] "RemoveContainer" containerID="fa86f3109bf152769bd9369dabc17d54290d99858877992d6ce0702550e664a2" Feb 17 14:06:02 crc kubenswrapper[4955]: E0217 14:06:02.232546 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 14:06:13 crc kubenswrapper[4955]: I0217 14:06:13.223203 4955 scope.go:117] "RemoveContainer" containerID="fa86f3109bf152769bd9369dabc17d54290d99858877992d6ce0702550e664a2" Feb 17 14:06:13 crc kubenswrapper[4955]: E0217 14:06:13.223971 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 14:06:25 crc kubenswrapper[4955]: I0217 14:06:25.222826 4955 scope.go:117] "RemoveContainer" containerID="fa86f3109bf152769bd9369dabc17d54290d99858877992d6ce0702550e664a2" Feb 17 14:06:25 crc kubenswrapper[4955]: E0217 14:06:25.224753 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 14:06:36 crc kubenswrapper[4955]: I0217 14:06:36.915546 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hp9l7"] Feb 17 14:06:36 crc kubenswrapper[4955]: E0217 14:06:36.916522 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="131a5d3f-e605-4310-a2ff-2044b7afcdbd" containerName="copy" Feb 17 14:06:36 crc kubenswrapper[4955]: I0217 14:06:36.916541 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="131a5d3f-e605-4310-a2ff-2044b7afcdbd" containerName="copy" Feb 17 14:06:36 crc kubenswrapper[4955]: E0217 14:06:36.916577 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0ab4d69-eeed-461b-bc7d-1005eacc732f" containerName="keystone-cron" Feb 17 14:06:36 crc kubenswrapper[4955]: I0217 14:06:36.916585 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0ab4d69-eeed-461b-bc7d-1005eacc732f" containerName="keystone-cron" Feb 17 14:06:36 crc kubenswrapper[4955]: E0217 14:06:36.916608 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="131a5d3f-e605-4310-a2ff-2044b7afcdbd" containerName="gather" Feb 17 14:06:36 crc kubenswrapper[4955]: I0217 14:06:36.916615 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="131a5d3f-e605-4310-a2ff-2044b7afcdbd" containerName="gather" Feb 17 14:06:36 crc kubenswrapper[4955]: I0217 14:06:36.916846 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="131a5d3f-e605-4310-a2ff-2044b7afcdbd" containerName="copy" Feb 17 14:06:36 crc kubenswrapper[4955]: I0217 14:06:36.916870 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0ab4d69-eeed-461b-bc7d-1005eacc732f" containerName="keystone-cron" Feb 17 14:06:36 crc kubenswrapper[4955]: I0217 14:06:36.916883 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="131a5d3f-e605-4310-a2ff-2044b7afcdbd" containerName="gather" Feb 17 14:06:36 crc kubenswrapper[4955]: I0217 14:06:36.918488 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hp9l7" Feb 17 14:06:36 crc kubenswrapper[4955]: I0217 14:06:36.930183 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hp9l7"] Feb 17 14:06:37 crc kubenswrapper[4955]: I0217 14:06:37.109742 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/583d8d26-1e56-461a-8af3-ee25a5c8de96-utilities\") pod \"redhat-operators-hp9l7\" (UID: \"583d8d26-1e56-461a-8af3-ee25a5c8de96\") " pod="openshift-marketplace/redhat-operators-hp9l7" Feb 17 14:06:37 crc kubenswrapper[4955]: I0217 14:06:37.110421 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/583d8d26-1e56-461a-8af3-ee25a5c8de96-catalog-content\") pod \"redhat-operators-hp9l7\" (UID: \"583d8d26-1e56-461a-8af3-ee25a5c8de96\") " pod="openshift-marketplace/redhat-operators-hp9l7" Feb 17 14:06:37 crc kubenswrapper[4955]: I0217 14:06:37.110514 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2r9r\" (UniqueName: \"kubernetes.io/projected/583d8d26-1e56-461a-8af3-ee25a5c8de96-kube-api-access-x2r9r\") pod \"redhat-operators-hp9l7\" (UID: \"583d8d26-1e56-461a-8af3-ee25a5c8de96\") " pod="openshift-marketplace/redhat-operators-hp9l7" Feb 17 14:06:37 crc kubenswrapper[4955]: I0217 14:06:37.213293 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/583d8d26-1e56-461a-8af3-ee25a5c8de96-utilities\") pod \"redhat-operators-hp9l7\" (UID: \"583d8d26-1e56-461a-8af3-ee25a5c8de96\") " pod="openshift-marketplace/redhat-operators-hp9l7" Feb 17 14:06:37 crc kubenswrapper[4955]: I0217 14:06:37.213647 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/583d8d26-1e56-461a-8af3-ee25a5c8de96-catalog-content\") pod \"redhat-operators-hp9l7\" (UID: \"583d8d26-1e56-461a-8af3-ee25a5c8de96\") " pod="openshift-marketplace/redhat-operators-hp9l7" Feb 17 14:06:37 crc kubenswrapper[4955]: I0217 14:06:37.213699 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2r9r\" (UniqueName: \"kubernetes.io/projected/583d8d26-1e56-461a-8af3-ee25a5c8de96-kube-api-access-x2r9r\") pod \"redhat-operators-hp9l7\" (UID: \"583d8d26-1e56-461a-8af3-ee25a5c8de96\") " pod="openshift-marketplace/redhat-operators-hp9l7" Feb 17 14:06:37 crc kubenswrapper[4955]: I0217 14:06:37.214096 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/583d8d26-1e56-461a-8af3-ee25a5c8de96-utilities\") pod \"redhat-operators-hp9l7\" (UID: \"583d8d26-1e56-461a-8af3-ee25a5c8de96\") " pod="openshift-marketplace/redhat-operators-hp9l7" Feb 17 14:06:37 crc kubenswrapper[4955]: I0217 14:06:37.214238 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/583d8d26-1e56-461a-8af3-ee25a5c8de96-catalog-content\") pod \"redhat-operators-hp9l7\" (UID: \"583d8d26-1e56-461a-8af3-ee25a5c8de96\") " pod="openshift-marketplace/redhat-operators-hp9l7" Feb 17 14:06:37 crc kubenswrapper[4955]: I0217 14:06:37.248639 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2r9r\" (UniqueName: \"kubernetes.io/projected/583d8d26-1e56-461a-8af3-ee25a5c8de96-kube-api-access-x2r9r\") pod \"redhat-operators-hp9l7\" (UID: \"583d8d26-1e56-461a-8af3-ee25a5c8de96\") " pod="openshift-marketplace/redhat-operators-hp9l7" Feb 17 14:06:37 crc kubenswrapper[4955]: I0217 14:06:37.544743 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hp9l7" Feb 17 14:06:38 crc kubenswrapper[4955]: I0217 14:06:38.034808 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hp9l7"] Feb 17 14:06:38 crc kubenswrapper[4955]: I0217 14:06:38.224131 4955 scope.go:117] "RemoveContainer" containerID="fa86f3109bf152769bd9369dabc17d54290d99858877992d6ce0702550e664a2" Feb 17 14:06:38 crc kubenswrapper[4955]: E0217 14:06:38.225218 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 14:06:38 crc kubenswrapper[4955]: I0217 14:06:38.545214 4955 generic.go:334] "Generic (PLEG): container finished" podID="583d8d26-1e56-461a-8af3-ee25a5c8de96" containerID="d77128bb22edf5cf9049850f1f25d270de4095b6a036edb0b6b44ffa88fcc774" exitCode=0 Feb 17 14:06:38 crc kubenswrapper[4955]: I0217 14:06:38.545267 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hp9l7" event={"ID":"583d8d26-1e56-461a-8af3-ee25a5c8de96","Type":"ContainerDied","Data":"d77128bb22edf5cf9049850f1f25d270de4095b6a036edb0b6b44ffa88fcc774"} Feb 17 14:06:38 crc kubenswrapper[4955]: I0217 14:06:38.545310 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hp9l7" event={"ID":"583d8d26-1e56-461a-8af3-ee25a5c8de96","Type":"ContainerStarted","Data":"cc560cb92647fad9de88724f7bc405ae394a5306325ab25314df1b51cd0a141a"} Feb 17 14:06:38 crc kubenswrapper[4955]: I0217 14:06:38.547732 4955 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 17 14:06:40 crc kubenswrapper[4955]: I0217 14:06:40.562148 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hp9l7" event={"ID":"583d8d26-1e56-461a-8af3-ee25a5c8de96","Type":"ContainerStarted","Data":"3f4d9c939dfcb377e5c21397676465958cf861e461d1c41bd76ccc4e462a1f53"} Feb 17 14:06:42 crc kubenswrapper[4955]: I0217 14:06:42.583483 4955 generic.go:334] "Generic (PLEG): container finished" podID="583d8d26-1e56-461a-8af3-ee25a5c8de96" containerID="3f4d9c939dfcb377e5c21397676465958cf861e461d1c41bd76ccc4e462a1f53" exitCode=0 Feb 17 14:06:42 crc kubenswrapper[4955]: I0217 14:06:42.583560 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hp9l7" event={"ID":"583d8d26-1e56-461a-8af3-ee25a5c8de96","Type":"ContainerDied","Data":"3f4d9c939dfcb377e5c21397676465958cf861e461d1c41bd76ccc4e462a1f53"} Feb 17 14:06:43 crc kubenswrapper[4955]: I0217 14:06:43.596196 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hp9l7" event={"ID":"583d8d26-1e56-461a-8af3-ee25a5c8de96","Type":"ContainerStarted","Data":"f0616f3df1f40788501404ca9137800d5b3c7c5f141956f7fcde904753d95272"} Feb 17 14:06:43 crc kubenswrapper[4955]: I0217 14:06:43.621742 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hp9l7" podStartSLOduration=3.20912993 podStartE2EDuration="7.621724157s" podCreationTimestamp="2026-02-17 14:06:36 +0000 UTC" firstStartedPulling="2026-02-17 14:06:38.547522029 +0000 UTC m=+3737.070251572" lastFinishedPulling="2026-02-17 14:06:42.960116256 +0000 UTC m=+3741.482845799" observedRunningTime="2026-02-17 14:06:43.615742766 +0000 UTC m=+3742.138472309" watchObservedRunningTime="2026-02-17 14:06:43.621724157 +0000 UTC m=+3742.144453690" Feb 17 14:06:47 crc kubenswrapper[4955]: I0217 14:06:47.545479 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hp9l7" Feb 17 14:06:47 crc kubenswrapper[4955]: I0217 14:06:47.546110 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hp9l7" Feb 17 14:06:48 crc kubenswrapper[4955]: I0217 14:06:48.602023 4955 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hp9l7" podUID="583d8d26-1e56-461a-8af3-ee25a5c8de96" containerName="registry-server" probeResult="failure" output=< Feb 17 14:06:48 crc kubenswrapper[4955]: timeout: failed to connect service ":50051" within 1s Feb 17 14:06:48 crc kubenswrapper[4955]: > Feb 17 14:06:49 crc kubenswrapper[4955]: I0217 14:06:49.223734 4955 scope.go:117] "RemoveContainer" containerID="fa86f3109bf152769bd9369dabc17d54290d99858877992d6ce0702550e664a2" Feb 17 14:06:49 crc kubenswrapper[4955]: E0217 14:06:49.224107 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 14:06:57 crc kubenswrapper[4955]: I0217 14:06:57.592295 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hp9l7" Feb 17 14:06:57 crc kubenswrapper[4955]: I0217 14:06:57.640809 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hp9l7" Feb 17 14:06:57 crc kubenswrapper[4955]: I0217 14:06:57.830196 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hp9l7"] Feb 17 14:06:58 crc kubenswrapper[4955]: I0217 14:06:58.720437 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hp9l7" podUID="583d8d26-1e56-461a-8af3-ee25a5c8de96" containerName="registry-server" containerID="cri-o://f0616f3df1f40788501404ca9137800d5b3c7c5f141956f7fcde904753d95272" gracePeriod=2 Feb 17 14:06:59 crc kubenswrapper[4955]: I0217 14:06:59.168857 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hp9l7" Feb 17 14:06:59 crc kubenswrapper[4955]: I0217 14:06:59.329744 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2r9r\" (UniqueName: \"kubernetes.io/projected/583d8d26-1e56-461a-8af3-ee25a5c8de96-kube-api-access-x2r9r\") pod \"583d8d26-1e56-461a-8af3-ee25a5c8de96\" (UID: \"583d8d26-1e56-461a-8af3-ee25a5c8de96\") " Feb 17 14:06:59 crc kubenswrapper[4955]: I0217 14:06:59.330001 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/583d8d26-1e56-461a-8af3-ee25a5c8de96-catalog-content\") pod \"583d8d26-1e56-461a-8af3-ee25a5c8de96\" (UID: \"583d8d26-1e56-461a-8af3-ee25a5c8de96\") " Feb 17 14:06:59 crc kubenswrapper[4955]: I0217 14:06:59.330058 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/583d8d26-1e56-461a-8af3-ee25a5c8de96-utilities\") pod \"583d8d26-1e56-461a-8af3-ee25a5c8de96\" (UID: \"583d8d26-1e56-461a-8af3-ee25a5c8de96\") " Feb 17 14:06:59 crc kubenswrapper[4955]: I0217 14:06:59.330890 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/583d8d26-1e56-461a-8af3-ee25a5c8de96-utilities" (OuterVolumeSpecName: "utilities") pod "583d8d26-1e56-461a-8af3-ee25a5c8de96" (UID: "583d8d26-1e56-461a-8af3-ee25a5c8de96"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 14:06:59 crc kubenswrapper[4955]: I0217 14:06:59.341606 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/583d8d26-1e56-461a-8af3-ee25a5c8de96-kube-api-access-x2r9r" (OuterVolumeSpecName: "kube-api-access-x2r9r") pod "583d8d26-1e56-461a-8af3-ee25a5c8de96" (UID: "583d8d26-1e56-461a-8af3-ee25a5c8de96"). InnerVolumeSpecName "kube-api-access-x2r9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 14:06:59 crc kubenswrapper[4955]: I0217 14:06:59.433514 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2r9r\" (UniqueName: \"kubernetes.io/projected/583d8d26-1e56-461a-8af3-ee25a5c8de96-kube-api-access-x2r9r\") on node \"crc\" DevicePath \"\"" Feb 17 14:06:59 crc kubenswrapper[4955]: I0217 14:06:59.433537 4955 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/583d8d26-1e56-461a-8af3-ee25a5c8de96-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 14:06:59 crc kubenswrapper[4955]: I0217 14:06:59.463361 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/583d8d26-1e56-461a-8af3-ee25a5c8de96-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "583d8d26-1e56-461a-8af3-ee25a5c8de96" (UID: "583d8d26-1e56-461a-8af3-ee25a5c8de96"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 14:06:59 crc kubenswrapper[4955]: I0217 14:06:59.535392 4955 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/583d8d26-1e56-461a-8af3-ee25a5c8de96-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 14:06:59 crc kubenswrapper[4955]: I0217 14:06:59.731852 4955 generic.go:334] "Generic (PLEG): container finished" podID="583d8d26-1e56-461a-8af3-ee25a5c8de96" containerID="f0616f3df1f40788501404ca9137800d5b3c7c5f141956f7fcde904753d95272" exitCode=0 Feb 17 14:06:59 crc kubenswrapper[4955]: I0217 14:06:59.731913 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hp9l7" event={"ID":"583d8d26-1e56-461a-8af3-ee25a5c8de96","Type":"ContainerDied","Data":"f0616f3df1f40788501404ca9137800d5b3c7c5f141956f7fcde904753d95272"} Feb 17 14:06:59 crc kubenswrapper[4955]: I0217 14:06:59.731953 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hp9l7" event={"ID":"583d8d26-1e56-461a-8af3-ee25a5c8de96","Type":"ContainerDied","Data":"cc560cb92647fad9de88724f7bc405ae394a5306325ab25314df1b51cd0a141a"} Feb 17 14:06:59 crc kubenswrapper[4955]: I0217 14:06:59.731981 4955 scope.go:117] "RemoveContainer" containerID="f0616f3df1f40788501404ca9137800d5b3c7c5f141956f7fcde904753d95272" Feb 17 14:06:59 crc kubenswrapper[4955]: I0217 14:06:59.732205 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hp9l7" Feb 17 14:06:59 crc kubenswrapper[4955]: I0217 14:06:59.771061 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hp9l7"] Feb 17 14:06:59 crc kubenswrapper[4955]: I0217 14:06:59.772965 4955 scope.go:117] "RemoveContainer" containerID="3f4d9c939dfcb377e5c21397676465958cf861e461d1c41bd76ccc4e462a1f53" Feb 17 14:06:59 crc kubenswrapper[4955]: I0217 14:06:59.792250 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hp9l7"] Feb 17 14:06:59 crc kubenswrapper[4955]: I0217 14:06:59.798233 4955 scope.go:117] "RemoveContainer" containerID="d77128bb22edf5cf9049850f1f25d270de4095b6a036edb0b6b44ffa88fcc774" Feb 17 14:06:59 crc kubenswrapper[4955]: I0217 14:06:59.838414 4955 scope.go:117] "RemoveContainer" containerID="f0616f3df1f40788501404ca9137800d5b3c7c5f141956f7fcde904753d95272" Feb 17 14:06:59 crc kubenswrapper[4955]: E0217 14:06:59.838844 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0616f3df1f40788501404ca9137800d5b3c7c5f141956f7fcde904753d95272\": container with ID starting with f0616f3df1f40788501404ca9137800d5b3c7c5f141956f7fcde904753d95272 not found: ID does not exist" containerID="f0616f3df1f40788501404ca9137800d5b3c7c5f141956f7fcde904753d95272" Feb 17 14:06:59 crc kubenswrapper[4955]: I0217 14:06:59.838886 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0616f3df1f40788501404ca9137800d5b3c7c5f141956f7fcde904753d95272"} err="failed to get container status \"f0616f3df1f40788501404ca9137800d5b3c7c5f141956f7fcde904753d95272\": rpc error: code = NotFound desc = could not find container \"f0616f3df1f40788501404ca9137800d5b3c7c5f141956f7fcde904753d95272\": container with ID starting with f0616f3df1f40788501404ca9137800d5b3c7c5f141956f7fcde904753d95272 not found: ID does not exist" Feb 17 14:06:59 crc kubenswrapper[4955]: I0217 14:06:59.838913 4955 scope.go:117] "RemoveContainer" containerID="3f4d9c939dfcb377e5c21397676465958cf861e461d1c41bd76ccc4e462a1f53" Feb 17 14:06:59 crc kubenswrapper[4955]: E0217 14:06:59.839320 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f4d9c939dfcb377e5c21397676465958cf861e461d1c41bd76ccc4e462a1f53\": container with ID starting with 3f4d9c939dfcb377e5c21397676465958cf861e461d1c41bd76ccc4e462a1f53 not found: ID does not exist" containerID="3f4d9c939dfcb377e5c21397676465958cf861e461d1c41bd76ccc4e462a1f53" Feb 17 14:06:59 crc kubenswrapper[4955]: I0217 14:06:59.839349 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f4d9c939dfcb377e5c21397676465958cf861e461d1c41bd76ccc4e462a1f53"} err="failed to get container status \"3f4d9c939dfcb377e5c21397676465958cf861e461d1c41bd76ccc4e462a1f53\": rpc error: code = NotFound desc = could not find container \"3f4d9c939dfcb377e5c21397676465958cf861e461d1c41bd76ccc4e462a1f53\": container with ID starting with 3f4d9c939dfcb377e5c21397676465958cf861e461d1c41bd76ccc4e462a1f53 not found: ID does not exist" Feb 17 14:06:59 crc kubenswrapper[4955]: I0217 14:06:59.839366 4955 scope.go:117] "RemoveContainer" containerID="d77128bb22edf5cf9049850f1f25d270de4095b6a036edb0b6b44ffa88fcc774" Feb 17 14:06:59 crc kubenswrapper[4955]: E0217 14:06:59.839632 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d77128bb22edf5cf9049850f1f25d270de4095b6a036edb0b6b44ffa88fcc774\": container with ID starting with d77128bb22edf5cf9049850f1f25d270de4095b6a036edb0b6b44ffa88fcc774 not found: ID does not exist" containerID="d77128bb22edf5cf9049850f1f25d270de4095b6a036edb0b6b44ffa88fcc774" Feb 17 14:06:59 crc kubenswrapper[4955]: I0217 14:06:59.839657 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d77128bb22edf5cf9049850f1f25d270de4095b6a036edb0b6b44ffa88fcc774"} err="failed to get container status \"d77128bb22edf5cf9049850f1f25d270de4095b6a036edb0b6b44ffa88fcc774\": rpc error: code = NotFound desc = could not find container \"d77128bb22edf5cf9049850f1f25d270de4095b6a036edb0b6b44ffa88fcc774\": container with ID starting with d77128bb22edf5cf9049850f1f25d270de4095b6a036edb0b6b44ffa88fcc774 not found: ID does not exist" Feb 17 14:07:00 crc kubenswrapper[4955]: I0217 14:07:00.237642 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="583d8d26-1e56-461a-8af3-ee25a5c8de96" path="/var/lib/kubelet/pods/583d8d26-1e56-461a-8af3-ee25a5c8de96/volumes" Feb 17 14:07:02 crc kubenswrapper[4955]: I0217 14:07:02.223471 4955 scope.go:117] "RemoveContainer" containerID="fa86f3109bf152769bd9369dabc17d54290d99858877992d6ce0702550e664a2" Feb 17 14:07:02 crc kubenswrapper[4955]: E0217 14:07:02.224170 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 14:07:15 crc kubenswrapper[4955]: I0217 14:07:15.246672 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-sncmd/must-gather-dd9pl"] Feb 17 14:07:15 crc kubenswrapper[4955]: E0217 14:07:15.247808 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="583d8d26-1e56-461a-8af3-ee25a5c8de96" containerName="extract-utilities" Feb 17 14:07:15 crc kubenswrapper[4955]: I0217 14:07:15.247825 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="583d8d26-1e56-461a-8af3-ee25a5c8de96" containerName="extract-utilities" Feb 17 14:07:15 crc kubenswrapper[4955]: E0217 14:07:15.247851 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="583d8d26-1e56-461a-8af3-ee25a5c8de96" containerName="extract-content" Feb 17 14:07:15 crc kubenswrapper[4955]: I0217 14:07:15.247862 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="583d8d26-1e56-461a-8af3-ee25a5c8de96" containerName="extract-content" Feb 17 14:07:15 crc kubenswrapper[4955]: E0217 14:07:15.247899 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="583d8d26-1e56-461a-8af3-ee25a5c8de96" containerName="registry-server" Feb 17 14:07:15 crc kubenswrapper[4955]: I0217 14:07:15.247907 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="583d8d26-1e56-461a-8af3-ee25a5c8de96" containerName="registry-server" Feb 17 14:07:15 crc kubenswrapper[4955]: I0217 14:07:15.265030 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="583d8d26-1e56-461a-8af3-ee25a5c8de96" containerName="registry-server" Feb 17 14:07:15 crc kubenswrapper[4955]: I0217 14:07:15.266854 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sncmd/must-gather-dd9pl" Feb 17 14:07:15 crc kubenswrapper[4955]: I0217 14:07:15.269526 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-sncmd"/"openshift-service-ca.crt" Feb 17 14:07:15 crc kubenswrapper[4955]: I0217 14:07:15.269831 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-sncmd"/"default-dockercfg-pmtwb" Feb 17 14:07:15 crc kubenswrapper[4955]: I0217 14:07:15.270296 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-sncmd"/"kube-root-ca.crt" Feb 17 14:07:15 crc kubenswrapper[4955]: I0217 14:07:15.295804 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-sncmd/must-gather-dd9pl"] Feb 17 14:07:15 crc kubenswrapper[4955]: I0217 14:07:15.440637 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2ntf\" (UniqueName: \"kubernetes.io/projected/0bdb56fb-1c91-4402-a733-8cbcb46e21ca-kube-api-access-j2ntf\") pod \"must-gather-dd9pl\" (UID: \"0bdb56fb-1c91-4402-a733-8cbcb46e21ca\") " pod="openshift-must-gather-sncmd/must-gather-dd9pl" Feb 17 14:07:15 crc kubenswrapper[4955]: I0217 14:07:15.440731 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0bdb56fb-1c91-4402-a733-8cbcb46e21ca-must-gather-output\") pod \"must-gather-dd9pl\" (UID: \"0bdb56fb-1c91-4402-a733-8cbcb46e21ca\") " pod="openshift-must-gather-sncmd/must-gather-dd9pl" Feb 17 14:07:15 crc kubenswrapper[4955]: I0217 14:07:15.542624 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2ntf\" (UniqueName: \"kubernetes.io/projected/0bdb56fb-1c91-4402-a733-8cbcb46e21ca-kube-api-access-j2ntf\") pod \"must-gather-dd9pl\" (UID: \"0bdb56fb-1c91-4402-a733-8cbcb46e21ca\") " pod="openshift-must-gather-sncmd/must-gather-dd9pl" Feb 17 14:07:15 crc kubenswrapper[4955]: I0217 14:07:15.542739 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0bdb56fb-1c91-4402-a733-8cbcb46e21ca-must-gather-output\") pod \"must-gather-dd9pl\" (UID: \"0bdb56fb-1c91-4402-a733-8cbcb46e21ca\") " pod="openshift-must-gather-sncmd/must-gather-dd9pl" Feb 17 14:07:15 crc kubenswrapper[4955]: I0217 14:07:15.543607 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0bdb56fb-1c91-4402-a733-8cbcb46e21ca-must-gather-output\") pod \"must-gather-dd9pl\" (UID: \"0bdb56fb-1c91-4402-a733-8cbcb46e21ca\") " pod="openshift-must-gather-sncmd/must-gather-dd9pl" Feb 17 14:07:15 crc kubenswrapper[4955]: I0217 14:07:15.570739 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2ntf\" (UniqueName: \"kubernetes.io/projected/0bdb56fb-1c91-4402-a733-8cbcb46e21ca-kube-api-access-j2ntf\") pod \"must-gather-dd9pl\" (UID: \"0bdb56fb-1c91-4402-a733-8cbcb46e21ca\") " pod="openshift-must-gather-sncmd/must-gather-dd9pl" Feb 17 14:07:15 crc kubenswrapper[4955]: I0217 14:07:15.594022 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sncmd/must-gather-dd9pl" Feb 17 14:07:16 crc kubenswrapper[4955]: I0217 14:07:16.086236 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-sncmd/must-gather-dd9pl"] Feb 17 14:07:16 crc kubenswrapper[4955]: I0217 14:07:16.878674 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sncmd/must-gather-dd9pl" event={"ID":"0bdb56fb-1c91-4402-a733-8cbcb46e21ca","Type":"ContainerStarted","Data":"1e5d05066dd216411721ec16d11df4de4f618415b573b58817c27c2628c874d4"} Feb 17 14:07:16 crc kubenswrapper[4955]: I0217 14:07:16.879054 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sncmd/must-gather-dd9pl" event={"ID":"0bdb56fb-1c91-4402-a733-8cbcb46e21ca","Type":"ContainerStarted","Data":"902293f5458457663ced1a4585a7672e6d53e697161b084cd04012aa0d2feb0f"} Feb 17 14:07:17 crc kubenswrapper[4955]: I0217 14:07:17.222593 4955 scope.go:117] "RemoveContainer" containerID="fa86f3109bf152769bd9369dabc17d54290d99858877992d6ce0702550e664a2" Feb 17 14:07:17 crc kubenswrapper[4955]: E0217 14:07:17.223269 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 14:07:17 crc kubenswrapper[4955]: I0217 14:07:17.891251 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sncmd/must-gather-dd9pl" event={"ID":"0bdb56fb-1c91-4402-a733-8cbcb46e21ca","Type":"ContainerStarted","Data":"ce8d24b61088b3bc0703e041dc3fea0ef78865e1dd1d0c9bfc1604ee13a3a69c"} Feb 17 14:07:17 crc kubenswrapper[4955]: I0217 14:07:17.913719 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-sncmd/must-gather-dd9pl" podStartSLOduration=2.913703866 podStartE2EDuration="2.913703866s" podCreationTimestamp="2026-02-17 14:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 14:07:17.909481315 +0000 UTC m=+3776.432210858" watchObservedRunningTime="2026-02-17 14:07:17.913703866 +0000 UTC m=+3776.436433409" Feb 17 14:07:20 crc kubenswrapper[4955]: I0217 14:07:20.569666 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-sncmd/crc-debug-kgmbf"] Feb 17 14:07:20 crc kubenswrapper[4955]: I0217 14:07:20.571916 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sncmd/crc-debug-kgmbf" Feb 17 14:07:20 crc kubenswrapper[4955]: I0217 14:07:20.642845 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/64b28f78-b649-4ce2-949e-29ebd8fe5062-host\") pod \"crc-debug-kgmbf\" (UID: \"64b28f78-b649-4ce2-949e-29ebd8fe5062\") " pod="openshift-must-gather-sncmd/crc-debug-kgmbf" Feb 17 14:07:20 crc kubenswrapper[4955]: I0217 14:07:20.643341 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xkp5\" (UniqueName: \"kubernetes.io/projected/64b28f78-b649-4ce2-949e-29ebd8fe5062-kube-api-access-8xkp5\") pod \"crc-debug-kgmbf\" (UID: \"64b28f78-b649-4ce2-949e-29ebd8fe5062\") " pod="openshift-must-gather-sncmd/crc-debug-kgmbf" Feb 17 14:07:20 crc kubenswrapper[4955]: I0217 14:07:20.745604 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/64b28f78-b649-4ce2-949e-29ebd8fe5062-host\") pod \"crc-debug-kgmbf\" (UID: \"64b28f78-b649-4ce2-949e-29ebd8fe5062\") " pod="openshift-must-gather-sncmd/crc-debug-kgmbf" Feb 17 14:07:20 crc kubenswrapper[4955]: I0217 14:07:20.745732 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xkp5\" (UniqueName: \"kubernetes.io/projected/64b28f78-b649-4ce2-949e-29ebd8fe5062-kube-api-access-8xkp5\") pod \"crc-debug-kgmbf\" (UID: \"64b28f78-b649-4ce2-949e-29ebd8fe5062\") " pod="openshift-must-gather-sncmd/crc-debug-kgmbf" Feb 17 14:07:20 crc kubenswrapper[4955]: I0217 14:07:20.745742 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/64b28f78-b649-4ce2-949e-29ebd8fe5062-host\") pod \"crc-debug-kgmbf\" (UID: \"64b28f78-b649-4ce2-949e-29ebd8fe5062\") " pod="openshift-must-gather-sncmd/crc-debug-kgmbf" Feb 17 14:07:20 crc kubenswrapper[4955]: I0217 14:07:20.770312 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xkp5\" (UniqueName: \"kubernetes.io/projected/64b28f78-b649-4ce2-949e-29ebd8fe5062-kube-api-access-8xkp5\") pod \"crc-debug-kgmbf\" (UID: \"64b28f78-b649-4ce2-949e-29ebd8fe5062\") " pod="openshift-must-gather-sncmd/crc-debug-kgmbf" Feb 17 14:07:20 crc kubenswrapper[4955]: I0217 14:07:20.901964 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sncmd/crc-debug-kgmbf" Feb 17 14:07:20 crc kubenswrapper[4955]: W0217 14:07:20.938322 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64b28f78_b649_4ce2_949e_29ebd8fe5062.slice/crio-f4ab12fed916622d861b00ed9c1ad4053bff1c6c4aee8ca36f07cca6ab5553e3 WatchSource:0}: Error finding container f4ab12fed916622d861b00ed9c1ad4053bff1c6c4aee8ca36f07cca6ab5553e3: Status 404 returned error can't find the container with id f4ab12fed916622d861b00ed9c1ad4053bff1c6c4aee8ca36f07cca6ab5553e3 Feb 17 14:07:21 crc kubenswrapper[4955]: I0217 14:07:21.928614 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sncmd/crc-debug-kgmbf" event={"ID":"64b28f78-b649-4ce2-949e-29ebd8fe5062","Type":"ContainerStarted","Data":"3ee69b90a7c7d296739f074bae24e7e64696ee6730c62e510684aa440cb2b8a0"} Feb 17 14:07:21 crc kubenswrapper[4955]: I0217 14:07:21.929169 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sncmd/crc-debug-kgmbf" event={"ID":"64b28f78-b649-4ce2-949e-29ebd8fe5062","Type":"ContainerStarted","Data":"f4ab12fed916622d861b00ed9c1ad4053bff1c6c4aee8ca36f07cca6ab5553e3"} Feb 17 14:07:21 crc kubenswrapper[4955]: I0217 14:07:21.946713 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-sncmd/crc-debug-kgmbf" podStartSLOduration=1.946689224 podStartE2EDuration="1.946689224s" podCreationTimestamp="2026-02-17 14:07:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 14:07:21.944264615 +0000 UTC m=+3780.466994168" watchObservedRunningTime="2026-02-17 14:07:21.946689224 +0000 UTC m=+3780.469418767" Feb 17 14:07:29 crc kubenswrapper[4955]: I0217 14:07:29.222994 4955 scope.go:117] "RemoveContainer" containerID="fa86f3109bf152769bd9369dabc17d54290d99858877992d6ce0702550e664a2" Feb 17 14:07:29 crc kubenswrapper[4955]: E0217 14:07:29.223907 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 14:07:41 crc kubenswrapper[4955]: I0217 14:07:41.223547 4955 scope.go:117] "RemoveContainer" containerID="fa86f3109bf152769bd9369dabc17d54290d99858877992d6ce0702550e664a2" Feb 17 14:07:41 crc kubenswrapper[4955]: E0217 14:07:41.224260 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 14:07:53 crc kubenswrapper[4955]: I0217 14:07:53.223502 4955 scope.go:117] "RemoveContainer" containerID="fa86f3109bf152769bd9369dabc17d54290d99858877992d6ce0702550e664a2" Feb 17 14:07:53 crc kubenswrapper[4955]: E0217 14:07:53.224303 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 14:07:57 crc kubenswrapper[4955]: I0217 14:07:57.251535 4955 generic.go:334] "Generic (PLEG): container finished" podID="64b28f78-b649-4ce2-949e-29ebd8fe5062" containerID="3ee69b90a7c7d296739f074bae24e7e64696ee6730c62e510684aa440cb2b8a0" exitCode=0 Feb 17 14:07:57 crc kubenswrapper[4955]: I0217 14:07:57.251654 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sncmd/crc-debug-kgmbf" event={"ID":"64b28f78-b649-4ce2-949e-29ebd8fe5062","Type":"ContainerDied","Data":"3ee69b90a7c7d296739f074bae24e7e64696ee6730c62e510684aa440cb2b8a0"} Feb 17 14:07:58 crc kubenswrapper[4955]: I0217 14:07:58.411503 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sncmd/crc-debug-kgmbf" Feb 17 14:07:58 crc kubenswrapper[4955]: I0217 14:07:58.442065 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-sncmd/crc-debug-kgmbf"] Feb 17 14:07:58 crc kubenswrapper[4955]: I0217 14:07:58.453082 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-sncmd/crc-debug-kgmbf"] Feb 17 14:07:58 crc kubenswrapper[4955]: I0217 14:07:58.534019 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/64b28f78-b649-4ce2-949e-29ebd8fe5062-host\") pod \"64b28f78-b649-4ce2-949e-29ebd8fe5062\" (UID: \"64b28f78-b649-4ce2-949e-29ebd8fe5062\") " Feb 17 14:07:58 crc kubenswrapper[4955]: I0217 14:07:58.534153 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xkp5\" (UniqueName: \"kubernetes.io/projected/64b28f78-b649-4ce2-949e-29ebd8fe5062-kube-api-access-8xkp5\") pod \"64b28f78-b649-4ce2-949e-29ebd8fe5062\" (UID: \"64b28f78-b649-4ce2-949e-29ebd8fe5062\") " Feb 17 14:07:58 crc kubenswrapper[4955]: I0217 14:07:58.534165 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/64b28f78-b649-4ce2-949e-29ebd8fe5062-host" (OuterVolumeSpecName: "host") pod "64b28f78-b649-4ce2-949e-29ebd8fe5062" (UID: "64b28f78-b649-4ce2-949e-29ebd8fe5062"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 14:07:58 crc kubenswrapper[4955]: I0217 14:07:58.535960 4955 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/64b28f78-b649-4ce2-949e-29ebd8fe5062-host\") on node \"crc\" DevicePath \"\"" Feb 17 14:07:58 crc kubenswrapper[4955]: I0217 14:07:58.541763 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64b28f78-b649-4ce2-949e-29ebd8fe5062-kube-api-access-8xkp5" (OuterVolumeSpecName: "kube-api-access-8xkp5") pod "64b28f78-b649-4ce2-949e-29ebd8fe5062" (UID: "64b28f78-b649-4ce2-949e-29ebd8fe5062"). InnerVolumeSpecName "kube-api-access-8xkp5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 14:07:58 crc kubenswrapper[4955]: I0217 14:07:58.637750 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xkp5\" (UniqueName: \"kubernetes.io/projected/64b28f78-b649-4ce2-949e-29ebd8fe5062-kube-api-access-8xkp5\") on node \"crc\" DevicePath \"\"" Feb 17 14:07:59 crc kubenswrapper[4955]: I0217 14:07:59.270744 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4ab12fed916622d861b00ed9c1ad4053bff1c6c4aee8ca36f07cca6ab5553e3" Feb 17 14:07:59 crc kubenswrapper[4955]: I0217 14:07:59.271476 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sncmd/crc-debug-kgmbf" Feb 17 14:07:59 crc kubenswrapper[4955]: I0217 14:07:59.732536 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-sncmd/crc-debug-46h9f"] Feb 17 14:07:59 crc kubenswrapper[4955]: E0217 14:07:59.733021 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64b28f78-b649-4ce2-949e-29ebd8fe5062" containerName="container-00" Feb 17 14:07:59 crc kubenswrapper[4955]: I0217 14:07:59.733038 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="64b28f78-b649-4ce2-949e-29ebd8fe5062" containerName="container-00" Feb 17 14:07:59 crc kubenswrapper[4955]: I0217 14:07:59.733266 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="64b28f78-b649-4ce2-949e-29ebd8fe5062" containerName="container-00" Feb 17 14:07:59 crc kubenswrapper[4955]: I0217 14:07:59.734125 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sncmd/crc-debug-46h9f" Feb 17 14:07:59 crc kubenswrapper[4955]: I0217 14:07:59.861513 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fbd8383c-94c7-49e9-80e9-a63c96476c69-host\") pod \"crc-debug-46h9f\" (UID: \"fbd8383c-94c7-49e9-80e9-a63c96476c69\") " pod="openshift-must-gather-sncmd/crc-debug-46h9f" Feb 17 14:07:59 crc kubenswrapper[4955]: I0217 14:07:59.861590 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vshm\" (UniqueName: \"kubernetes.io/projected/fbd8383c-94c7-49e9-80e9-a63c96476c69-kube-api-access-8vshm\") pod \"crc-debug-46h9f\" (UID: \"fbd8383c-94c7-49e9-80e9-a63c96476c69\") " pod="openshift-must-gather-sncmd/crc-debug-46h9f" Feb 17 14:07:59 crc kubenswrapper[4955]: I0217 14:07:59.963467 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vshm\" (UniqueName: \"kubernetes.io/projected/fbd8383c-94c7-49e9-80e9-a63c96476c69-kube-api-access-8vshm\") pod \"crc-debug-46h9f\" (UID: \"fbd8383c-94c7-49e9-80e9-a63c96476c69\") " pod="openshift-must-gather-sncmd/crc-debug-46h9f" Feb 17 14:07:59 crc kubenswrapper[4955]: I0217 14:07:59.963635 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fbd8383c-94c7-49e9-80e9-a63c96476c69-host\") pod \"crc-debug-46h9f\" (UID: \"fbd8383c-94c7-49e9-80e9-a63c96476c69\") " pod="openshift-must-gather-sncmd/crc-debug-46h9f" Feb 17 14:07:59 crc kubenswrapper[4955]: I0217 14:07:59.963742 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fbd8383c-94c7-49e9-80e9-a63c96476c69-host\") pod \"crc-debug-46h9f\" (UID: \"fbd8383c-94c7-49e9-80e9-a63c96476c69\") " pod="openshift-must-gather-sncmd/crc-debug-46h9f" Feb 17 14:07:59 crc kubenswrapper[4955]: I0217 14:07:59.982672 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vshm\" (UniqueName: \"kubernetes.io/projected/fbd8383c-94c7-49e9-80e9-a63c96476c69-kube-api-access-8vshm\") pod \"crc-debug-46h9f\" (UID: \"fbd8383c-94c7-49e9-80e9-a63c96476c69\") " pod="openshift-must-gather-sncmd/crc-debug-46h9f" Feb 17 14:08:00 crc kubenswrapper[4955]: I0217 14:08:00.050169 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sncmd/crc-debug-46h9f" Feb 17 14:08:00 crc kubenswrapper[4955]: I0217 14:08:00.234836 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64b28f78-b649-4ce2-949e-29ebd8fe5062" path="/var/lib/kubelet/pods/64b28f78-b649-4ce2-949e-29ebd8fe5062/volumes" Feb 17 14:08:00 crc kubenswrapper[4955]: I0217 14:08:00.280184 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sncmd/crc-debug-46h9f" event={"ID":"fbd8383c-94c7-49e9-80e9-a63c96476c69","Type":"ContainerStarted","Data":"253d6a4a4cf6e3b16468d917971dc5e21c9105d2d6327593e54ef07d96636ce5"} Feb 17 14:08:01 crc kubenswrapper[4955]: I0217 14:08:01.289631 4955 generic.go:334] "Generic (PLEG): container finished" podID="fbd8383c-94c7-49e9-80e9-a63c96476c69" containerID="7cef3d0826b115af833f956740b6e85ca1d41ec37c7cfe7a3ab909fef22e17ac" exitCode=0 Feb 17 14:08:01 crc kubenswrapper[4955]: I0217 14:08:01.289732 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sncmd/crc-debug-46h9f" event={"ID":"fbd8383c-94c7-49e9-80e9-a63c96476c69","Type":"ContainerDied","Data":"7cef3d0826b115af833f956740b6e85ca1d41ec37c7cfe7a3ab909fef22e17ac"} Feb 17 14:08:01 crc kubenswrapper[4955]: I0217 14:08:01.734353 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-sncmd/crc-debug-46h9f"] Feb 17 14:08:01 crc kubenswrapper[4955]: I0217 14:08:01.744215 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-sncmd/crc-debug-46h9f"] Feb 17 14:08:02 crc kubenswrapper[4955]: I0217 14:08:02.398642 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sncmd/crc-debug-46h9f" Feb 17 14:08:02 crc kubenswrapper[4955]: I0217 14:08:02.512618 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8vshm\" (UniqueName: \"kubernetes.io/projected/fbd8383c-94c7-49e9-80e9-a63c96476c69-kube-api-access-8vshm\") pod \"fbd8383c-94c7-49e9-80e9-a63c96476c69\" (UID: \"fbd8383c-94c7-49e9-80e9-a63c96476c69\") " Feb 17 14:08:02 crc kubenswrapper[4955]: I0217 14:08:02.512798 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fbd8383c-94c7-49e9-80e9-a63c96476c69-host\") pod \"fbd8383c-94c7-49e9-80e9-a63c96476c69\" (UID: \"fbd8383c-94c7-49e9-80e9-a63c96476c69\") " Feb 17 14:08:02 crc kubenswrapper[4955]: I0217 14:08:02.512971 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fbd8383c-94c7-49e9-80e9-a63c96476c69-host" (OuterVolumeSpecName: "host") pod "fbd8383c-94c7-49e9-80e9-a63c96476c69" (UID: "fbd8383c-94c7-49e9-80e9-a63c96476c69"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 14:08:02 crc kubenswrapper[4955]: I0217 14:08:02.513463 4955 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fbd8383c-94c7-49e9-80e9-a63c96476c69-host\") on node \"crc\" DevicePath \"\"" Feb 17 14:08:02 crc kubenswrapper[4955]: I0217 14:08:02.524034 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbd8383c-94c7-49e9-80e9-a63c96476c69-kube-api-access-8vshm" (OuterVolumeSpecName: "kube-api-access-8vshm") pod "fbd8383c-94c7-49e9-80e9-a63c96476c69" (UID: "fbd8383c-94c7-49e9-80e9-a63c96476c69"). InnerVolumeSpecName "kube-api-access-8vshm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 14:08:02 crc kubenswrapper[4955]: I0217 14:08:02.614701 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8vshm\" (UniqueName: \"kubernetes.io/projected/fbd8383c-94c7-49e9-80e9-a63c96476c69-kube-api-access-8vshm\") on node \"crc\" DevicePath \"\"" Feb 17 14:08:02 crc kubenswrapper[4955]: I0217 14:08:02.939890 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-sncmd/crc-debug-k527r"] Feb 17 14:08:02 crc kubenswrapper[4955]: E0217 14:08:02.940775 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbd8383c-94c7-49e9-80e9-a63c96476c69" containerName="container-00" Feb 17 14:08:02 crc kubenswrapper[4955]: I0217 14:08:02.940867 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbd8383c-94c7-49e9-80e9-a63c96476c69" containerName="container-00" Feb 17 14:08:02 crc kubenswrapper[4955]: I0217 14:08:02.941122 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbd8383c-94c7-49e9-80e9-a63c96476c69" containerName="container-00" Feb 17 14:08:02 crc kubenswrapper[4955]: I0217 14:08:02.941721 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sncmd/crc-debug-k527r" Feb 17 14:08:03 crc kubenswrapper[4955]: I0217 14:08:03.020879 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfg4m\" (UniqueName: \"kubernetes.io/projected/8d0a24cf-b58d-4493-9395-f3775cf25594-kube-api-access-dfg4m\") pod \"crc-debug-k527r\" (UID: \"8d0a24cf-b58d-4493-9395-f3775cf25594\") " pod="openshift-must-gather-sncmd/crc-debug-k527r" Feb 17 14:08:03 crc kubenswrapper[4955]: I0217 14:08:03.020996 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8d0a24cf-b58d-4493-9395-f3775cf25594-host\") pod \"crc-debug-k527r\" (UID: \"8d0a24cf-b58d-4493-9395-f3775cf25594\") " pod="openshift-must-gather-sncmd/crc-debug-k527r" Feb 17 14:08:03 crc kubenswrapper[4955]: I0217 14:08:03.123180 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfg4m\" (UniqueName: \"kubernetes.io/projected/8d0a24cf-b58d-4493-9395-f3775cf25594-kube-api-access-dfg4m\") pod \"crc-debug-k527r\" (UID: \"8d0a24cf-b58d-4493-9395-f3775cf25594\") " pod="openshift-must-gather-sncmd/crc-debug-k527r" Feb 17 14:08:03 crc kubenswrapper[4955]: I0217 14:08:03.123565 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8d0a24cf-b58d-4493-9395-f3775cf25594-host\") pod \"crc-debug-k527r\" (UID: \"8d0a24cf-b58d-4493-9395-f3775cf25594\") " pod="openshift-must-gather-sncmd/crc-debug-k527r" Feb 17 14:08:03 crc kubenswrapper[4955]: I0217 14:08:03.127121 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8d0a24cf-b58d-4493-9395-f3775cf25594-host\") pod \"crc-debug-k527r\" (UID: \"8d0a24cf-b58d-4493-9395-f3775cf25594\") " pod="openshift-must-gather-sncmd/crc-debug-k527r" Feb 17 14:08:03 crc kubenswrapper[4955]: I0217 14:08:03.157502 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfg4m\" (UniqueName: \"kubernetes.io/projected/8d0a24cf-b58d-4493-9395-f3775cf25594-kube-api-access-dfg4m\") pod \"crc-debug-k527r\" (UID: \"8d0a24cf-b58d-4493-9395-f3775cf25594\") " pod="openshift-must-gather-sncmd/crc-debug-k527r" Feb 17 14:08:03 crc kubenswrapper[4955]: I0217 14:08:03.259294 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sncmd/crc-debug-k527r" Feb 17 14:08:03 crc kubenswrapper[4955]: I0217 14:08:03.310361 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sncmd/crc-debug-46h9f" Feb 17 14:08:03 crc kubenswrapper[4955]: W0217 14:08:03.310893 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8d0a24cf_b58d_4493_9395_f3775cf25594.slice/crio-1547f87fd47d1901ef5d304677bfd5d8c99bf4b57d1456315d31c9cde583bf41 WatchSource:0}: Error finding container 1547f87fd47d1901ef5d304677bfd5d8c99bf4b57d1456315d31c9cde583bf41: Status 404 returned error can't find the container with id 1547f87fd47d1901ef5d304677bfd5d8c99bf4b57d1456315d31c9cde583bf41 Feb 17 14:08:03 crc kubenswrapper[4955]: I0217 14:08:03.310295 4955 scope.go:117] "RemoveContainer" containerID="7cef3d0826b115af833f956740b6e85ca1d41ec37c7cfe7a3ab909fef22e17ac" Feb 17 14:08:04 crc kubenswrapper[4955]: I0217 14:08:04.222922 4955 scope.go:117] "RemoveContainer" containerID="fa86f3109bf152769bd9369dabc17d54290d99858877992d6ce0702550e664a2" Feb 17 14:08:04 crc kubenswrapper[4955]: E0217 14:08:04.224120 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 14:08:04 crc kubenswrapper[4955]: I0217 14:08:04.233296 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fbd8383c-94c7-49e9-80e9-a63c96476c69" path="/var/lib/kubelet/pods/fbd8383c-94c7-49e9-80e9-a63c96476c69/volumes" Feb 17 14:08:04 crc kubenswrapper[4955]: I0217 14:08:04.320116 4955 generic.go:334] "Generic (PLEG): container finished" podID="8d0a24cf-b58d-4493-9395-f3775cf25594" containerID="29e80b937c69fba2569169e33f138ec18a36f69a83e119c2ce4cb695531ed9d2" exitCode=0 Feb 17 14:08:04 crc kubenswrapper[4955]: I0217 14:08:04.320213 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sncmd/crc-debug-k527r" event={"ID":"8d0a24cf-b58d-4493-9395-f3775cf25594","Type":"ContainerDied","Data":"29e80b937c69fba2569169e33f138ec18a36f69a83e119c2ce4cb695531ed9d2"} Feb 17 14:08:04 crc kubenswrapper[4955]: I0217 14:08:04.320270 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sncmd/crc-debug-k527r" event={"ID":"8d0a24cf-b58d-4493-9395-f3775cf25594","Type":"ContainerStarted","Data":"1547f87fd47d1901ef5d304677bfd5d8c99bf4b57d1456315d31c9cde583bf41"} Feb 17 14:08:04 crc kubenswrapper[4955]: I0217 14:08:04.359312 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-sncmd/crc-debug-k527r"] Feb 17 14:08:04 crc kubenswrapper[4955]: I0217 14:08:04.367339 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-sncmd/crc-debug-k527r"] Feb 17 14:08:05 crc kubenswrapper[4955]: I0217 14:08:05.435128 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sncmd/crc-debug-k527r" Feb 17 14:08:05 crc kubenswrapper[4955]: I0217 14:08:05.569945 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8d0a24cf-b58d-4493-9395-f3775cf25594-host\") pod \"8d0a24cf-b58d-4493-9395-f3775cf25594\" (UID: \"8d0a24cf-b58d-4493-9395-f3775cf25594\") " Feb 17 14:08:05 crc kubenswrapper[4955]: I0217 14:08:05.570013 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dfg4m\" (UniqueName: \"kubernetes.io/projected/8d0a24cf-b58d-4493-9395-f3775cf25594-kube-api-access-dfg4m\") pod \"8d0a24cf-b58d-4493-9395-f3775cf25594\" (UID: \"8d0a24cf-b58d-4493-9395-f3775cf25594\") " Feb 17 14:08:05 crc kubenswrapper[4955]: I0217 14:08:05.570073 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8d0a24cf-b58d-4493-9395-f3775cf25594-host" (OuterVolumeSpecName: "host") pod "8d0a24cf-b58d-4493-9395-f3775cf25594" (UID: "8d0a24cf-b58d-4493-9395-f3775cf25594"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 17 14:08:05 crc kubenswrapper[4955]: I0217 14:08:05.570515 4955 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8d0a24cf-b58d-4493-9395-f3775cf25594-host\") on node \"crc\" DevicePath \"\"" Feb 17 14:08:05 crc kubenswrapper[4955]: I0217 14:08:05.580232 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d0a24cf-b58d-4493-9395-f3775cf25594-kube-api-access-dfg4m" (OuterVolumeSpecName: "kube-api-access-dfg4m") pod "8d0a24cf-b58d-4493-9395-f3775cf25594" (UID: "8d0a24cf-b58d-4493-9395-f3775cf25594"). InnerVolumeSpecName "kube-api-access-dfg4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 14:08:05 crc kubenswrapper[4955]: I0217 14:08:05.672414 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dfg4m\" (UniqueName: \"kubernetes.io/projected/8d0a24cf-b58d-4493-9395-f3775cf25594-kube-api-access-dfg4m\") on node \"crc\" DevicePath \"\"" Feb 17 14:08:06 crc kubenswrapper[4955]: I0217 14:08:06.237932 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d0a24cf-b58d-4493-9395-f3775cf25594" path="/var/lib/kubelet/pods/8d0a24cf-b58d-4493-9395-f3775cf25594/volumes" Feb 17 14:08:06 crc kubenswrapper[4955]: I0217 14:08:06.338290 4955 scope.go:117] "RemoveContainer" containerID="29e80b937c69fba2569169e33f138ec18a36f69a83e119c2ce4cb695531ed9d2" Feb 17 14:08:06 crc kubenswrapper[4955]: I0217 14:08:06.338374 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sncmd/crc-debug-k527r" Feb 17 14:08:16 crc kubenswrapper[4955]: I0217 14:08:16.223349 4955 scope.go:117] "RemoveContainer" containerID="fa86f3109bf152769bd9369dabc17d54290d99858877992d6ce0702550e664a2" Feb 17 14:08:16 crc kubenswrapper[4955]: E0217 14:08:16.224137 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 14:08:28 crc kubenswrapper[4955]: I0217 14:08:28.223679 4955 scope.go:117] "RemoveContainer" containerID="fa86f3109bf152769bd9369dabc17d54290d99858877992d6ce0702550e664a2" Feb 17 14:08:28 crc kubenswrapper[4955]: E0217 14:08:28.224809 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 14:08:33 crc kubenswrapper[4955]: I0217 14:08:33.637132 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-57f5fb6b66-7vp2j_87c5d914-d03a-41e9-9b1c-8c34a5e0758b/barbican-api/0.log" Feb 17 14:08:33 crc kubenswrapper[4955]: I0217 14:08:33.856973 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-57f5fb6b66-7vp2j_87c5d914-d03a-41e9-9b1c-8c34a5e0758b/barbican-api-log/0.log" Feb 17 14:08:33 crc kubenswrapper[4955]: I0217 14:08:33.878749 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7789497c74-pgff5_8468e621-9a50-49f7-ba03-996bc8564f43/barbican-keystone-listener-log/0.log" Feb 17 14:08:33 crc kubenswrapper[4955]: I0217 14:08:33.919357 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7789497c74-pgff5_8468e621-9a50-49f7-ba03-996bc8564f43/barbican-keystone-listener/0.log" Feb 17 14:08:34 crc kubenswrapper[4955]: I0217 14:08:34.050080 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6ff67b7c6c-9dcnd_20f0b6cf-c788-45bd-bdb2-cb2e01608a08/barbican-worker/0.log" Feb 17 14:08:34 crc kubenswrapper[4955]: I0217 14:08:34.054036 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6ff67b7c6c-9dcnd_20f0b6cf-c788-45bd-bdb2-cb2e01608a08/barbican-worker-log/0.log" Feb 17 14:08:34 crc kubenswrapper[4955]: I0217 14:08:34.275509 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-zn5l6_ea41b3c5-2c55-4294-b2dc-a25ab083c780/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Feb 17 14:08:34 crc kubenswrapper[4955]: I0217 14:08:34.340549 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c46f44da-c85d-4a6c-a5db-28e900c054da/ceilometer-central-agent/0.log" Feb 17 14:08:34 crc kubenswrapper[4955]: I0217 14:08:34.631190 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c46f44da-c85d-4a6c-a5db-28e900c054da/ceilometer-notification-agent/0.log" Feb 17 14:08:34 crc kubenswrapper[4955]: I0217 14:08:34.654198 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c46f44da-c85d-4a6c-a5db-28e900c054da/proxy-httpd/0.log" Feb 17 14:08:34 crc kubenswrapper[4955]: I0217 14:08:34.734668 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c46f44da-c85d-4a6c-a5db-28e900c054da/sg-core/0.log" Feb 17 14:08:34 crc kubenswrapper[4955]: I0217 14:08:34.828867 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_b1148d9f-bf24-449c-83b4-9a64d0572df0/cinder-api-log/0.log" Feb 17 14:08:34 crc kubenswrapper[4955]: I0217 14:08:34.884220 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_b1148d9f-bf24-449c-83b4-9a64d0572df0/cinder-api/0.log" Feb 17 14:08:35 crc kubenswrapper[4955]: I0217 14:08:35.014273 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7/cinder-scheduler/0.log" Feb 17 14:08:35 crc kubenswrapper[4955]: I0217 14:08:35.090030 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_2a5b18d1-9eae-4cb4-a432-9e3c5b68e4e7/probe/0.log" Feb 17 14:08:35 crc kubenswrapper[4955]: I0217 14:08:35.255694 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-ncz59_c74ecb0c-c07e-4196-a0a8-7c107404b60f/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 17 14:08:35 crc kubenswrapper[4955]: I0217 14:08:35.290858 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-sj68p_5e268a8e-d845-4915-b6ff-f10bc386cfc5/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 17 14:08:35 crc kubenswrapper[4955]: I0217 14:08:35.454007 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-hktkl_d64b1ace-43dd-409e-a0ca-f7573676d9b3/init/0.log" Feb 17 14:08:35 crc kubenswrapper[4955]: I0217 14:08:35.613900 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-hktkl_d64b1ace-43dd-409e-a0ca-f7573676d9b3/init/0.log" Feb 17 14:08:35 crc kubenswrapper[4955]: I0217 14:08:35.725441 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-hktkl_d64b1ace-43dd-409e-a0ca-f7573676d9b3/dnsmasq-dns/0.log" Feb 17 14:08:35 crc kubenswrapper[4955]: I0217 14:08:35.751827 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-bpbzk_682aa1ae-e0dc-476d-862d-45071ee0609a/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Feb 17 14:08:35 crc kubenswrapper[4955]: I0217 14:08:35.902035 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_72354e86-fe70-4224-b6ea-c7fed8a90f5d/glance-httpd/0.log" Feb 17 14:08:35 crc kubenswrapper[4955]: I0217 14:08:35.913892 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_72354e86-fe70-4224-b6ea-c7fed8a90f5d/glance-log/0.log" Feb 17 14:08:36 crc kubenswrapper[4955]: I0217 14:08:36.109058 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_539c0e46-aba7-4b5b-a176-3e9ac9089560/glance-log/0.log" Feb 17 14:08:36 crc kubenswrapper[4955]: I0217 14:08:36.137738 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_539c0e46-aba7-4b5b-a176-3e9ac9089560/glance-httpd/0.log" Feb 17 14:08:36 crc kubenswrapper[4955]: I0217 14:08:36.382497 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7fffbcf7cb-wr74b_d6cb09d2-ad7d-4486-a41d-89023789fd87/horizon/0.log" Feb 17 14:08:36 crc kubenswrapper[4955]: I0217 14:08:36.398950 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-4zrqx_f3e71e93-00f0-46f0-b87c-df0ddde2af3d/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Feb 17 14:08:36 crc kubenswrapper[4955]: I0217 14:08:36.725486 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-5h2ng_86978b4a-2cb7-4f99-bec3-0afeacdeb7b6/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 17 14:08:36 crc kubenswrapper[4955]: I0217 14:08:36.739111 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7fffbcf7cb-wr74b_d6cb09d2-ad7d-4486-a41d-89023789fd87/horizon-log/0.log" Feb 17 14:08:37 crc kubenswrapper[4955]: I0217 14:08:37.016863 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29522281-wgswt_f0ab4d69-eeed-461b-bc7d-1005eacc732f/keystone-cron/0.log" Feb 17 14:08:37 crc kubenswrapper[4955]: I0217 14:08:37.042260 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-c5b8c48f6-zr5fn_fc7f3f32-cb4f-4b66-8ba6-4e35b2555b75/keystone-api/0.log" Feb 17 14:08:37 crc kubenswrapper[4955]: I0217 14:08:37.146989 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_3e1af961-6b57-4a77-ae1c-dcfacea5c8c6/kube-state-metrics/0.log" Feb 17 14:08:37 crc kubenswrapper[4955]: I0217 14:08:37.264149 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-fq9zr_e3944196-3712-4506-bde5-758adf20aa63/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Feb 17 14:08:37 crc kubenswrapper[4955]: I0217 14:08:37.595347 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7c5ddc6895-552ts_25c07d2a-4f61-48fc-88e3-0ed671ad9cff/neutron-api/0.log" Feb 17 14:08:37 crc kubenswrapper[4955]: I0217 14:08:37.669026 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7c5ddc6895-552ts_25c07d2a-4f61-48fc-88e3-0ed671ad9cff/neutron-httpd/0.log" Feb 17 14:08:37 crc kubenswrapper[4955]: I0217 14:08:37.841091 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-8r6hf_5ff9b191-a542-41b9-a5c4-479d29e04fc3/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Feb 17 14:08:38 crc kubenswrapper[4955]: I0217 14:08:38.497557 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_a1a37e38-9f12-49ad-8464-f744f7f5fc2f/nova-api-log/0.log" Feb 17 14:08:38 crc kubenswrapper[4955]: I0217 14:08:38.530988 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_115dedcd-9ffe-408e-9b6b-878f53c6c252/nova-cell0-conductor-conductor/0.log" Feb 17 14:08:38 crc kubenswrapper[4955]: I0217 14:08:38.778649 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_2cc025ff-ae43-4877-940e-dab395505735/nova-cell1-conductor-conductor/0.log" Feb 17 14:08:38 crc kubenswrapper[4955]: I0217 14:08:38.888970 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_3614da91-13c7-4d09-b508-933322e35f08/nova-cell1-novncproxy-novncproxy/0.log" Feb 17 14:08:38 crc kubenswrapper[4955]: I0217 14:08:38.950645 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_a1a37e38-9f12-49ad-8464-f744f7f5fc2f/nova-api-api/0.log" Feb 17 14:08:39 crc kubenswrapper[4955]: I0217 14:08:39.063998 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-v478l_1ca453e4-29ab-4f17-b10b-f23c3a0f6a4f/nova-edpm-deployment-openstack-edpm-ipam/0.log" Feb 17 14:08:39 crc kubenswrapper[4955]: I0217 14:08:39.299551 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_ff800864-d970-46b0-bb2e-eb213d423da4/nova-metadata-log/0.log" Feb 17 14:08:39 crc kubenswrapper[4955]: I0217 14:08:39.628761 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1c93012c-0538-49a7-b414-6b01bb6c6dce/mysql-bootstrap/0.log" Feb 17 14:08:39 crc kubenswrapper[4955]: I0217 14:08:39.684978 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_ca665950-2b74-4ac5-8d48-eacdac1f65ae/nova-scheduler-scheduler/0.log" Feb 17 14:08:39 crc kubenswrapper[4955]: I0217 14:08:39.740061 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1c93012c-0538-49a7-b414-6b01bb6c6dce/mysql-bootstrap/0.log" Feb 17 14:08:39 crc kubenswrapper[4955]: I0217 14:08:39.856643 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1c93012c-0538-49a7-b414-6b01bb6c6dce/galera/0.log" Feb 17 14:08:40 crc kubenswrapper[4955]: I0217 14:08:40.008067 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_def1308a-11ee-4465-89f0-e78ffc4b8b51/mysql-bootstrap/0.log" Feb 17 14:08:40 crc kubenswrapper[4955]: I0217 14:08:40.222735 4955 scope.go:117] "RemoveContainer" containerID="fa86f3109bf152769bd9369dabc17d54290d99858877992d6ce0702550e664a2" Feb 17 14:08:40 crc kubenswrapper[4955]: E0217 14:08:40.223283 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 14:08:40 crc kubenswrapper[4955]: I0217 14:08:40.227026 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_def1308a-11ee-4465-89f0-e78ffc4b8b51/mysql-bootstrap/0.log" Feb 17 14:08:40 crc kubenswrapper[4955]: I0217 14:08:40.233421 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_def1308a-11ee-4465-89f0-e78ffc4b8b51/galera/0.log" Feb 17 14:08:40 crc kubenswrapper[4955]: I0217 14:08:40.489500 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_c2c452a7-9ff5-4f92-a5a9-87925a22302f/openstackclient/0.log" Feb 17 14:08:40 crc kubenswrapper[4955]: I0217 14:08:40.545102 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_ff800864-d970-46b0-bb2e-eb213d423da4/nova-metadata-metadata/0.log" Feb 17 14:08:40 crc kubenswrapper[4955]: I0217 14:08:40.560846 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-f8cjz_452bb98a-815b-4154-84c5-6b31338fca29/ovn-controller/0.log" Feb 17 14:08:40 crc kubenswrapper[4955]: I0217 14:08:40.839101 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-m27b4_9534cd01-b88e-48a4-a2ca-dd968e045dbc/ovsdb-server-init/0.log" Feb 17 14:08:40 crc kubenswrapper[4955]: I0217 14:08:40.842625 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-2s2tl_c2a9d7a6-de50-4a56-b18a-ed69f82ea656/openstack-network-exporter/0.log" Feb 17 14:08:41 crc kubenswrapper[4955]: I0217 14:08:41.070138 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-m27b4_9534cd01-b88e-48a4-a2ca-dd968e045dbc/ovsdb-server-init/0.log" Feb 17 14:08:41 crc kubenswrapper[4955]: I0217 14:08:41.105658 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-m27b4_9534cd01-b88e-48a4-a2ca-dd968e045dbc/ovs-vswitchd/0.log" Feb 17 14:08:41 crc kubenswrapper[4955]: I0217 14:08:41.142841 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-m27b4_9534cd01-b88e-48a4-a2ca-dd968e045dbc/ovsdb-server/0.log" Feb 17 14:08:41 crc kubenswrapper[4955]: I0217 14:08:41.291511 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-5zsq2_27e2834f-9629-4902-b1b9-e613e29164c0/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Feb 17 14:08:41 crc kubenswrapper[4955]: I0217 14:08:41.305111 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_94e88c0f-4478-4de1-9340-99904aac77e9/openstack-network-exporter/0.log" Feb 17 14:08:41 crc kubenswrapper[4955]: I0217 14:08:41.386134 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_94e88c0f-4478-4de1-9340-99904aac77e9/ovn-northd/0.log" Feb 17 14:08:41 crc kubenswrapper[4955]: I0217 14:08:41.509606 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_b3384848-891d-46ff-a869-61d657304b74/openstack-network-exporter/0.log" Feb 17 14:08:41 crc kubenswrapper[4955]: I0217 14:08:41.559953 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_b3384848-891d-46ff-a869-61d657304b74/ovsdbserver-nb/0.log" Feb 17 14:08:41 crc kubenswrapper[4955]: I0217 14:08:41.700911 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_aa892f09-0663-4fb6-b3ad-a15c300b5b65/openstack-network-exporter/0.log" Feb 17 14:08:41 crc kubenswrapper[4955]: I0217 14:08:41.755561 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_aa892f09-0663-4fb6-b3ad-a15c300b5b65/ovsdbserver-sb/0.log" Feb 17 14:08:41 crc kubenswrapper[4955]: I0217 14:08:41.914353 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5d56c994-tx7w7_54f84d50-4d2b-4d71-91e0-e6a7a00f568e/placement-api/0.log" Feb 17 14:08:41 crc kubenswrapper[4955]: I0217 14:08:41.990554 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5d56c994-tx7w7_54f84d50-4d2b-4d71-91e0-e6a7a00f568e/placement-log/0.log" Feb 17 14:08:42 crc kubenswrapper[4955]: I0217 14:08:42.049003 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_cd70e7eb-fb12-4b99-ab83-67d4abb9392b/setup-container/0.log" Feb 17 14:08:42 crc kubenswrapper[4955]: I0217 14:08:42.221315 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_cd70e7eb-fb12-4b99-ab83-67d4abb9392b/setup-container/0.log" Feb 17 14:08:42 crc kubenswrapper[4955]: I0217 14:08:42.297744 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_cd70e7eb-fb12-4b99-ab83-67d4abb9392b/rabbitmq/0.log" Feb 17 14:08:42 crc kubenswrapper[4955]: I0217 14:08:42.354236 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_c8c91914-fefa-479e-b497-de0f1f2978a2/setup-container/0.log" Feb 17 14:08:42 crc kubenswrapper[4955]: I0217 14:08:42.532470 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_c8c91914-fefa-479e-b497-de0f1f2978a2/setup-container/0.log" Feb 17 14:08:42 crc kubenswrapper[4955]: I0217 14:08:42.608419 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-mqnxx_5fe6a78c-e6a0-441f-b374-ca6603778eab/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 17 14:08:42 crc kubenswrapper[4955]: I0217 14:08:42.629202 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_c8c91914-fefa-479e-b497-de0f1f2978a2/rabbitmq/0.log" Feb 17 14:08:42 crc kubenswrapper[4955]: I0217 14:08:42.833026 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-cv8cz_8e826189-c723-4c9e-8ab8-3e60c06b7e42/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Feb 17 14:08:42 crc kubenswrapper[4955]: I0217 14:08:42.876393 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-g2nhl_e1bffb7a-a33a-466d-b382-a6896e83d8d9/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Feb 17 14:08:43 crc kubenswrapper[4955]: I0217 14:08:43.024886 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-qplrv_15be064a-4dca-49b2-aac9-6891c1afb237/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 17 14:08:43 crc kubenswrapper[4955]: I0217 14:08:43.202742 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-t5v8s_e6728d3d-3842-41be-a973-efb3d3fe67b9/ssh-known-hosts-edpm-deployment/0.log" Feb 17 14:08:43 crc kubenswrapper[4955]: I0217 14:08:43.420251 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5bcc76b6cf-4r2g9_0fa0efd9-940a-4580-ab68-c5954c0e732e/proxy-server/0.log" Feb 17 14:08:43 crc kubenswrapper[4955]: I0217 14:08:43.422586 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5bcc76b6cf-4r2g9_0fa0efd9-940a-4580-ab68-c5954c0e732e/proxy-httpd/0.log" Feb 17 14:08:43 crc kubenswrapper[4955]: I0217 14:08:43.597936 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-56p7b_46beefa1-21de-4818-a70f-e9f840a920ec/swift-ring-rebalance/0.log" Feb 17 14:08:43 crc kubenswrapper[4955]: I0217 14:08:43.674896 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fb6d9434-4476-438b-90e4-4781bd85b0bb/account-auditor/0.log" Feb 17 14:08:43 crc kubenswrapper[4955]: I0217 14:08:43.784386 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fb6d9434-4476-438b-90e4-4781bd85b0bb/account-reaper/0.log" Feb 17 14:08:43 crc kubenswrapper[4955]: I0217 14:08:43.836892 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fb6d9434-4476-438b-90e4-4781bd85b0bb/account-replicator/0.log" Feb 17 14:08:43 crc kubenswrapper[4955]: I0217 14:08:43.877502 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fb6d9434-4476-438b-90e4-4781bd85b0bb/account-server/0.log" Feb 17 14:08:43 crc kubenswrapper[4955]: I0217 14:08:43.966498 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fb6d9434-4476-438b-90e4-4781bd85b0bb/container-auditor/0.log" Feb 17 14:08:44 crc kubenswrapper[4955]: I0217 14:08:44.070137 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fb6d9434-4476-438b-90e4-4781bd85b0bb/container-server/0.log" Feb 17 14:08:44 crc kubenswrapper[4955]: I0217 14:08:44.085745 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fb6d9434-4476-438b-90e4-4781bd85b0bb/container-updater/0.log" Feb 17 14:08:44 crc kubenswrapper[4955]: I0217 14:08:44.094659 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fb6d9434-4476-438b-90e4-4781bd85b0bb/container-replicator/0.log" Feb 17 14:08:44 crc kubenswrapper[4955]: I0217 14:08:44.223154 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fb6d9434-4476-438b-90e4-4781bd85b0bb/object-auditor/0.log" Feb 17 14:08:44 crc kubenswrapper[4955]: I0217 14:08:44.260183 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fb6d9434-4476-438b-90e4-4781bd85b0bb/object-expirer/0.log" Feb 17 14:08:44 crc kubenswrapper[4955]: I0217 14:08:44.310612 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fb6d9434-4476-438b-90e4-4781bd85b0bb/object-replicator/0.log" Feb 17 14:08:44 crc kubenswrapper[4955]: I0217 14:08:44.321457 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fb6d9434-4476-438b-90e4-4781bd85b0bb/object-server/0.log" Feb 17 14:08:44 crc kubenswrapper[4955]: I0217 14:08:44.426892 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fb6d9434-4476-438b-90e4-4781bd85b0bb/object-updater/0.log" Feb 17 14:08:44 crc kubenswrapper[4955]: I0217 14:08:44.452516 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fb6d9434-4476-438b-90e4-4781bd85b0bb/rsync/0.log" Feb 17 14:08:44 crc kubenswrapper[4955]: I0217 14:08:44.487701 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fb6d9434-4476-438b-90e4-4781bd85b0bb/swift-recon-cron/0.log" Feb 17 14:08:44 crc kubenswrapper[4955]: I0217 14:08:44.784903 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_de1d0057-d623-4185-b31a-572dfaabcce2/tempest-tests-tempest-tests-runner/0.log" Feb 17 14:08:44 crc kubenswrapper[4955]: I0217 14:08:44.923171 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-npz67_0acb7fb8-671d-4e90-8be4-f5f507308f11/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Feb 17 14:08:44 crc kubenswrapper[4955]: I0217 14:08:44.996504 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_35a9a907-8d3b-49dd-9dd3-0f2e1a5aa7bc/test-operator-logs-container/0.log" Feb 17 14:08:45 crc kubenswrapper[4955]: I0217 14:08:45.142399 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-jr9rx_b743858f-111b-4d79-aeaf-6b9608fecbac/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 17 14:08:51 crc kubenswrapper[4955]: I0217 14:08:51.224060 4955 scope.go:117] "RemoveContainer" containerID="fa86f3109bf152769bd9369dabc17d54290d99858877992d6ce0702550e664a2" Feb 17 14:08:51 crc kubenswrapper[4955]: E0217 14:08:51.224610 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 14:08:55 crc kubenswrapper[4955]: I0217 14:08:55.296258 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_7dab7b98-7d1a-4940-a4c3-8bbbd57fe121/memcached/0.log" Feb 17 14:09:02 crc kubenswrapper[4955]: I0217 14:09:02.234627 4955 scope.go:117] "RemoveContainer" containerID="fa86f3109bf152769bd9369dabc17d54290d99858877992d6ce0702550e664a2" Feb 17 14:09:02 crc kubenswrapper[4955]: E0217 14:09:02.235426 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 14:09:09 crc kubenswrapper[4955]: I0217 14:09:09.563175 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p_f328d9c1-907b-4705-8580-0b5da54053c6/util/0.log" Feb 17 14:09:09 crc kubenswrapper[4955]: I0217 14:09:09.802376 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p_f328d9c1-907b-4705-8580-0b5da54053c6/pull/0.log" Feb 17 14:09:09 crc kubenswrapper[4955]: I0217 14:09:09.848966 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p_f328d9c1-907b-4705-8580-0b5da54053c6/util/0.log" Feb 17 14:09:10 crc kubenswrapper[4955]: I0217 14:09:10.000631 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p_f328d9c1-907b-4705-8580-0b5da54053c6/pull/0.log" Feb 17 14:09:10 crc kubenswrapper[4955]: I0217 14:09:10.228724 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p_f328d9c1-907b-4705-8580-0b5da54053c6/pull/0.log" Feb 17 14:09:10 crc kubenswrapper[4955]: I0217 14:09:10.263141 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p_f328d9c1-907b-4705-8580-0b5da54053c6/util/0.log" Feb 17 14:09:10 crc kubenswrapper[4955]: I0217 14:09:10.399822 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d568eacd9f626ba30177714fde257cb60e0ca7b4cf776f079ee7fac03ets59p_f328d9c1-907b-4705-8580-0b5da54053c6/extract/0.log" Feb 17 14:09:10 crc kubenswrapper[4955]: I0217 14:09:10.660883 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-6d8bf5c495-2c8ww_1e4fd408-9f09-4289-92ce-7c5ae18f235a/manager/0.log" Feb 17 14:09:10 crc kubenswrapper[4955]: I0217 14:09:10.722356 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-5d946d989d-lpxnk_6d1e254f-7d68-45c5-bda0-e3b33155ae56/manager/0.log" Feb 17 14:09:10 crc kubenswrapper[4955]: I0217 14:09:10.907484 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-69f49c598c-75hpg_071c3b13-9304-48b6-bf22-d89dd62825a0/manager/0.log" Feb 17 14:09:11 crc kubenswrapper[4955]: I0217 14:09:11.034069 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987464f4-g967h_5c339089-ba43-481b-aaa8-f8bf09fd106d/manager/0.log" Feb 17 14:09:11 crc kubenswrapper[4955]: I0217 14:09:11.207589 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5b9b8895d5-k2x44_6ade752d-a3d0-4449-8028-67405763ccc8/manager/0.log" Feb 17 14:09:11 crc kubenswrapper[4955]: I0217 14:09:11.476121 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-554564d7fc-h2rm6_9019acce-c356-421a-bd47-a163a8857f30/manager/0.log" Feb 17 14:09:11 crc kubenswrapper[4955]: I0217 14:09:11.660888 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-79d975b745-gpgxj_cc84cdf7-92aa-4134-a784-d821c076cde3/manager/0.log" Feb 17 14:09:11 crc kubenswrapper[4955]: I0217 14:09:11.773815 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-b4d948c87-mkqrg_d6decb51-f3f9-4183-a20a-6aa4491dba61/manager/0.log" Feb 17 14:09:11 crc kubenswrapper[4955]: I0217 14:09:11.870935 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-54f6768c69-4f8kb_01eb7df7-9e0a-4e78-85a3-0d456dec139f/manager/0.log" Feb 17 14:09:12 crc kubenswrapper[4955]: I0217 14:09:12.080771 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6994f66f48-nzn4k_308b7713-e37c-4890-8b56-11379bad3f8d/manager/0.log" Feb 17 14:09:12 crc kubenswrapper[4955]: I0217 14:09:12.409545 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-64ddbf8bb-xkvlp_037562f9-c04f-4e9d-8fa5-3e98ac7c8fc5/manager/0.log" Feb 17 14:09:12 crc kubenswrapper[4955]: I0217 14:09:12.495375 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-567668f5cf-5fxtz_744f25f9-1641-4864-8a55-ea14ead19cae/manager/0.log" Feb 17 14:09:12 crc kubenswrapper[4955]: I0217 14:09:12.749125 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-7c6767dc9cxkmq8_cd2ef1c9-23d1-4671-bc71-2c077afdfe08/manager/0.log" Feb 17 14:09:13 crc kubenswrapper[4955]: I0217 14:09:13.181376 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-57f9c85f89-8zxwf_5b01c583-8db4-4939-ac7b-e0d16b9f72bf/operator/0.log" Feb 17 14:09:13 crc kubenswrapper[4955]: I0217 14:09:13.521687 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-5qzcl_38d53049-7d41-4938-b2d6-c7c7568ba144/registry-server/0.log" Feb 17 14:09:13 crc kubenswrapper[4955]: I0217 14:09:13.842700 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-d44cf6b75-vq5wq_17cecde4-2eb0-445e-bcf4-f73f719627df/manager/0.log" Feb 17 14:09:14 crc kubenswrapper[4955]: I0217 14:09:14.076762 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-8497b45c89-fqrz9_788f9fac-dd10-49c4-8a25-4578d69aeb2b/manager/0.log" Feb 17 14:09:14 crc kubenswrapper[4955]: I0217 14:09:14.080720 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69f8888797-wkdhh_3f752291-5ba5-4074-a126-e99f6f0242aa/manager/0.log" Feb 17 14:09:14 crc kubenswrapper[4955]: I0217 14:09:14.311697 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-nmxwn_afaac24a-0894-4001-9b50-d3d7c2de90bc/operator/0.log" Feb 17 14:09:14 crc kubenswrapper[4955]: I0217 14:09:14.395499 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-68f46476f-5qz9n_60b19c88-d91f-422e-9012-ad97daf661bb/manager/0.log" Feb 17 14:09:14 crc kubenswrapper[4955]: I0217 14:09:14.606557 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7f45b4ff68-q68lr_72fe0652-eee2-4e49-83b9-3c1790e8ca9e/manager/0.log" Feb 17 14:09:14 crc kubenswrapper[4955]: I0217 14:09:14.761935 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-7866795846-xqxjt_90e44f80-9802-4134-8aa3-519bc93f4bd3/manager/0.log" Feb 17 14:09:14 crc kubenswrapper[4955]: I0217 14:09:14.894531 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-5db88f68c-v6jkn_a1f4d718-9535-4290-94eb-93d6288d6047/manager/0.log" Feb 17 14:09:15 crc kubenswrapper[4955]: I0217 14:09:15.121991 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5699594bc4-nrc2h_c93c2a50-6396-4a32-9b88-dcdc341ec6c5/manager/0.log" Feb 17 14:09:16 crc kubenswrapper[4955]: I0217 14:09:16.228202 4955 scope.go:117] "RemoveContainer" containerID="fa86f3109bf152769bd9369dabc17d54290d99858877992d6ce0702550e664a2" Feb 17 14:09:16 crc kubenswrapper[4955]: E0217 14:09:16.228689 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 14:09:16 crc kubenswrapper[4955]: I0217 14:09:16.352894 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-868647ff47-ffkxx_602419fb-5033-4889-b587-e1eb97001910/manager/0.log" Feb 17 14:09:28 crc kubenswrapper[4955]: I0217 14:09:28.223467 4955 scope.go:117] "RemoveContainer" containerID="fa86f3109bf152769bd9369dabc17d54290d99858877992d6ce0702550e664a2" Feb 17 14:09:28 crc kubenswrapper[4955]: E0217 14:09:28.224089 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 14:09:34 crc kubenswrapper[4955]: I0217 14:09:34.938966 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-vtd5z_8f5feb72-96ca-4e4d-9566-0757c7dd93f9/control-plane-machine-set-operator/0.log" Feb 17 14:09:35 crc kubenswrapper[4955]: I0217 14:09:35.139713 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-d5sv6_847ef69d-8cd1-4831-8250-05f2cf789d88/kube-rbac-proxy/0.log" Feb 17 14:09:35 crc kubenswrapper[4955]: I0217 14:09:35.201230 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-d5sv6_847ef69d-8cd1-4831-8250-05f2cf789d88/machine-api-operator/0.log" Feb 17 14:09:43 crc kubenswrapper[4955]: I0217 14:09:43.223178 4955 scope.go:117] "RemoveContainer" containerID="fa86f3109bf152769bd9369dabc17d54290d99858877992d6ce0702550e664a2" Feb 17 14:09:44 crc kubenswrapper[4955]: I0217 14:09:44.214442 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" event={"ID":"5bf36887-2a2b-4193-abdb-a15beaf46300","Type":"ContainerStarted","Data":"91eab0ac0f5514297337a55f7c52c0dd490bc24e8afd932deb11adb7e1d444e2"} Feb 17 14:09:47 crc kubenswrapper[4955]: I0217 14:09:47.228977 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-lg6zf_54536ade-766d-4cf8-b1b5-057ccb961b13/cert-manager-controller/0.log" Feb 17 14:09:47 crc kubenswrapper[4955]: I0217 14:09:47.396920 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-lmhj4_de5cd487-58d9-4c86-a4e2-f6b000a351f5/cert-manager-cainjector/0.log" Feb 17 14:09:47 crc kubenswrapper[4955]: I0217 14:09:47.430964 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-69dts_d4a28595-c677-4966-8d3a-1ad5d299fd51/cert-manager-webhook/0.log" Feb 17 14:09:59 crc kubenswrapper[4955]: I0217 14:09:59.944021 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5c78fc5d65-bdkvr_0afd3165-a15b-406f-88fb-4ce82ab17355/nmstate-console-plugin/0.log" Feb 17 14:10:00 crc kubenswrapper[4955]: I0217 14:10:00.066519 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-9vdw4_bb2572e6-ed65-4a6c-a3ca-f72c6cb4b5c0/nmstate-handler/0.log" Feb 17 14:10:00 crc kubenswrapper[4955]: I0217 14:10:00.244611 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58c85c668d-b4ngx_bcf404eb-4186-40c9-ba70-768b78c8a60d/kube-rbac-proxy/0.log" Feb 17 14:10:00 crc kubenswrapper[4955]: I0217 14:10:00.272691 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58c85c668d-b4ngx_bcf404eb-4186-40c9-ba70-768b78c8a60d/nmstate-metrics/0.log" Feb 17 14:10:00 crc kubenswrapper[4955]: I0217 14:10:00.310874 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-694c9596b7-j9h2x_4bd50965-e917-4e34-80b2-4e54e1d67feb/nmstate-operator/0.log" Feb 17 14:10:00 crc kubenswrapper[4955]: I0217 14:10:00.491768 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-866bcb46dc-d7hqh_3aca323f-af5c-452c-8a4e-8c2b55e39dd5/nmstate-webhook/0.log" Feb 17 14:10:28 crc kubenswrapper[4955]: I0217 14:10:28.052497 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-69bbfbf88f-pvzn9_8c11830d-6699-432a-a2d3-14b135bc1665/kube-rbac-proxy/0.log" Feb 17 14:10:28 crc kubenswrapper[4955]: I0217 14:10:28.074010 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-69bbfbf88f-pvzn9_8c11830d-6699-432a-a2d3-14b135bc1665/controller/0.log" Feb 17 14:10:28 crc kubenswrapper[4955]: I0217 14:10:28.262980 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vzhr7_720ac59e-55ea-44be-b7af-9b57fec4901b/cp-frr-files/0.log" Feb 17 14:10:28 crc kubenswrapper[4955]: I0217 14:10:28.493437 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vzhr7_720ac59e-55ea-44be-b7af-9b57fec4901b/cp-reloader/0.log" Feb 17 14:10:28 crc kubenswrapper[4955]: I0217 14:10:28.494174 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vzhr7_720ac59e-55ea-44be-b7af-9b57fec4901b/cp-frr-files/0.log" Feb 17 14:10:28 crc kubenswrapper[4955]: I0217 14:10:28.495288 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vzhr7_720ac59e-55ea-44be-b7af-9b57fec4901b/cp-metrics/0.log" Feb 17 14:10:28 crc kubenswrapper[4955]: I0217 14:10:28.525646 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vzhr7_720ac59e-55ea-44be-b7af-9b57fec4901b/cp-reloader/0.log" Feb 17 14:10:28 crc kubenswrapper[4955]: I0217 14:10:28.720815 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vzhr7_720ac59e-55ea-44be-b7af-9b57fec4901b/cp-frr-files/0.log" Feb 17 14:10:28 crc kubenswrapper[4955]: I0217 14:10:28.720969 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vzhr7_720ac59e-55ea-44be-b7af-9b57fec4901b/cp-metrics/0.log" Feb 17 14:10:28 crc kubenswrapper[4955]: I0217 14:10:28.721155 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vzhr7_720ac59e-55ea-44be-b7af-9b57fec4901b/cp-reloader/0.log" Feb 17 14:10:28 crc kubenswrapper[4955]: I0217 14:10:28.725105 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vzhr7_720ac59e-55ea-44be-b7af-9b57fec4901b/cp-metrics/0.log" Feb 17 14:10:28 crc kubenswrapper[4955]: I0217 14:10:28.977883 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vzhr7_720ac59e-55ea-44be-b7af-9b57fec4901b/controller/0.log" Feb 17 14:10:28 crc kubenswrapper[4955]: I0217 14:10:28.981692 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vzhr7_720ac59e-55ea-44be-b7af-9b57fec4901b/cp-frr-files/0.log" Feb 17 14:10:28 crc kubenswrapper[4955]: I0217 14:10:28.982871 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vzhr7_720ac59e-55ea-44be-b7af-9b57fec4901b/cp-metrics/0.log" Feb 17 14:10:28 crc kubenswrapper[4955]: I0217 14:10:28.998237 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vzhr7_720ac59e-55ea-44be-b7af-9b57fec4901b/cp-reloader/0.log" Feb 17 14:10:29 crc kubenswrapper[4955]: I0217 14:10:29.193228 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vzhr7_720ac59e-55ea-44be-b7af-9b57fec4901b/frr-metrics/0.log" Feb 17 14:10:29 crc kubenswrapper[4955]: I0217 14:10:29.231013 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vzhr7_720ac59e-55ea-44be-b7af-9b57fec4901b/kube-rbac-proxy/0.log" Feb 17 14:10:29 crc kubenswrapper[4955]: I0217 14:10:29.258925 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vzhr7_720ac59e-55ea-44be-b7af-9b57fec4901b/kube-rbac-proxy-frr/0.log" Feb 17 14:10:29 crc kubenswrapper[4955]: I0217 14:10:29.483893 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-78b44bf5bb-rcv4v_92393c39-fc3d-490c-93a3-c11ddef89c02/frr-k8s-webhook-server/0.log" Feb 17 14:10:29 crc kubenswrapper[4955]: I0217 14:10:29.542105 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vzhr7_720ac59e-55ea-44be-b7af-9b57fec4901b/reloader/0.log" Feb 17 14:10:29 crc kubenswrapper[4955]: I0217 14:10:29.736867 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-754bdc4867-j8r95_6f6cf1f8-d54b-4fad-8313-c5f235a9df88/manager/0.log" Feb 17 14:10:29 crc kubenswrapper[4955]: I0217 14:10:29.885325 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-579d9646f7-zpz6n_4b384d59-9dc8-4d70-8b12-bd97252d1cfa/webhook-server/0.log" Feb 17 14:10:30 crc kubenswrapper[4955]: I0217 14:10:30.029109 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-dl5vp_618eb288-1aa4-43a9-b26a-9e2f5378411d/kube-rbac-proxy/0.log" Feb 17 14:10:30 crc kubenswrapper[4955]: I0217 14:10:30.594717 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-dl5vp_618eb288-1aa4-43a9-b26a-9e2f5378411d/speaker/0.log" Feb 17 14:10:30 crc kubenswrapper[4955]: I0217 14:10:30.737580 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vzhr7_720ac59e-55ea-44be-b7af-9b57fec4901b/frr/0.log" Feb 17 14:10:45 crc kubenswrapper[4955]: I0217 14:10:45.254186 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l_102e3d48-f56c-4e7c-956e-23484bf45776/util/0.log" Feb 17 14:10:45 crc kubenswrapper[4955]: I0217 14:10:45.504126 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l_102e3d48-f56c-4e7c-956e-23484bf45776/pull/0.log" Feb 17 14:10:45 crc kubenswrapper[4955]: I0217 14:10:45.529551 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l_102e3d48-f56c-4e7c-956e-23484bf45776/pull/0.log" Feb 17 14:10:45 crc kubenswrapper[4955]: I0217 14:10:45.551408 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l_102e3d48-f56c-4e7c-956e-23484bf45776/util/0.log" Feb 17 14:10:45 crc kubenswrapper[4955]: I0217 14:10:45.662739 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l_102e3d48-f56c-4e7c-956e-23484bf45776/util/0.log" Feb 17 14:10:45 crc kubenswrapper[4955]: I0217 14:10:45.701720 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l_102e3d48-f56c-4e7c-956e-23484bf45776/extract/0.log" Feb 17 14:10:45 crc kubenswrapper[4955]: I0217 14:10:45.729394 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2139hm9l_102e3d48-f56c-4e7c-956e-23484bf45776/pull/0.log" Feb 17 14:10:45 crc kubenswrapper[4955]: I0217 14:10:45.852914 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-krpv5_47a043d1-03b2-4a46-8c24-f1d1fef7efa6/extract-utilities/0.log" Feb 17 14:10:46 crc kubenswrapper[4955]: I0217 14:10:46.020553 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-krpv5_47a043d1-03b2-4a46-8c24-f1d1fef7efa6/extract-content/0.log" Feb 17 14:10:46 crc kubenswrapper[4955]: I0217 14:10:46.050269 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-krpv5_47a043d1-03b2-4a46-8c24-f1d1fef7efa6/extract-content/0.log" Feb 17 14:10:46 crc kubenswrapper[4955]: I0217 14:10:46.065572 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-krpv5_47a043d1-03b2-4a46-8c24-f1d1fef7efa6/extract-utilities/0.log" Feb 17 14:10:46 crc kubenswrapper[4955]: I0217 14:10:46.236399 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-krpv5_47a043d1-03b2-4a46-8c24-f1d1fef7efa6/extract-content/0.log" Feb 17 14:10:46 crc kubenswrapper[4955]: I0217 14:10:46.269388 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-krpv5_47a043d1-03b2-4a46-8c24-f1d1fef7efa6/extract-utilities/0.log" Feb 17 14:10:46 crc kubenswrapper[4955]: I0217 14:10:46.475961 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2jxwm_59516804-ad0e-428f-bf9a-1b5561f7ec1a/extract-utilities/0.log" Feb 17 14:10:46 crc kubenswrapper[4955]: I0217 14:10:46.689416 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2jxwm_59516804-ad0e-428f-bf9a-1b5561f7ec1a/extract-utilities/0.log" Feb 17 14:10:46 crc kubenswrapper[4955]: I0217 14:10:46.750118 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2jxwm_59516804-ad0e-428f-bf9a-1b5561f7ec1a/extract-content/0.log" Feb 17 14:10:46 crc kubenswrapper[4955]: I0217 14:10:46.764241 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-krpv5_47a043d1-03b2-4a46-8c24-f1d1fef7efa6/registry-server/0.log" Feb 17 14:10:46 crc kubenswrapper[4955]: I0217 14:10:46.769079 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2jxwm_59516804-ad0e-428f-bf9a-1b5561f7ec1a/extract-content/0.log" Feb 17 14:10:46 crc kubenswrapper[4955]: I0217 14:10:46.962564 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2jxwm_59516804-ad0e-428f-bf9a-1b5561f7ec1a/extract-utilities/0.log" Feb 17 14:10:46 crc kubenswrapper[4955]: I0217 14:10:46.989668 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2jxwm_59516804-ad0e-428f-bf9a-1b5561f7ec1a/extract-content/0.log" Feb 17 14:10:47 crc kubenswrapper[4955]: I0217 14:10:47.151241 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j_c5bff73f-3f9e-4313-93ac-6276ceaf16d7/util/0.log" Feb 17 14:10:47 crc kubenswrapper[4955]: I0217 14:10:47.381973 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j_c5bff73f-3f9e-4313-93ac-6276ceaf16d7/pull/0.log" Feb 17 14:10:47 crc kubenswrapper[4955]: I0217 14:10:47.423573 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j_c5bff73f-3f9e-4313-93ac-6276ceaf16d7/pull/0.log" Feb 17 14:10:47 crc kubenswrapper[4955]: I0217 14:10:47.496210 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j_c5bff73f-3f9e-4313-93ac-6276ceaf16d7/util/0.log" Feb 17 14:10:47 crc kubenswrapper[4955]: I0217 14:10:47.567001 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2jxwm_59516804-ad0e-428f-bf9a-1b5561f7ec1a/registry-server/0.log" Feb 17 14:10:47 crc kubenswrapper[4955]: I0217 14:10:47.787019 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j_c5bff73f-3f9e-4313-93ac-6276ceaf16d7/pull/0.log" Feb 17 14:10:47 crc kubenswrapper[4955]: I0217 14:10:47.812334 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j_c5bff73f-3f9e-4313-93ac-6276ceaf16d7/util/0.log" Feb 17 14:10:47 crc kubenswrapper[4955]: I0217 14:10:47.864557 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca67w4j_c5bff73f-3f9e-4313-93ac-6276ceaf16d7/extract/0.log" Feb 17 14:10:47 crc kubenswrapper[4955]: I0217 14:10:47.991829 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-spj8f_f7954d5c-c985-4421-96e9-24ae232c8ba1/marketplace-operator/0.log" Feb 17 14:10:48 crc kubenswrapper[4955]: I0217 14:10:48.113239 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jmrnj_1cf150d2-8598-45d3-b193-4d27e90ca648/extract-utilities/0.log" Feb 17 14:10:48 crc kubenswrapper[4955]: I0217 14:10:48.259352 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jmrnj_1cf150d2-8598-45d3-b193-4d27e90ca648/extract-utilities/0.log" Feb 17 14:10:48 crc kubenswrapper[4955]: I0217 14:10:48.315625 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jmrnj_1cf150d2-8598-45d3-b193-4d27e90ca648/extract-content/0.log" Feb 17 14:10:48 crc kubenswrapper[4955]: I0217 14:10:48.315659 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jmrnj_1cf150d2-8598-45d3-b193-4d27e90ca648/extract-content/0.log" Feb 17 14:10:48 crc kubenswrapper[4955]: I0217 14:10:48.481677 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jmrnj_1cf150d2-8598-45d3-b193-4d27e90ca648/extract-content/0.log" Feb 17 14:10:48 crc kubenswrapper[4955]: I0217 14:10:48.500505 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jmrnj_1cf150d2-8598-45d3-b193-4d27e90ca648/extract-utilities/0.log" Feb 17 14:10:48 crc kubenswrapper[4955]: I0217 14:10:48.690339 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qz5rg_a5c8c29f-7770-4ba6-aad9-12604d10981f/extract-utilities/0.log" Feb 17 14:10:50 crc kubenswrapper[4955]: I0217 14:10:48.904767 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qz5rg_a5c8c29f-7770-4ba6-aad9-12604d10981f/extract-utilities/0.log" Feb 17 14:10:50 crc kubenswrapper[4955]: I0217 14:10:48.922014 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qz5rg_a5c8c29f-7770-4ba6-aad9-12604d10981f/extract-content/0.log" Feb 17 14:10:50 crc kubenswrapper[4955]: I0217 14:10:48.927844 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qz5rg_a5c8c29f-7770-4ba6-aad9-12604d10981f/extract-content/0.log" Feb 17 14:10:50 crc kubenswrapper[4955]: I0217 14:10:49.150536 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qz5rg_a5c8c29f-7770-4ba6-aad9-12604d10981f/extract-utilities/0.log" Feb 17 14:10:50 crc kubenswrapper[4955]: I0217 14:10:49.220024 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qz5rg_a5c8c29f-7770-4ba6-aad9-12604d10981f/extract-content/0.log" Feb 17 14:10:50 crc kubenswrapper[4955]: I0217 14:10:49.224512 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jmrnj_1cf150d2-8598-45d3-b193-4d27e90ca648/registry-server/0.log" Feb 17 14:10:50 crc kubenswrapper[4955]: I0217 14:10:49.842981 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qz5rg_a5c8c29f-7770-4ba6-aad9-12604d10981f/registry-server/0.log" Feb 17 14:11:25 crc kubenswrapper[4955]: E0217 14:11:25.892600 4955 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.46:40124->38.102.83.46:34749: write tcp 38.102.83.46:40124->38.102.83.46:34749: write: broken pipe Feb 17 14:12:04 crc kubenswrapper[4955]: I0217 14:12:04.575317 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 14:12:04 crc kubenswrapper[4955]: I0217 14:12:04.576032 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 14:12:10 crc kubenswrapper[4955]: I0217 14:12:10.404392 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zmnm5"] Feb 17 14:12:10 crc kubenswrapper[4955]: E0217 14:12:10.405544 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d0a24cf-b58d-4493-9395-f3775cf25594" containerName="container-00" Feb 17 14:12:10 crc kubenswrapper[4955]: I0217 14:12:10.405560 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d0a24cf-b58d-4493-9395-f3775cf25594" containerName="container-00" Feb 17 14:12:10 crc kubenswrapper[4955]: I0217 14:12:10.405853 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d0a24cf-b58d-4493-9395-f3775cf25594" containerName="container-00" Feb 17 14:12:10 crc kubenswrapper[4955]: I0217 14:12:10.408491 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zmnm5" Feb 17 14:12:10 crc kubenswrapper[4955]: I0217 14:12:10.428562 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zmnm5"] Feb 17 14:12:10 crc kubenswrapper[4955]: I0217 14:12:10.565121 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3-catalog-content\") pod \"community-operators-zmnm5\" (UID: \"03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3\") " pod="openshift-marketplace/community-operators-zmnm5" Feb 17 14:12:10 crc kubenswrapper[4955]: I0217 14:12:10.565181 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3-utilities\") pod \"community-operators-zmnm5\" (UID: \"03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3\") " pod="openshift-marketplace/community-operators-zmnm5" Feb 17 14:12:10 crc kubenswrapper[4955]: I0217 14:12:10.565316 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbjwl\" (UniqueName: \"kubernetes.io/projected/03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3-kube-api-access-jbjwl\") pod \"community-operators-zmnm5\" (UID: \"03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3\") " pod="openshift-marketplace/community-operators-zmnm5" Feb 17 14:12:10 crc kubenswrapper[4955]: I0217 14:12:10.600225 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5xj6c"] Feb 17 14:12:10 crc kubenswrapper[4955]: I0217 14:12:10.602173 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5xj6c" Feb 17 14:12:10 crc kubenswrapper[4955]: I0217 14:12:10.619631 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5xj6c"] Feb 17 14:12:10 crc kubenswrapper[4955]: I0217 14:12:10.667509 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dl69p\" (UniqueName: \"kubernetes.io/projected/becde1e2-5d09-45d1-b498-32172a394218-kube-api-access-dl69p\") pod \"redhat-marketplace-5xj6c\" (UID: \"becde1e2-5d09-45d1-b498-32172a394218\") " pod="openshift-marketplace/redhat-marketplace-5xj6c" Feb 17 14:12:10 crc kubenswrapper[4955]: I0217 14:12:10.667599 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbjwl\" (UniqueName: \"kubernetes.io/projected/03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3-kube-api-access-jbjwl\") pod \"community-operators-zmnm5\" (UID: \"03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3\") " pod="openshift-marketplace/community-operators-zmnm5" Feb 17 14:12:10 crc kubenswrapper[4955]: I0217 14:12:10.667698 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3-catalog-content\") pod \"community-operators-zmnm5\" (UID: \"03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3\") " pod="openshift-marketplace/community-operators-zmnm5" Feb 17 14:12:10 crc kubenswrapper[4955]: I0217 14:12:10.667722 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/becde1e2-5d09-45d1-b498-32172a394218-utilities\") pod \"redhat-marketplace-5xj6c\" (UID: \"becde1e2-5d09-45d1-b498-32172a394218\") " pod="openshift-marketplace/redhat-marketplace-5xj6c" Feb 17 14:12:10 crc kubenswrapper[4955]: I0217 14:12:10.667743 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3-utilities\") pod \"community-operators-zmnm5\" (UID: \"03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3\") " pod="openshift-marketplace/community-operators-zmnm5" Feb 17 14:12:10 crc kubenswrapper[4955]: I0217 14:12:10.667768 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/becde1e2-5d09-45d1-b498-32172a394218-catalog-content\") pod \"redhat-marketplace-5xj6c\" (UID: \"becde1e2-5d09-45d1-b498-32172a394218\") " pod="openshift-marketplace/redhat-marketplace-5xj6c" Feb 17 14:12:10 crc kubenswrapper[4955]: I0217 14:12:10.668684 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3-catalog-content\") pod \"community-operators-zmnm5\" (UID: \"03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3\") " pod="openshift-marketplace/community-operators-zmnm5" Feb 17 14:12:10 crc kubenswrapper[4955]: I0217 14:12:10.668930 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3-utilities\") pod \"community-operators-zmnm5\" (UID: \"03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3\") " pod="openshift-marketplace/community-operators-zmnm5" Feb 17 14:12:10 crc kubenswrapper[4955]: I0217 14:12:10.688752 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbjwl\" (UniqueName: \"kubernetes.io/projected/03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3-kube-api-access-jbjwl\") pod \"community-operators-zmnm5\" (UID: \"03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3\") " pod="openshift-marketplace/community-operators-zmnm5" Feb 17 14:12:10 crc kubenswrapper[4955]: I0217 14:12:10.736703 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zmnm5" Feb 17 14:12:10 crc kubenswrapper[4955]: I0217 14:12:10.771186 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/becde1e2-5d09-45d1-b498-32172a394218-utilities\") pod \"redhat-marketplace-5xj6c\" (UID: \"becde1e2-5d09-45d1-b498-32172a394218\") " pod="openshift-marketplace/redhat-marketplace-5xj6c" Feb 17 14:12:10 crc kubenswrapper[4955]: I0217 14:12:10.771270 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/becde1e2-5d09-45d1-b498-32172a394218-catalog-content\") pod \"redhat-marketplace-5xj6c\" (UID: \"becde1e2-5d09-45d1-b498-32172a394218\") " pod="openshift-marketplace/redhat-marketplace-5xj6c" Feb 17 14:12:10 crc kubenswrapper[4955]: I0217 14:12:10.771330 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dl69p\" (UniqueName: \"kubernetes.io/projected/becde1e2-5d09-45d1-b498-32172a394218-kube-api-access-dl69p\") pod \"redhat-marketplace-5xj6c\" (UID: \"becde1e2-5d09-45d1-b498-32172a394218\") " pod="openshift-marketplace/redhat-marketplace-5xj6c" Feb 17 14:12:10 crc kubenswrapper[4955]: I0217 14:12:10.771976 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/becde1e2-5d09-45d1-b498-32172a394218-utilities\") pod \"redhat-marketplace-5xj6c\" (UID: \"becde1e2-5d09-45d1-b498-32172a394218\") " pod="openshift-marketplace/redhat-marketplace-5xj6c" Feb 17 14:12:10 crc kubenswrapper[4955]: I0217 14:12:10.772305 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/becde1e2-5d09-45d1-b498-32172a394218-catalog-content\") pod \"redhat-marketplace-5xj6c\" (UID: \"becde1e2-5d09-45d1-b498-32172a394218\") " pod="openshift-marketplace/redhat-marketplace-5xj6c" Feb 17 14:12:10 crc kubenswrapper[4955]: I0217 14:12:10.792659 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dl69p\" (UniqueName: \"kubernetes.io/projected/becde1e2-5d09-45d1-b498-32172a394218-kube-api-access-dl69p\") pod \"redhat-marketplace-5xj6c\" (UID: \"becde1e2-5d09-45d1-b498-32172a394218\") " pod="openshift-marketplace/redhat-marketplace-5xj6c" Feb 17 14:12:10 crc kubenswrapper[4955]: I0217 14:12:10.933465 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5xj6c" Feb 17 14:12:11 crc kubenswrapper[4955]: I0217 14:12:11.428921 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zmnm5"] Feb 17 14:12:11 crc kubenswrapper[4955]: I0217 14:12:11.574734 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5xj6c"] Feb 17 14:12:11 crc kubenswrapper[4955]: W0217 14:12:11.580798 4955 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbecde1e2_5d09_45d1_b498_32172a394218.slice/crio-8322da6eb65b61c431a9ae37a608df442e398cbc87f8f3a9d9363a53a1a74189 WatchSource:0}: Error finding container 8322da6eb65b61c431a9ae37a608df442e398cbc87f8f3a9d9363a53a1a74189: Status 404 returned error can't find the container with id 8322da6eb65b61c431a9ae37a608df442e398cbc87f8f3a9d9363a53a1a74189 Feb 17 14:12:11 crc kubenswrapper[4955]: I0217 14:12:11.826148 4955 generic.go:334] "Generic (PLEG): container finished" podID="03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3" containerID="b0277e95434105261fc8e84c066dc0691c0a46871a03a802978f471b108973bd" exitCode=0 Feb 17 14:12:11 crc kubenswrapper[4955]: I0217 14:12:11.826246 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zmnm5" event={"ID":"03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3","Type":"ContainerDied","Data":"b0277e95434105261fc8e84c066dc0691c0a46871a03a802978f471b108973bd"} Feb 17 14:12:11 crc kubenswrapper[4955]: I0217 14:12:11.826507 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zmnm5" event={"ID":"03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3","Type":"ContainerStarted","Data":"1fc91de1ef5de61cb92702c7bc1995d3b324978fd78575c1ecaa0cbacf724f11"} Feb 17 14:12:11 crc kubenswrapper[4955]: I0217 14:12:11.828003 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5xj6c" event={"ID":"becde1e2-5d09-45d1-b498-32172a394218","Type":"ContainerStarted","Data":"8322da6eb65b61c431a9ae37a608df442e398cbc87f8f3a9d9363a53a1a74189"} Feb 17 14:12:11 crc kubenswrapper[4955]: I0217 14:12:11.828297 4955 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 17 14:12:12 crc kubenswrapper[4955]: I0217 14:12:12.839130 4955 generic.go:334] "Generic (PLEG): container finished" podID="becde1e2-5d09-45d1-b498-32172a394218" containerID="9ed714ed7dd8cb7f1c280590b9fa7aa090c65c792f7d42ed20c2f26ade2ce3fe" exitCode=0 Feb 17 14:12:12 crc kubenswrapper[4955]: I0217 14:12:12.839172 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5xj6c" event={"ID":"becde1e2-5d09-45d1-b498-32172a394218","Type":"ContainerDied","Data":"9ed714ed7dd8cb7f1c280590b9fa7aa090c65c792f7d42ed20c2f26ade2ce3fe"} Feb 17 14:12:13 crc kubenswrapper[4955]: I0217 14:12:13.848639 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zmnm5" event={"ID":"03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3","Type":"ContainerStarted","Data":"fa4c75728340f90283d908ea77c8e8b4c831e1f68ab7bc62a37f32c1b650c605"} Feb 17 14:12:14 crc kubenswrapper[4955]: I0217 14:12:14.405163 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vm2zj"] Feb 17 14:12:14 crc kubenswrapper[4955]: I0217 14:12:14.407925 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vm2zj" Feb 17 14:12:14 crc kubenswrapper[4955]: I0217 14:12:14.419337 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vm2zj"] Feb 17 14:12:14 crc kubenswrapper[4955]: I0217 14:12:14.467559 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c0840d8-0253-42de-93db-69d0af8b65d8-catalog-content\") pod \"certified-operators-vm2zj\" (UID: \"3c0840d8-0253-42de-93db-69d0af8b65d8\") " pod="openshift-marketplace/certified-operators-vm2zj" Feb 17 14:12:14 crc kubenswrapper[4955]: I0217 14:12:14.467739 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxbqz\" (UniqueName: \"kubernetes.io/projected/3c0840d8-0253-42de-93db-69d0af8b65d8-kube-api-access-nxbqz\") pod \"certified-operators-vm2zj\" (UID: \"3c0840d8-0253-42de-93db-69d0af8b65d8\") " pod="openshift-marketplace/certified-operators-vm2zj" Feb 17 14:12:14 crc kubenswrapper[4955]: I0217 14:12:14.467806 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c0840d8-0253-42de-93db-69d0af8b65d8-utilities\") pod \"certified-operators-vm2zj\" (UID: \"3c0840d8-0253-42de-93db-69d0af8b65d8\") " pod="openshift-marketplace/certified-operators-vm2zj" Feb 17 14:12:14 crc kubenswrapper[4955]: I0217 14:12:14.570993 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c0840d8-0253-42de-93db-69d0af8b65d8-catalog-content\") pod \"certified-operators-vm2zj\" (UID: \"3c0840d8-0253-42de-93db-69d0af8b65d8\") " pod="openshift-marketplace/certified-operators-vm2zj" Feb 17 14:12:14 crc kubenswrapper[4955]: I0217 14:12:14.571153 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxbqz\" (UniqueName: \"kubernetes.io/projected/3c0840d8-0253-42de-93db-69d0af8b65d8-kube-api-access-nxbqz\") pod \"certified-operators-vm2zj\" (UID: \"3c0840d8-0253-42de-93db-69d0af8b65d8\") " pod="openshift-marketplace/certified-operators-vm2zj" Feb 17 14:12:14 crc kubenswrapper[4955]: I0217 14:12:14.571206 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c0840d8-0253-42de-93db-69d0af8b65d8-utilities\") pod \"certified-operators-vm2zj\" (UID: \"3c0840d8-0253-42de-93db-69d0af8b65d8\") " pod="openshift-marketplace/certified-operators-vm2zj" Feb 17 14:12:14 crc kubenswrapper[4955]: I0217 14:12:14.571672 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c0840d8-0253-42de-93db-69d0af8b65d8-catalog-content\") pod \"certified-operators-vm2zj\" (UID: \"3c0840d8-0253-42de-93db-69d0af8b65d8\") " pod="openshift-marketplace/certified-operators-vm2zj" Feb 17 14:12:14 crc kubenswrapper[4955]: I0217 14:12:14.571757 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c0840d8-0253-42de-93db-69d0af8b65d8-utilities\") pod \"certified-operators-vm2zj\" (UID: \"3c0840d8-0253-42de-93db-69d0af8b65d8\") " pod="openshift-marketplace/certified-operators-vm2zj" Feb 17 14:12:14 crc kubenswrapper[4955]: I0217 14:12:14.597420 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxbqz\" (UniqueName: \"kubernetes.io/projected/3c0840d8-0253-42de-93db-69d0af8b65d8-kube-api-access-nxbqz\") pod \"certified-operators-vm2zj\" (UID: \"3c0840d8-0253-42de-93db-69d0af8b65d8\") " pod="openshift-marketplace/certified-operators-vm2zj" Feb 17 14:12:14 crc kubenswrapper[4955]: I0217 14:12:14.743002 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vm2zj" Feb 17 14:12:14 crc kubenswrapper[4955]: I0217 14:12:14.877456 4955 generic.go:334] "Generic (PLEG): container finished" podID="03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3" containerID="fa4c75728340f90283d908ea77c8e8b4c831e1f68ab7bc62a37f32c1b650c605" exitCode=0 Feb 17 14:12:14 crc kubenswrapper[4955]: I0217 14:12:14.877818 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zmnm5" event={"ID":"03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3","Type":"ContainerDied","Data":"fa4c75728340f90283d908ea77c8e8b4c831e1f68ab7bc62a37f32c1b650c605"} Feb 17 14:12:14 crc kubenswrapper[4955]: I0217 14:12:14.915371 4955 generic.go:334] "Generic (PLEG): container finished" podID="becde1e2-5d09-45d1-b498-32172a394218" containerID="8d31c0eef724bc0b378fa316f3b6f2a5a44501b4549981259bc5373e8213baf4" exitCode=0 Feb 17 14:12:14 crc kubenswrapper[4955]: I0217 14:12:14.915544 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5xj6c" event={"ID":"becde1e2-5d09-45d1-b498-32172a394218","Type":"ContainerDied","Data":"8d31c0eef724bc0b378fa316f3b6f2a5a44501b4549981259bc5373e8213baf4"} Feb 17 14:12:15 crc kubenswrapper[4955]: I0217 14:12:15.360426 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vm2zj"] Feb 17 14:12:15 crc kubenswrapper[4955]: I0217 14:12:15.927890 4955 generic.go:334] "Generic (PLEG): container finished" podID="3c0840d8-0253-42de-93db-69d0af8b65d8" containerID="c34925fd4e677a2ee7f9e6285781f088ec41c0158148755ff6abed37afdd47c1" exitCode=0 Feb 17 14:12:15 crc kubenswrapper[4955]: I0217 14:12:15.927971 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vm2zj" event={"ID":"3c0840d8-0253-42de-93db-69d0af8b65d8","Type":"ContainerDied","Data":"c34925fd4e677a2ee7f9e6285781f088ec41c0158148755ff6abed37afdd47c1"} Feb 17 14:12:15 crc kubenswrapper[4955]: I0217 14:12:15.928160 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vm2zj" event={"ID":"3c0840d8-0253-42de-93db-69d0af8b65d8","Type":"ContainerStarted","Data":"1e3e8b99731f8715637545da88eaaf42590536161d29350eb8416737df23e47c"} Feb 17 14:12:16 crc kubenswrapper[4955]: I0217 14:12:16.942641 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zmnm5" event={"ID":"03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3","Type":"ContainerStarted","Data":"387344b0d3c9ee96f857eea0bd20a9508f00e2be510aef81e827fc663eb14493"} Feb 17 14:12:16 crc kubenswrapper[4955]: I0217 14:12:16.945452 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5xj6c" event={"ID":"becde1e2-5d09-45d1-b498-32172a394218","Type":"ContainerStarted","Data":"63ca367efb5f1f3a6f1d0659f09b617f5dc6a428cef18e7389a609f30a47ecd1"} Feb 17 14:12:16 crc kubenswrapper[4955]: I0217 14:12:16.975994 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zmnm5" podStartSLOduration=3.186897084 podStartE2EDuration="6.975967619s" podCreationTimestamp="2026-02-17 14:12:10 +0000 UTC" firstStartedPulling="2026-02-17 14:12:11.828076413 +0000 UTC m=+4070.350805946" lastFinishedPulling="2026-02-17 14:12:15.617146948 +0000 UTC m=+4074.139876481" observedRunningTime="2026-02-17 14:12:16.962136003 +0000 UTC m=+4075.484865546" watchObservedRunningTime="2026-02-17 14:12:16.975967619 +0000 UTC m=+4075.498697162" Feb 17 14:12:16 crc kubenswrapper[4955]: I0217 14:12:16.996246 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5xj6c" podStartSLOduration=3.515190403 podStartE2EDuration="6.996215548s" podCreationTimestamp="2026-02-17 14:12:10 +0000 UTC" firstStartedPulling="2026-02-17 14:12:12.840821337 +0000 UTC m=+4071.363550870" lastFinishedPulling="2026-02-17 14:12:16.321846472 +0000 UTC m=+4074.844576015" observedRunningTime="2026-02-17 14:12:16.980771566 +0000 UTC m=+4075.503501119" watchObservedRunningTime="2026-02-17 14:12:16.996215548 +0000 UTC m=+4075.518945101" Feb 17 14:12:18 crc kubenswrapper[4955]: I0217 14:12:18.965307 4955 generic.go:334] "Generic (PLEG): container finished" podID="3c0840d8-0253-42de-93db-69d0af8b65d8" containerID="db45e6a529bb706d3560ba09745e4240dfbc9c5a81a7427c41f2a32bc39e75be" exitCode=0 Feb 17 14:12:18 crc kubenswrapper[4955]: I0217 14:12:18.965393 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vm2zj" event={"ID":"3c0840d8-0253-42de-93db-69d0af8b65d8","Type":"ContainerDied","Data":"db45e6a529bb706d3560ba09745e4240dfbc9c5a81a7427c41f2a32bc39e75be"} Feb 17 14:12:19 crc kubenswrapper[4955]: I0217 14:12:19.988151 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vm2zj" event={"ID":"3c0840d8-0253-42de-93db-69d0af8b65d8","Type":"ContainerStarted","Data":"3a35e498e4044bb8af4ff6d4f2e61db3f1d6d133ee08751392f370386f2698b7"} Feb 17 14:12:20 crc kubenswrapper[4955]: I0217 14:12:20.015600 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vm2zj" podStartSLOduration=3.479189215 podStartE2EDuration="6.015576967s" podCreationTimestamp="2026-02-17 14:12:14 +0000 UTC" firstStartedPulling="2026-02-17 14:12:16.947293818 +0000 UTC m=+4075.470023361" lastFinishedPulling="2026-02-17 14:12:19.48368157 +0000 UTC m=+4078.006411113" observedRunningTime="2026-02-17 14:12:20.005247552 +0000 UTC m=+4078.527977115" watchObservedRunningTime="2026-02-17 14:12:20.015576967 +0000 UTC m=+4078.538306520" Feb 17 14:12:20 crc kubenswrapper[4955]: I0217 14:12:20.737930 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zmnm5" Feb 17 14:12:20 crc kubenswrapper[4955]: I0217 14:12:20.738202 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zmnm5" Feb 17 14:12:20 crc kubenswrapper[4955]: I0217 14:12:20.789083 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zmnm5" Feb 17 14:12:20 crc kubenswrapper[4955]: I0217 14:12:20.934450 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5xj6c" Feb 17 14:12:20 crc kubenswrapper[4955]: I0217 14:12:20.934737 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5xj6c" Feb 17 14:12:20 crc kubenswrapper[4955]: I0217 14:12:20.983553 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5xj6c" Feb 17 14:12:22 crc kubenswrapper[4955]: I0217 14:12:22.051716 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5xj6c" Feb 17 14:12:22 crc kubenswrapper[4955]: I0217 14:12:22.054099 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zmnm5" Feb 17 14:12:22 crc kubenswrapper[4955]: I0217 14:12:22.788572 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5xj6c"] Feb 17 14:12:24 crc kubenswrapper[4955]: I0217 14:12:24.020688 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5xj6c" podUID="becde1e2-5d09-45d1-b498-32172a394218" containerName="registry-server" containerID="cri-o://63ca367efb5f1f3a6f1d0659f09b617f5dc6a428cef18e7389a609f30a47ecd1" gracePeriod=2 Feb 17 14:12:24 crc kubenswrapper[4955]: I0217 14:12:24.579131 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5xj6c" Feb 17 14:12:24 crc kubenswrapper[4955]: I0217 14:12:24.693113 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/becde1e2-5d09-45d1-b498-32172a394218-catalog-content\") pod \"becde1e2-5d09-45d1-b498-32172a394218\" (UID: \"becde1e2-5d09-45d1-b498-32172a394218\") " Feb 17 14:12:24 crc kubenswrapper[4955]: I0217 14:12:24.693169 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dl69p\" (UniqueName: \"kubernetes.io/projected/becde1e2-5d09-45d1-b498-32172a394218-kube-api-access-dl69p\") pod \"becde1e2-5d09-45d1-b498-32172a394218\" (UID: \"becde1e2-5d09-45d1-b498-32172a394218\") " Feb 17 14:12:24 crc kubenswrapper[4955]: I0217 14:12:24.693246 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/becde1e2-5d09-45d1-b498-32172a394218-utilities\") pod \"becde1e2-5d09-45d1-b498-32172a394218\" (UID: \"becde1e2-5d09-45d1-b498-32172a394218\") " Feb 17 14:12:24 crc kubenswrapper[4955]: I0217 14:12:24.694405 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/becde1e2-5d09-45d1-b498-32172a394218-utilities" (OuterVolumeSpecName: "utilities") pod "becde1e2-5d09-45d1-b498-32172a394218" (UID: "becde1e2-5d09-45d1-b498-32172a394218"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 14:12:24 crc kubenswrapper[4955]: I0217 14:12:24.710370 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/becde1e2-5d09-45d1-b498-32172a394218-kube-api-access-dl69p" (OuterVolumeSpecName: "kube-api-access-dl69p") pod "becde1e2-5d09-45d1-b498-32172a394218" (UID: "becde1e2-5d09-45d1-b498-32172a394218"). InnerVolumeSpecName "kube-api-access-dl69p". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 14:12:24 crc kubenswrapper[4955]: I0217 14:12:24.744175 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vm2zj" Feb 17 14:12:24 crc kubenswrapper[4955]: I0217 14:12:24.744241 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vm2zj" Feb 17 14:12:24 crc kubenswrapper[4955]: I0217 14:12:24.790197 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vm2zj" Feb 17 14:12:24 crc kubenswrapper[4955]: I0217 14:12:24.795040 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dl69p\" (UniqueName: \"kubernetes.io/projected/becde1e2-5d09-45d1-b498-32172a394218-kube-api-access-dl69p\") on node \"crc\" DevicePath \"\"" Feb 17 14:12:24 crc kubenswrapper[4955]: I0217 14:12:24.795077 4955 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/becde1e2-5d09-45d1-b498-32172a394218-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 14:12:25 crc kubenswrapper[4955]: I0217 14:12:25.034285 4955 generic.go:334] "Generic (PLEG): container finished" podID="becde1e2-5d09-45d1-b498-32172a394218" containerID="63ca367efb5f1f3a6f1d0659f09b617f5dc6a428cef18e7389a609f30a47ecd1" exitCode=0 Feb 17 14:12:25 crc kubenswrapper[4955]: I0217 14:12:25.034373 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5xj6c" Feb 17 14:12:25 crc kubenswrapper[4955]: I0217 14:12:25.034438 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5xj6c" event={"ID":"becde1e2-5d09-45d1-b498-32172a394218","Type":"ContainerDied","Data":"63ca367efb5f1f3a6f1d0659f09b617f5dc6a428cef18e7389a609f30a47ecd1"} Feb 17 14:12:25 crc kubenswrapper[4955]: I0217 14:12:25.034487 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5xj6c" event={"ID":"becde1e2-5d09-45d1-b498-32172a394218","Type":"ContainerDied","Data":"8322da6eb65b61c431a9ae37a608df442e398cbc87f8f3a9d9363a53a1a74189"} Feb 17 14:12:25 crc kubenswrapper[4955]: I0217 14:12:25.034525 4955 scope.go:117] "RemoveContainer" containerID="63ca367efb5f1f3a6f1d0659f09b617f5dc6a428cef18e7389a609f30a47ecd1" Feb 17 14:12:25 crc kubenswrapper[4955]: I0217 14:12:25.057248 4955 scope.go:117] "RemoveContainer" containerID="8d31c0eef724bc0b378fa316f3b6f2a5a44501b4549981259bc5373e8213baf4" Feb 17 14:12:25 crc kubenswrapper[4955]: I0217 14:12:25.091099 4955 scope.go:117] "RemoveContainer" containerID="9ed714ed7dd8cb7f1c280590b9fa7aa090c65c792f7d42ed20c2f26ade2ce3fe" Feb 17 14:12:25 crc kubenswrapper[4955]: I0217 14:12:25.109154 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vm2zj" Feb 17 14:12:25 crc kubenswrapper[4955]: I0217 14:12:25.132876 4955 scope.go:117] "RemoveContainer" containerID="63ca367efb5f1f3a6f1d0659f09b617f5dc6a428cef18e7389a609f30a47ecd1" Feb 17 14:12:25 crc kubenswrapper[4955]: E0217 14:12:25.133563 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63ca367efb5f1f3a6f1d0659f09b617f5dc6a428cef18e7389a609f30a47ecd1\": container with ID starting with 63ca367efb5f1f3a6f1d0659f09b617f5dc6a428cef18e7389a609f30a47ecd1 not found: ID does not exist" containerID="63ca367efb5f1f3a6f1d0659f09b617f5dc6a428cef18e7389a609f30a47ecd1" Feb 17 14:12:25 crc kubenswrapper[4955]: I0217 14:12:25.133623 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63ca367efb5f1f3a6f1d0659f09b617f5dc6a428cef18e7389a609f30a47ecd1"} err="failed to get container status \"63ca367efb5f1f3a6f1d0659f09b617f5dc6a428cef18e7389a609f30a47ecd1\": rpc error: code = NotFound desc = could not find container \"63ca367efb5f1f3a6f1d0659f09b617f5dc6a428cef18e7389a609f30a47ecd1\": container with ID starting with 63ca367efb5f1f3a6f1d0659f09b617f5dc6a428cef18e7389a609f30a47ecd1 not found: ID does not exist" Feb 17 14:12:25 crc kubenswrapper[4955]: I0217 14:12:25.133652 4955 scope.go:117] "RemoveContainer" containerID="8d31c0eef724bc0b378fa316f3b6f2a5a44501b4549981259bc5373e8213baf4" Feb 17 14:12:25 crc kubenswrapper[4955]: E0217 14:12:25.135919 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d31c0eef724bc0b378fa316f3b6f2a5a44501b4549981259bc5373e8213baf4\": container with ID starting with 8d31c0eef724bc0b378fa316f3b6f2a5a44501b4549981259bc5373e8213baf4 not found: ID does not exist" containerID="8d31c0eef724bc0b378fa316f3b6f2a5a44501b4549981259bc5373e8213baf4" Feb 17 14:12:25 crc kubenswrapper[4955]: I0217 14:12:25.135949 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d31c0eef724bc0b378fa316f3b6f2a5a44501b4549981259bc5373e8213baf4"} err="failed to get container status \"8d31c0eef724bc0b378fa316f3b6f2a5a44501b4549981259bc5373e8213baf4\": rpc error: code = NotFound desc = could not find container \"8d31c0eef724bc0b378fa316f3b6f2a5a44501b4549981259bc5373e8213baf4\": container with ID starting with 8d31c0eef724bc0b378fa316f3b6f2a5a44501b4549981259bc5373e8213baf4 not found: ID does not exist" Feb 17 14:12:25 crc kubenswrapper[4955]: I0217 14:12:25.135969 4955 scope.go:117] "RemoveContainer" containerID="9ed714ed7dd8cb7f1c280590b9fa7aa090c65c792f7d42ed20c2f26ade2ce3fe" Feb 17 14:12:25 crc kubenswrapper[4955]: E0217 14:12:25.136324 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ed714ed7dd8cb7f1c280590b9fa7aa090c65c792f7d42ed20c2f26ade2ce3fe\": container with ID starting with 9ed714ed7dd8cb7f1c280590b9fa7aa090c65c792f7d42ed20c2f26ade2ce3fe not found: ID does not exist" containerID="9ed714ed7dd8cb7f1c280590b9fa7aa090c65c792f7d42ed20c2f26ade2ce3fe" Feb 17 14:12:25 crc kubenswrapper[4955]: I0217 14:12:25.136343 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ed714ed7dd8cb7f1c280590b9fa7aa090c65c792f7d42ed20c2f26ade2ce3fe"} err="failed to get container status \"9ed714ed7dd8cb7f1c280590b9fa7aa090c65c792f7d42ed20c2f26ade2ce3fe\": rpc error: code = NotFound desc = could not find container \"9ed714ed7dd8cb7f1c280590b9fa7aa090c65c792f7d42ed20c2f26ade2ce3fe\": container with ID starting with 9ed714ed7dd8cb7f1c280590b9fa7aa090c65c792f7d42ed20c2f26ade2ce3fe not found: ID does not exist" Feb 17 14:12:25 crc kubenswrapper[4955]: I0217 14:12:25.468177 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/becde1e2-5d09-45d1-b498-32172a394218-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "becde1e2-5d09-45d1-b498-32172a394218" (UID: "becde1e2-5d09-45d1-b498-32172a394218"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 14:12:25 crc kubenswrapper[4955]: I0217 14:12:25.507261 4955 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/becde1e2-5d09-45d1-b498-32172a394218-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 14:12:25 crc kubenswrapper[4955]: I0217 14:12:25.680660 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5xj6c"] Feb 17 14:12:25 crc kubenswrapper[4955]: I0217 14:12:25.689192 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5xj6c"] Feb 17 14:12:25 crc kubenswrapper[4955]: E0217 14:12:25.758372 4955 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbecde1e2_5d09_45d1_b498_32172a394218.slice/crio-8322da6eb65b61c431a9ae37a608df442e398cbc87f8f3a9d9363a53a1a74189\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbecde1e2_5d09_45d1_b498_32172a394218.slice\": RecentStats: unable to find data in memory cache]" Feb 17 14:12:26 crc kubenswrapper[4955]: I0217 14:12:26.236224 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="becde1e2-5d09-45d1-b498-32172a394218" path="/var/lib/kubelet/pods/becde1e2-5d09-45d1-b498-32172a394218/volumes" Feb 17 14:12:26 crc kubenswrapper[4955]: I0217 14:12:26.788961 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zmnm5"] Feb 17 14:12:26 crc kubenswrapper[4955]: I0217 14:12:26.789445 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zmnm5" podUID="03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3" containerName="registry-server" containerID="cri-o://387344b0d3c9ee96f857eea0bd20a9508f00e2be510aef81e827fc663eb14493" gracePeriod=2 Feb 17 14:12:27 crc kubenswrapper[4955]: I0217 14:12:27.062355 4955 generic.go:334] "Generic (PLEG): container finished" podID="03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3" containerID="387344b0d3c9ee96f857eea0bd20a9508f00e2be510aef81e827fc663eb14493" exitCode=0 Feb 17 14:12:27 crc kubenswrapper[4955]: I0217 14:12:27.062410 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zmnm5" event={"ID":"03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3","Type":"ContainerDied","Data":"387344b0d3c9ee96f857eea0bd20a9508f00e2be510aef81e827fc663eb14493"} Feb 17 14:12:27 crc kubenswrapper[4955]: I0217 14:12:27.274937 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zmnm5" Feb 17 14:12:27 crc kubenswrapper[4955]: I0217 14:12:27.443394 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbjwl\" (UniqueName: \"kubernetes.io/projected/03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3-kube-api-access-jbjwl\") pod \"03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3\" (UID: \"03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3\") " Feb 17 14:12:27 crc kubenswrapper[4955]: I0217 14:12:27.443510 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3-utilities\") pod \"03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3\" (UID: \"03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3\") " Feb 17 14:12:27 crc kubenswrapper[4955]: I0217 14:12:27.443537 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3-catalog-content\") pod \"03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3\" (UID: \"03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3\") " Feb 17 14:12:27 crc kubenswrapper[4955]: I0217 14:12:27.444875 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3-utilities" (OuterVolumeSpecName: "utilities") pod "03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3" (UID: "03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 14:12:27 crc kubenswrapper[4955]: I0217 14:12:27.448679 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3-kube-api-access-jbjwl" (OuterVolumeSpecName: "kube-api-access-jbjwl") pod "03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3" (UID: "03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3"). InnerVolumeSpecName "kube-api-access-jbjwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 14:12:27 crc kubenswrapper[4955]: I0217 14:12:27.505015 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3" (UID: "03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 14:12:27 crc kubenswrapper[4955]: I0217 14:12:27.545444 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbjwl\" (UniqueName: \"kubernetes.io/projected/03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3-kube-api-access-jbjwl\") on node \"crc\" DevicePath \"\"" Feb 17 14:12:27 crc kubenswrapper[4955]: I0217 14:12:27.545490 4955 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 14:12:27 crc kubenswrapper[4955]: I0217 14:12:27.545501 4955 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 14:12:28 crc kubenswrapper[4955]: I0217 14:12:28.081335 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zmnm5" event={"ID":"03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3","Type":"ContainerDied","Data":"1fc91de1ef5de61cb92702c7bc1995d3b324978fd78575c1ecaa0cbacf724f11"} Feb 17 14:12:28 crc kubenswrapper[4955]: I0217 14:12:28.081405 4955 scope.go:117] "RemoveContainer" containerID="387344b0d3c9ee96f857eea0bd20a9508f00e2be510aef81e827fc663eb14493" Feb 17 14:12:28 crc kubenswrapper[4955]: I0217 14:12:28.081755 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zmnm5" Feb 17 14:12:28 crc kubenswrapper[4955]: I0217 14:12:28.128908 4955 scope.go:117] "RemoveContainer" containerID="fa4c75728340f90283d908ea77c8e8b4c831e1f68ab7bc62a37f32c1b650c605" Feb 17 14:12:28 crc kubenswrapper[4955]: I0217 14:12:28.131914 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zmnm5"] Feb 17 14:12:28 crc kubenswrapper[4955]: I0217 14:12:28.142081 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zmnm5"] Feb 17 14:12:28 crc kubenswrapper[4955]: I0217 14:12:28.236604 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3" path="/var/lib/kubelet/pods/03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3/volumes" Feb 17 14:12:28 crc kubenswrapper[4955]: I0217 14:12:28.575153 4955 scope.go:117] "RemoveContainer" containerID="b0277e95434105261fc8e84c066dc0691c0a46871a03a802978f471b108973bd" Feb 17 14:12:31 crc kubenswrapper[4955]: I0217 14:12:31.387288 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vm2zj"] Feb 17 14:12:31 crc kubenswrapper[4955]: I0217 14:12:31.388016 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vm2zj" podUID="3c0840d8-0253-42de-93db-69d0af8b65d8" containerName="registry-server" containerID="cri-o://3a35e498e4044bb8af4ff6d4f2e61db3f1d6d133ee08751392f370386f2698b7" gracePeriod=2 Feb 17 14:12:31 crc kubenswrapper[4955]: I0217 14:12:31.888486 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vm2zj" Feb 17 14:12:32 crc kubenswrapper[4955]: I0217 14:12:32.038324 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c0840d8-0253-42de-93db-69d0af8b65d8-catalog-content\") pod \"3c0840d8-0253-42de-93db-69d0af8b65d8\" (UID: \"3c0840d8-0253-42de-93db-69d0af8b65d8\") " Feb 17 14:12:32 crc kubenswrapper[4955]: I0217 14:12:32.038406 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxbqz\" (UniqueName: \"kubernetes.io/projected/3c0840d8-0253-42de-93db-69d0af8b65d8-kube-api-access-nxbqz\") pod \"3c0840d8-0253-42de-93db-69d0af8b65d8\" (UID: \"3c0840d8-0253-42de-93db-69d0af8b65d8\") " Feb 17 14:12:32 crc kubenswrapper[4955]: I0217 14:12:32.038614 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c0840d8-0253-42de-93db-69d0af8b65d8-utilities\") pod \"3c0840d8-0253-42de-93db-69d0af8b65d8\" (UID: \"3c0840d8-0253-42de-93db-69d0af8b65d8\") " Feb 17 14:12:32 crc kubenswrapper[4955]: I0217 14:12:32.040290 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c0840d8-0253-42de-93db-69d0af8b65d8-utilities" (OuterVolumeSpecName: "utilities") pod "3c0840d8-0253-42de-93db-69d0af8b65d8" (UID: "3c0840d8-0253-42de-93db-69d0af8b65d8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 14:12:32 crc kubenswrapper[4955]: I0217 14:12:32.054282 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c0840d8-0253-42de-93db-69d0af8b65d8-kube-api-access-nxbqz" (OuterVolumeSpecName: "kube-api-access-nxbqz") pod "3c0840d8-0253-42de-93db-69d0af8b65d8" (UID: "3c0840d8-0253-42de-93db-69d0af8b65d8"). InnerVolumeSpecName "kube-api-access-nxbqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 14:12:32 crc kubenswrapper[4955]: I0217 14:12:32.097732 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c0840d8-0253-42de-93db-69d0af8b65d8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3c0840d8-0253-42de-93db-69d0af8b65d8" (UID: "3c0840d8-0253-42de-93db-69d0af8b65d8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 14:12:32 crc kubenswrapper[4955]: I0217 14:12:32.114607 4955 generic.go:334] "Generic (PLEG): container finished" podID="3c0840d8-0253-42de-93db-69d0af8b65d8" containerID="3a35e498e4044bb8af4ff6d4f2e61db3f1d6d133ee08751392f370386f2698b7" exitCode=0 Feb 17 14:12:32 crc kubenswrapper[4955]: I0217 14:12:32.114655 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vm2zj" event={"ID":"3c0840d8-0253-42de-93db-69d0af8b65d8","Type":"ContainerDied","Data":"3a35e498e4044bb8af4ff6d4f2e61db3f1d6d133ee08751392f370386f2698b7"} Feb 17 14:12:32 crc kubenswrapper[4955]: I0217 14:12:32.114682 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vm2zj" event={"ID":"3c0840d8-0253-42de-93db-69d0af8b65d8","Type":"ContainerDied","Data":"1e3e8b99731f8715637545da88eaaf42590536161d29350eb8416737df23e47c"} Feb 17 14:12:32 crc kubenswrapper[4955]: I0217 14:12:32.114657 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vm2zj" Feb 17 14:12:32 crc kubenswrapper[4955]: I0217 14:12:32.114702 4955 scope.go:117] "RemoveContainer" containerID="3a35e498e4044bb8af4ff6d4f2e61db3f1d6d133ee08751392f370386f2698b7" Feb 17 14:12:32 crc kubenswrapper[4955]: I0217 14:12:32.134215 4955 scope.go:117] "RemoveContainer" containerID="db45e6a529bb706d3560ba09745e4240dfbc9c5a81a7427c41f2a32bc39e75be" Feb 17 14:12:32 crc kubenswrapper[4955]: I0217 14:12:32.141907 4955 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c0840d8-0253-42de-93db-69d0af8b65d8-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 14:12:32 crc kubenswrapper[4955]: I0217 14:12:32.141942 4955 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c0840d8-0253-42de-93db-69d0af8b65d8-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 14:12:32 crc kubenswrapper[4955]: I0217 14:12:32.141958 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nxbqz\" (UniqueName: \"kubernetes.io/projected/3c0840d8-0253-42de-93db-69d0af8b65d8-kube-api-access-nxbqz\") on node \"crc\" DevicePath \"\"" Feb 17 14:12:32 crc kubenswrapper[4955]: I0217 14:12:32.156829 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vm2zj"] Feb 17 14:12:32 crc kubenswrapper[4955]: I0217 14:12:32.171677 4955 scope.go:117] "RemoveContainer" containerID="c34925fd4e677a2ee7f9e6285781f088ec41c0158148755ff6abed37afdd47c1" Feb 17 14:12:32 crc kubenswrapper[4955]: I0217 14:12:32.190983 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vm2zj"] Feb 17 14:12:32 crc kubenswrapper[4955]: I0217 14:12:32.207065 4955 scope.go:117] "RemoveContainer" containerID="3a35e498e4044bb8af4ff6d4f2e61db3f1d6d133ee08751392f370386f2698b7" Feb 17 14:12:32 crc kubenswrapper[4955]: E0217 14:12:32.207661 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a35e498e4044bb8af4ff6d4f2e61db3f1d6d133ee08751392f370386f2698b7\": container with ID starting with 3a35e498e4044bb8af4ff6d4f2e61db3f1d6d133ee08751392f370386f2698b7 not found: ID does not exist" containerID="3a35e498e4044bb8af4ff6d4f2e61db3f1d6d133ee08751392f370386f2698b7" Feb 17 14:12:32 crc kubenswrapper[4955]: I0217 14:12:32.207719 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a35e498e4044bb8af4ff6d4f2e61db3f1d6d133ee08751392f370386f2698b7"} err="failed to get container status \"3a35e498e4044bb8af4ff6d4f2e61db3f1d6d133ee08751392f370386f2698b7\": rpc error: code = NotFound desc = could not find container \"3a35e498e4044bb8af4ff6d4f2e61db3f1d6d133ee08751392f370386f2698b7\": container with ID starting with 3a35e498e4044bb8af4ff6d4f2e61db3f1d6d133ee08751392f370386f2698b7 not found: ID does not exist" Feb 17 14:12:32 crc kubenswrapper[4955]: I0217 14:12:32.207746 4955 scope.go:117] "RemoveContainer" containerID="db45e6a529bb706d3560ba09745e4240dfbc9c5a81a7427c41f2a32bc39e75be" Feb 17 14:12:32 crc kubenswrapper[4955]: E0217 14:12:32.208172 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db45e6a529bb706d3560ba09745e4240dfbc9c5a81a7427c41f2a32bc39e75be\": container with ID starting with db45e6a529bb706d3560ba09745e4240dfbc9c5a81a7427c41f2a32bc39e75be not found: ID does not exist" containerID="db45e6a529bb706d3560ba09745e4240dfbc9c5a81a7427c41f2a32bc39e75be" Feb 17 14:12:32 crc kubenswrapper[4955]: I0217 14:12:32.208188 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db45e6a529bb706d3560ba09745e4240dfbc9c5a81a7427c41f2a32bc39e75be"} err="failed to get container status \"db45e6a529bb706d3560ba09745e4240dfbc9c5a81a7427c41f2a32bc39e75be\": rpc error: code = NotFound desc = could not find container \"db45e6a529bb706d3560ba09745e4240dfbc9c5a81a7427c41f2a32bc39e75be\": container with ID starting with db45e6a529bb706d3560ba09745e4240dfbc9c5a81a7427c41f2a32bc39e75be not found: ID does not exist" Feb 17 14:12:32 crc kubenswrapper[4955]: I0217 14:12:32.208202 4955 scope.go:117] "RemoveContainer" containerID="c34925fd4e677a2ee7f9e6285781f088ec41c0158148755ff6abed37afdd47c1" Feb 17 14:12:32 crc kubenswrapper[4955]: E0217 14:12:32.208408 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c34925fd4e677a2ee7f9e6285781f088ec41c0158148755ff6abed37afdd47c1\": container with ID starting with c34925fd4e677a2ee7f9e6285781f088ec41c0158148755ff6abed37afdd47c1 not found: ID does not exist" containerID="c34925fd4e677a2ee7f9e6285781f088ec41c0158148755ff6abed37afdd47c1" Feb 17 14:12:32 crc kubenswrapper[4955]: I0217 14:12:32.208421 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c34925fd4e677a2ee7f9e6285781f088ec41c0158148755ff6abed37afdd47c1"} err="failed to get container status \"c34925fd4e677a2ee7f9e6285781f088ec41c0158148755ff6abed37afdd47c1\": rpc error: code = NotFound desc = could not find container \"c34925fd4e677a2ee7f9e6285781f088ec41c0158148755ff6abed37afdd47c1\": container with ID starting with c34925fd4e677a2ee7f9e6285781f088ec41c0158148755ff6abed37afdd47c1 not found: ID does not exist" Feb 17 14:12:32 crc kubenswrapper[4955]: I0217 14:12:32.232346 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c0840d8-0253-42de-93db-69d0af8b65d8" path="/var/lib/kubelet/pods/3c0840d8-0253-42de-93db-69d0af8b65d8/volumes" Feb 17 14:12:34 crc kubenswrapper[4955]: I0217 14:12:34.575322 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 14:12:34 crc kubenswrapper[4955]: I0217 14:12:34.575424 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 14:12:46 crc kubenswrapper[4955]: I0217 14:12:46.229609 4955 generic.go:334] "Generic (PLEG): container finished" podID="0bdb56fb-1c91-4402-a733-8cbcb46e21ca" containerID="1e5d05066dd216411721ec16d11df4de4f618415b573b58817c27c2628c874d4" exitCode=0 Feb 17 14:12:46 crc kubenswrapper[4955]: I0217 14:12:46.236066 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-sncmd/must-gather-dd9pl" event={"ID":"0bdb56fb-1c91-4402-a733-8cbcb46e21ca","Type":"ContainerDied","Data":"1e5d05066dd216411721ec16d11df4de4f618415b573b58817c27c2628c874d4"} Feb 17 14:12:46 crc kubenswrapper[4955]: I0217 14:12:46.236638 4955 scope.go:117] "RemoveContainer" containerID="1e5d05066dd216411721ec16d11df4de4f618415b573b58817c27c2628c874d4" Feb 17 14:12:46 crc kubenswrapper[4955]: I0217 14:12:46.457559 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-sncmd_must-gather-dd9pl_0bdb56fb-1c91-4402-a733-8cbcb46e21ca/gather/0.log" Feb 17 14:12:57 crc kubenswrapper[4955]: I0217 14:12:57.934644 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-sncmd/must-gather-dd9pl"] Feb 17 14:12:57 crc kubenswrapper[4955]: I0217 14:12:57.935499 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-sncmd/must-gather-dd9pl" podUID="0bdb56fb-1c91-4402-a733-8cbcb46e21ca" containerName="copy" containerID="cri-o://ce8d24b61088b3bc0703e041dc3fea0ef78865e1dd1d0c9bfc1604ee13a3a69c" gracePeriod=2 Feb 17 14:12:57 crc kubenswrapper[4955]: I0217 14:12:57.942518 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-sncmd/must-gather-dd9pl"] Feb 17 14:12:58 crc kubenswrapper[4955]: I0217 14:12:58.337889 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-sncmd_must-gather-dd9pl_0bdb56fb-1c91-4402-a733-8cbcb46e21ca/copy/0.log" Feb 17 14:12:58 crc kubenswrapper[4955]: I0217 14:12:58.338703 4955 generic.go:334] "Generic (PLEG): container finished" podID="0bdb56fb-1c91-4402-a733-8cbcb46e21ca" containerID="ce8d24b61088b3bc0703e041dc3fea0ef78865e1dd1d0c9bfc1604ee13a3a69c" exitCode=143 Feb 17 14:12:58 crc kubenswrapper[4955]: I0217 14:12:58.450542 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-sncmd_must-gather-dd9pl_0bdb56fb-1c91-4402-a733-8cbcb46e21ca/copy/0.log" Feb 17 14:12:58 crc kubenswrapper[4955]: I0217 14:12:58.450954 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sncmd/must-gather-dd9pl" Feb 17 14:12:58 crc kubenswrapper[4955]: I0217 14:12:58.547700 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0bdb56fb-1c91-4402-a733-8cbcb46e21ca-must-gather-output\") pod \"0bdb56fb-1c91-4402-a733-8cbcb46e21ca\" (UID: \"0bdb56fb-1c91-4402-a733-8cbcb46e21ca\") " Feb 17 14:12:58 crc kubenswrapper[4955]: I0217 14:12:58.548228 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2ntf\" (UniqueName: \"kubernetes.io/projected/0bdb56fb-1c91-4402-a733-8cbcb46e21ca-kube-api-access-j2ntf\") pod \"0bdb56fb-1c91-4402-a733-8cbcb46e21ca\" (UID: \"0bdb56fb-1c91-4402-a733-8cbcb46e21ca\") " Feb 17 14:12:58 crc kubenswrapper[4955]: I0217 14:12:58.554576 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bdb56fb-1c91-4402-a733-8cbcb46e21ca-kube-api-access-j2ntf" (OuterVolumeSpecName: "kube-api-access-j2ntf") pod "0bdb56fb-1c91-4402-a733-8cbcb46e21ca" (UID: "0bdb56fb-1c91-4402-a733-8cbcb46e21ca"). InnerVolumeSpecName "kube-api-access-j2ntf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 14:12:58 crc kubenswrapper[4955]: I0217 14:12:58.650015 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2ntf\" (UniqueName: \"kubernetes.io/projected/0bdb56fb-1c91-4402-a733-8cbcb46e21ca-kube-api-access-j2ntf\") on node \"crc\" DevicePath \"\"" Feb 17 14:12:58 crc kubenswrapper[4955]: I0217 14:12:58.707776 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0bdb56fb-1c91-4402-a733-8cbcb46e21ca-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "0bdb56fb-1c91-4402-a733-8cbcb46e21ca" (UID: "0bdb56fb-1c91-4402-a733-8cbcb46e21ca"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 14:12:58 crc kubenswrapper[4955]: I0217 14:12:58.752019 4955 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0bdb56fb-1c91-4402-a733-8cbcb46e21ca-must-gather-output\") on node \"crc\" DevicePath \"\"" Feb 17 14:12:59 crc kubenswrapper[4955]: I0217 14:12:59.347685 4955 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-sncmd_must-gather-dd9pl_0bdb56fb-1c91-4402-a733-8cbcb46e21ca/copy/0.log" Feb 17 14:12:59 crc kubenswrapper[4955]: I0217 14:12:59.348107 4955 scope.go:117] "RemoveContainer" containerID="ce8d24b61088b3bc0703e041dc3fea0ef78865e1dd1d0c9bfc1604ee13a3a69c" Feb 17 14:12:59 crc kubenswrapper[4955]: I0217 14:12:59.348196 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-sncmd/must-gather-dd9pl" Feb 17 14:12:59 crc kubenswrapper[4955]: I0217 14:12:59.366452 4955 scope.go:117] "RemoveContainer" containerID="1e5d05066dd216411721ec16d11df4de4f618415b573b58817c27c2628c874d4" Feb 17 14:13:00 crc kubenswrapper[4955]: I0217 14:13:00.251654 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0bdb56fb-1c91-4402-a733-8cbcb46e21ca" path="/var/lib/kubelet/pods/0bdb56fb-1c91-4402-a733-8cbcb46e21ca/volumes" Feb 17 14:13:04 crc kubenswrapper[4955]: I0217 14:13:04.574678 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 14:13:04 crc kubenswrapper[4955]: I0217 14:13:04.575360 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 14:13:04 crc kubenswrapper[4955]: I0217 14:13:04.575415 4955 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" Feb 17 14:13:04 crc kubenswrapper[4955]: I0217 14:13:04.576183 4955 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"91eab0ac0f5514297337a55f7c52c0dd490bc24e8afd932deb11adb7e1d444e2"} pod="openshift-machine-config-operator/machine-config-daemon-29qxq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 17 14:13:04 crc kubenswrapper[4955]: I0217 14:13:04.576249 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" containerID="cri-o://91eab0ac0f5514297337a55f7c52c0dd490bc24e8afd932deb11adb7e1d444e2" gracePeriod=600 Feb 17 14:13:05 crc kubenswrapper[4955]: I0217 14:13:05.403629 4955 generic.go:334] "Generic (PLEG): container finished" podID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerID="91eab0ac0f5514297337a55f7c52c0dd490bc24e8afd932deb11adb7e1d444e2" exitCode=0 Feb 17 14:13:05 crc kubenswrapper[4955]: I0217 14:13:05.403712 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" event={"ID":"5bf36887-2a2b-4193-abdb-a15beaf46300","Type":"ContainerDied","Data":"91eab0ac0f5514297337a55f7c52c0dd490bc24e8afd932deb11adb7e1d444e2"} Feb 17 14:13:05 crc kubenswrapper[4955]: I0217 14:13:05.404045 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" event={"ID":"5bf36887-2a2b-4193-abdb-a15beaf46300","Type":"ContainerStarted","Data":"0f74b5105f865cc62e93aedd90435f5aa7c0ed8ad937b4e2a170054c21f3de5a"} Feb 17 14:13:05 crc kubenswrapper[4955]: I0217 14:13:05.404068 4955 scope.go:117] "RemoveContainer" containerID="fa86f3109bf152769bd9369dabc17d54290d99858877992d6ce0702550e664a2" Feb 17 14:13:52 crc kubenswrapper[4955]: I0217 14:13:52.152125 4955 scope.go:117] "RemoveContainer" containerID="3ee69b90a7c7d296739f074bae24e7e64696ee6730c62e510684aa440cb2b8a0" Feb 17 14:15:00 crc kubenswrapper[4955]: I0217 14:15:00.186448 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29522295-vhzjg"] Feb 17 14:15:00 crc kubenswrapper[4955]: E0217 14:15:00.187295 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3" containerName="extract-utilities" Feb 17 14:15:00 crc kubenswrapper[4955]: I0217 14:15:00.187307 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3" containerName="extract-utilities" Feb 17 14:15:00 crc kubenswrapper[4955]: E0217 14:15:00.187319 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="becde1e2-5d09-45d1-b498-32172a394218" containerName="registry-server" Feb 17 14:15:00 crc kubenswrapper[4955]: I0217 14:15:00.187325 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="becde1e2-5d09-45d1-b498-32172a394218" containerName="registry-server" Feb 17 14:15:00 crc kubenswrapper[4955]: E0217 14:15:00.187333 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3" containerName="registry-server" Feb 17 14:15:00 crc kubenswrapper[4955]: I0217 14:15:00.187339 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3" containerName="registry-server" Feb 17 14:15:00 crc kubenswrapper[4955]: E0217 14:15:00.187353 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bdb56fb-1c91-4402-a733-8cbcb46e21ca" containerName="gather" Feb 17 14:15:00 crc kubenswrapper[4955]: I0217 14:15:00.187360 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bdb56fb-1c91-4402-a733-8cbcb46e21ca" containerName="gather" Feb 17 14:15:00 crc kubenswrapper[4955]: E0217 14:15:00.187374 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="becde1e2-5d09-45d1-b498-32172a394218" containerName="extract-utilities" Feb 17 14:15:00 crc kubenswrapper[4955]: I0217 14:15:00.187380 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="becde1e2-5d09-45d1-b498-32172a394218" containerName="extract-utilities" Feb 17 14:15:00 crc kubenswrapper[4955]: E0217 14:15:00.187390 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c0840d8-0253-42de-93db-69d0af8b65d8" containerName="registry-server" Feb 17 14:15:00 crc kubenswrapper[4955]: I0217 14:15:00.187396 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c0840d8-0253-42de-93db-69d0af8b65d8" containerName="registry-server" Feb 17 14:15:00 crc kubenswrapper[4955]: E0217 14:15:00.187409 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c0840d8-0253-42de-93db-69d0af8b65d8" containerName="extract-utilities" Feb 17 14:15:00 crc kubenswrapper[4955]: I0217 14:15:00.187415 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c0840d8-0253-42de-93db-69d0af8b65d8" containerName="extract-utilities" Feb 17 14:15:00 crc kubenswrapper[4955]: E0217 14:15:00.187426 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c0840d8-0253-42de-93db-69d0af8b65d8" containerName="extract-content" Feb 17 14:15:00 crc kubenswrapper[4955]: I0217 14:15:00.187432 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c0840d8-0253-42de-93db-69d0af8b65d8" containerName="extract-content" Feb 17 14:15:00 crc kubenswrapper[4955]: E0217 14:15:00.187441 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="becde1e2-5d09-45d1-b498-32172a394218" containerName="extract-content" Feb 17 14:15:00 crc kubenswrapper[4955]: I0217 14:15:00.187447 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="becde1e2-5d09-45d1-b498-32172a394218" containerName="extract-content" Feb 17 14:15:00 crc kubenswrapper[4955]: E0217 14:15:00.187459 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3" containerName="extract-content" Feb 17 14:15:00 crc kubenswrapper[4955]: I0217 14:15:00.187465 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3" containerName="extract-content" Feb 17 14:15:00 crc kubenswrapper[4955]: E0217 14:15:00.187479 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bdb56fb-1c91-4402-a733-8cbcb46e21ca" containerName="copy" Feb 17 14:15:00 crc kubenswrapper[4955]: I0217 14:15:00.187484 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bdb56fb-1c91-4402-a733-8cbcb46e21ca" containerName="copy" Feb 17 14:15:00 crc kubenswrapper[4955]: I0217 14:15:00.187644 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bdb56fb-1c91-4402-a733-8cbcb46e21ca" containerName="gather" Feb 17 14:15:00 crc kubenswrapper[4955]: I0217 14:15:00.187657 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="03e9edbf-e333-4a3a-b7d5-0c34f21b2fd3" containerName="registry-server" Feb 17 14:15:00 crc kubenswrapper[4955]: I0217 14:15:00.187668 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c0840d8-0253-42de-93db-69d0af8b65d8" containerName="registry-server" Feb 17 14:15:00 crc kubenswrapper[4955]: I0217 14:15:00.187678 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bdb56fb-1c91-4402-a733-8cbcb46e21ca" containerName="copy" Feb 17 14:15:00 crc kubenswrapper[4955]: I0217 14:15:00.187686 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="becde1e2-5d09-45d1-b498-32172a394218" containerName="registry-server" Feb 17 14:15:00 crc kubenswrapper[4955]: I0217 14:15:00.188300 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29522295-vhzjg" Feb 17 14:15:00 crc kubenswrapper[4955]: I0217 14:15:00.190593 4955 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 17 14:15:00 crc kubenswrapper[4955]: I0217 14:15:00.190966 4955 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 17 14:15:00 crc kubenswrapper[4955]: I0217 14:15:00.198908 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29522295-vhzjg"] Feb 17 14:15:00 crc kubenswrapper[4955]: I0217 14:15:00.330958 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/92437800-eebd-4c82-8a93-0fc6bad6bd11-config-volume\") pod \"collect-profiles-29522295-vhzjg\" (UID: \"92437800-eebd-4c82-8a93-0fc6bad6bd11\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522295-vhzjg" Feb 17 14:15:00 crc kubenswrapper[4955]: I0217 14:15:00.331060 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/92437800-eebd-4c82-8a93-0fc6bad6bd11-secret-volume\") pod \"collect-profiles-29522295-vhzjg\" (UID: \"92437800-eebd-4c82-8a93-0fc6bad6bd11\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522295-vhzjg" Feb 17 14:15:00 crc kubenswrapper[4955]: I0217 14:15:00.331089 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4pms\" (UniqueName: \"kubernetes.io/projected/92437800-eebd-4c82-8a93-0fc6bad6bd11-kube-api-access-w4pms\") pod \"collect-profiles-29522295-vhzjg\" (UID: \"92437800-eebd-4c82-8a93-0fc6bad6bd11\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522295-vhzjg" Feb 17 14:15:00 crc kubenswrapper[4955]: I0217 14:15:00.432342 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/92437800-eebd-4c82-8a93-0fc6bad6bd11-config-volume\") pod \"collect-profiles-29522295-vhzjg\" (UID: \"92437800-eebd-4c82-8a93-0fc6bad6bd11\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522295-vhzjg" Feb 17 14:15:00 crc kubenswrapper[4955]: I0217 14:15:00.432651 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/92437800-eebd-4c82-8a93-0fc6bad6bd11-secret-volume\") pod \"collect-profiles-29522295-vhzjg\" (UID: \"92437800-eebd-4c82-8a93-0fc6bad6bd11\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522295-vhzjg" Feb 17 14:15:00 crc kubenswrapper[4955]: I0217 14:15:00.432798 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4pms\" (UniqueName: \"kubernetes.io/projected/92437800-eebd-4c82-8a93-0fc6bad6bd11-kube-api-access-w4pms\") pod \"collect-profiles-29522295-vhzjg\" (UID: \"92437800-eebd-4c82-8a93-0fc6bad6bd11\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522295-vhzjg" Feb 17 14:15:00 crc kubenswrapper[4955]: I0217 14:15:00.433362 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/92437800-eebd-4c82-8a93-0fc6bad6bd11-config-volume\") pod \"collect-profiles-29522295-vhzjg\" (UID: \"92437800-eebd-4c82-8a93-0fc6bad6bd11\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522295-vhzjg" Feb 17 14:15:00 crc kubenswrapper[4955]: I0217 14:15:00.438634 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/92437800-eebd-4c82-8a93-0fc6bad6bd11-secret-volume\") pod \"collect-profiles-29522295-vhzjg\" (UID: \"92437800-eebd-4c82-8a93-0fc6bad6bd11\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522295-vhzjg" Feb 17 14:15:00 crc kubenswrapper[4955]: I0217 14:15:00.451525 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4pms\" (UniqueName: \"kubernetes.io/projected/92437800-eebd-4c82-8a93-0fc6bad6bd11-kube-api-access-w4pms\") pod \"collect-profiles-29522295-vhzjg\" (UID: \"92437800-eebd-4c82-8a93-0fc6bad6bd11\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29522295-vhzjg" Feb 17 14:15:00 crc kubenswrapper[4955]: I0217 14:15:00.519268 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29522295-vhzjg" Feb 17 14:15:01 crc kubenswrapper[4955]: I0217 14:15:01.003935 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29522295-vhzjg"] Feb 17 14:15:01 crc kubenswrapper[4955]: I0217 14:15:01.495848 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29522295-vhzjg" event={"ID":"92437800-eebd-4c82-8a93-0fc6bad6bd11","Type":"ContainerStarted","Data":"befbaa5325bc36e3978b8b97db86fb02a0c1bbba11febe45e956784e5269b24a"} Feb 17 14:15:01 crc kubenswrapper[4955]: I0217 14:15:01.495901 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29522295-vhzjg" event={"ID":"92437800-eebd-4c82-8a93-0fc6bad6bd11","Type":"ContainerStarted","Data":"6c10e370e3d618e879589fefaa894cbaaf71e447df978ce1f3bef4596adda15d"} Feb 17 14:15:01 crc kubenswrapper[4955]: I0217 14:15:01.521252 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29522295-vhzjg" podStartSLOduration=1.521232875 podStartE2EDuration="1.521232875s" podCreationTimestamp="2026-02-17 14:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 14:15:01.516133259 +0000 UTC m=+4240.038862802" watchObservedRunningTime="2026-02-17 14:15:01.521232875 +0000 UTC m=+4240.043962418" Feb 17 14:15:02 crc kubenswrapper[4955]: I0217 14:15:02.504713 4955 generic.go:334] "Generic (PLEG): container finished" podID="92437800-eebd-4c82-8a93-0fc6bad6bd11" containerID="befbaa5325bc36e3978b8b97db86fb02a0c1bbba11febe45e956784e5269b24a" exitCode=0 Feb 17 14:15:02 crc kubenswrapper[4955]: I0217 14:15:02.505039 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29522295-vhzjg" event={"ID":"92437800-eebd-4c82-8a93-0fc6bad6bd11","Type":"ContainerDied","Data":"befbaa5325bc36e3978b8b97db86fb02a0c1bbba11febe45e956784e5269b24a"} Feb 17 14:15:04 crc kubenswrapper[4955]: I0217 14:15:04.057934 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29522295-vhzjg" Feb 17 14:15:04 crc kubenswrapper[4955]: I0217 14:15:04.266946 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/92437800-eebd-4c82-8a93-0fc6bad6bd11-secret-volume\") pod \"92437800-eebd-4c82-8a93-0fc6bad6bd11\" (UID: \"92437800-eebd-4c82-8a93-0fc6bad6bd11\") " Feb 17 14:15:04 crc kubenswrapper[4955]: I0217 14:15:04.267031 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4pms\" (UniqueName: \"kubernetes.io/projected/92437800-eebd-4c82-8a93-0fc6bad6bd11-kube-api-access-w4pms\") pod \"92437800-eebd-4c82-8a93-0fc6bad6bd11\" (UID: \"92437800-eebd-4c82-8a93-0fc6bad6bd11\") " Feb 17 14:15:04 crc kubenswrapper[4955]: I0217 14:15:04.267056 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/92437800-eebd-4c82-8a93-0fc6bad6bd11-config-volume\") pod \"92437800-eebd-4c82-8a93-0fc6bad6bd11\" (UID: \"92437800-eebd-4c82-8a93-0fc6bad6bd11\") " Feb 17 14:15:04 crc kubenswrapper[4955]: I0217 14:15:04.267897 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92437800-eebd-4c82-8a93-0fc6bad6bd11-config-volume" (OuterVolumeSpecName: "config-volume") pod "92437800-eebd-4c82-8a93-0fc6bad6bd11" (UID: "92437800-eebd-4c82-8a93-0fc6bad6bd11"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 17 14:15:04 crc kubenswrapper[4955]: I0217 14:15:04.273696 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92437800-eebd-4c82-8a93-0fc6bad6bd11-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "92437800-eebd-4c82-8a93-0fc6bad6bd11" (UID: "92437800-eebd-4c82-8a93-0fc6bad6bd11"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 17 14:15:04 crc kubenswrapper[4955]: I0217 14:15:04.274424 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92437800-eebd-4c82-8a93-0fc6bad6bd11-kube-api-access-w4pms" (OuterVolumeSpecName: "kube-api-access-w4pms") pod "92437800-eebd-4c82-8a93-0fc6bad6bd11" (UID: "92437800-eebd-4c82-8a93-0fc6bad6bd11"). InnerVolumeSpecName "kube-api-access-w4pms". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 14:15:04 crc kubenswrapper[4955]: I0217 14:15:04.370712 4955 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/92437800-eebd-4c82-8a93-0fc6bad6bd11-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 17 14:15:04 crc kubenswrapper[4955]: I0217 14:15:04.370746 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4pms\" (UniqueName: \"kubernetes.io/projected/92437800-eebd-4c82-8a93-0fc6bad6bd11-kube-api-access-w4pms\") on node \"crc\" DevicePath \"\"" Feb 17 14:15:04 crc kubenswrapper[4955]: I0217 14:15:04.370762 4955 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/92437800-eebd-4c82-8a93-0fc6bad6bd11-config-volume\") on node \"crc\" DevicePath \"\"" Feb 17 14:15:04 crc kubenswrapper[4955]: I0217 14:15:04.524389 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29522295-vhzjg" event={"ID":"92437800-eebd-4c82-8a93-0fc6bad6bd11","Type":"ContainerDied","Data":"6c10e370e3d618e879589fefaa894cbaaf71e447df978ce1f3bef4596adda15d"} Feb 17 14:15:04 crc kubenswrapper[4955]: I0217 14:15:04.524743 4955 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c10e370e3d618e879589fefaa894cbaaf71e447df978ce1f3bef4596adda15d" Feb 17 14:15:04 crc kubenswrapper[4955]: I0217 14:15:04.524460 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29522295-vhzjg" Feb 17 14:15:04 crc kubenswrapper[4955]: I0217 14:15:04.577109 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 14:15:04 crc kubenswrapper[4955]: I0217 14:15:04.577179 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 14:15:04 crc kubenswrapper[4955]: I0217 14:15:04.593500 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29522250-m57xx"] Feb 17 14:15:04 crc kubenswrapper[4955]: I0217 14:15:04.602231 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29522250-m57xx"] Feb 17 14:15:06 crc kubenswrapper[4955]: I0217 14:15:06.233177 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1fabad5b-2fa4-47a1-93bc-70d5b6299a41" path="/var/lib/kubelet/pods/1fabad5b-2fa4-47a1-93bc-70d5b6299a41/volumes" Feb 17 14:15:34 crc kubenswrapper[4955]: I0217 14:15:34.574989 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 14:15:34 crc kubenswrapper[4955]: I0217 14:15:34.575610 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 14:15:52 crc kubenswrapper[4955]: I0217 14:15:52.237122 4955 scope.go:117] "RemoveContainer" containerID="15fa28350c7f25ce62e418796a1b0df9b9d93076a4bf3a9597ac4c87f2caaea8" Feb 17 14:16:04 crc kubenswrapper[4955]: I0217 14:16:04.575195 4955 patch_prober.go:28] interesting pod/machine-config-daemon-29qxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 17 14:16:04 crc kubenswrapper[4955]: I0217 14:16:04.575851 4955 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 17 14:16:04 crc kubenswrapper[4955]: I0217 14:16:04.575926 4955 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" Feb 17 14:16:04 crc kubenswrapper[4955]: I0217 14:16:04.576738 4955 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0f74b5105f865cc62e93aedd90435f5aa7c0ed8ad937b4e2a170054c21f3de5a"} pod="openshift-machine-config-operator/machine-config-daemon-29qxq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 17 14:16:04 crc kubenswrapper[4955]: I0217 14:16:04.576821 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerName="machine-config-daemon" containerID="cri-o://0f74b5105f865cc62e93aedd90435f5aa7c0ed8ad937b4e2a170054c21f3de5a" gracePeriod=600 Feb 17 14:16:04 crc kubenswrapper[4955]: E0217 14:16:04.716942 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 14:16:05 crc kubenswrapper[4955]: I0217 14:16:05.178806 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" event={"ID":"5bf36887-2a2b-4193-abdb-a15beaf46300","Type":"ContainerDied","Data":"0f74b5105f865cc62e93aedd90435f5aa7c0ed8ad937b4e2a170054c21f3de5a"} Feb 17 14:16:05 crc kubenswrapper[4955]: I0217 14:16:05.178879 4955 scope.go:117] "RemoveContainer" containerID="91eab0ac0f5514297337a55f7c52c0dd490bc24e8afd932deb11adb7e1d444e2" Feb 17 14:16:05 crc kubenswrapper[4955]: I0217 14:16:05.178805 4955 generic.go:334] "Generic (PLEG): container finished" podID="5bf36887-2a2b-4193-abdb-a15beaf46300" containerID="0f74b5105f865cc62e93aedd90435f5aa7c0ed8ad937b4e2a170054c21f3de5a" exitCode=0 Feb 17 14:16:05 crc kubenswrapper[4955]: I0217 14:16:05.179566 4955 scope.go:117] "RemoveContainer" containerID="0f74b5105f865cc62e93aedd90435f5aa7c0ed8ad937b4e2a170054c21f3de5a" Feb 17 14:16:05 crc kubenswrapper[4955]: E0217 14:16:05.179882 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 14:16:19 crc kubenswrapper[4955]: I0217 14:16:19.224112 4955 scope.go:117] "RemoveContainer" containerID="0f74b5105f865cc62e93aedd90435f5aa7c0ed8ad937b4e2a170054c21f3de5a" Feb 17 14:16:19 crc kubenswrapper[4955]: E0217 14:16:19.225420 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 14:16:30 crc kubenswrapper[4955]: I0217 14:16:30.224122 4955 scope.go:117] "RemoveContainer" containerID="0f74b5105f865cc62e93aedd90435f5aa7c0ed8ad937b4e2a170054c21f3de5a" Feb 17 14:16:30 crc kubenswrapper[4955]: E0217 14:16:30.225095 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 14:16:43 crc kubenswrapper[4955]: I0217 14:16:43.223234 4955 scope.go:117] "RemoveContainer" containerID="0f74b5105f865cc62e93aedd90435f5aa7c0ed8ad937b4e2a170054c21f3de5a" Feb 17 14:16:43 crc kubenswrapper[4955]: E0217 14:16:43.224506 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 14:16:47 crc kubenswrapper[4955]: I0217 14:16:47.888270 4955 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ck72j"] Feb 17 14:16:47 crc kubenswrapper[4955]: E0217 14:16:47.890062 4955 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92437800-eebd-4c82-8a93-0fc6bad6bd11" containerName="collect-profiles" Feb 17 14:16:47 crc kubenswrapper[4955]: I0217 14:16:47.890081 4955 state_mem.go:107] "Deleted CPUSet assignment" podUID="92437800-eebd-4c82-8a93-0fc6bad6bd11" containerName="collect-profiles" Feb 17 14:16:47 crc kubenswrapper[4955]: I0217 14:16:47.892118 4955 memory_manager.go:354] "RemoveStaleState removing state" podUID="92437800-eebd-4c82-8a93-0fc6bad6bd11" containerName="collect-profiles" Feb 17 14:16:47 crc kubenswrapper[4955]: I0217 14:16:47.902688 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ck72j" Feb 17 14:16:47 crc kubenswrapper[4955]: I0217 14:16:47.943860 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ck72j"] Feb 17 14:16:48 crc kubenswrapper[4955]: I0217 14:16:48.003551 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8wkm\" (UniqueName: \"kubernetes.io/projected/960f0b07-5fd6-4c0b-8808-909939ff95db-kube-api-access-h8wkm\") pod \"redhat-operators-ck72j\" (UID: \"960f0b07-5fd6-4c0b-8808-909939ff95db\") " pod="openshift-marketplace/redhat-operators-ck72j" Feb 17 14:16:48 crc kubenswrapper[4955]: I0217 14:16:48.003611 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/960f0b07-5fd6-4c0b-8808-909939ff95db-catalog-content\") pod \"redhat-operators-ck72j\" (UID: \"960f0b07-5fd6-4c0b-8808-909939ff95db\") " pod="openshift-marketplace/redhat-operators-ck72j" Feb 17 14:16:48 crc kubenswrapper[4955]: I0217 14:16:48.003955 4955 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/960f0b07-5fd6-4c0b-8808-909939ff95db-utilities\") pod \"redhat-operators-ck72j\" (UID: \"960f0b07-5fd6-4c0b-8808-909939ff95db\") " pod="openshift-marketplace/redhat-operators-ck72j" Feb 17 14:16:48 crc kubenswrapper[4955]: I0217 14:16:48.105663 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/960f0b07-5fd6-4c0b-8808-909939ff95db-utilities\") pod \"redhat-operators-ck72j\" (UID: \"960f0b07-5fd6-4c0b-8808-909939ff95db\") " pod="openshift-marketplace/redhat-operators-ck72j" Feb 17 14:16:48 crc kubenswrapper[4955]: I0217 14:16:48.105821 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8wkm\" (UniqueName: \"kubernetes.io/projected/960f0b07-5fd6-4c0b-8808-909939ff95db-kube-api-access-h8wkm\") pod \"redhat-operators-ck72j\" (UID: \"960f0b07-5fd6-4c0b-8808-909939ff95db\") " pod="openshift-marketplace/redhat-operators-ck72j" Feb 17 14:16:48 crc kubenswrapper[4955]: I0217 14:16:48.105842 4955 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/960f0b07-5fd6-4c0b-8808-909939ff95db-catalog-content\") pod \"redhat-operators-ck72j\" (UID: \"960f0b07-5fd6-4c0b-8808-909939ff95db\") " pod="openshift-marketplace/redhat-operators-ck72j" Feb 17 14:16:48 crc kubenswrapper[4955]: I0217 14:16:48.106292 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/960f0b07-5fd6-4c0b-8808-909939ff95db-catalog-content\") pod \"redhat-operators-ck72j\" (UID: \"960f0b07-5fd6-4c0b-8808-909939ff95db\") " pod="openshift-marketplace/redhat-operators-ck72j" Feb 17 14:16:48 crc kubenswrapper[4955]: I0217 14:16:48.106291 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/960f0b07-5fd6-4c0b-8808-909939ff95db-utilities\") pod \"redhat-operators-ck72j\" (UID: \"960f0b07-5fd6-4c0b-8808-909939ff95db\") " pod="openshift-marketplace/redhat-operators-ck72j" Feb 17 14:16:48 crc kubenswrapper[4955]: I0217 14:16:48.126084 4955 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8wkm\" (UniqueName: \"kubernetes.io/projected/960f0b07-5fd6-4c0b-8808-909939ff95db-kube-api-access-h8wkm\") pod \"redhat-operators-ck72j\" (UID: \"960f0b07-5fd6-4c0b-8808-909939ff95db\") " pod="openshift-marketplace/redhat-operators-ck72j" Feb 17 14:16:48 crc kubenswrapper[4955]: I0217 14:16:48.254957 4955 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ck72j" Feb 17 14:16:48 crc kubenswrapper[4955]: I0217 14:16:48.795041 4955 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ck72j"] Feb 17 14:16:49 crc kubenswrapper[4955]: I0217 14:16:49.572190 4955 generic.go:334] "Generic (PLEG): container finished" podID="960f0b07-5fd6-4c0b-8808-909939ff95db" containerID="ef653b120804ee71403c9fd2db7307d834710dfe6208507b42b396d0e475f2c2" exitCode=0 Feb 17 14:16:49 crc kubenswrapper[4955]: I0217 14:16:49.572437 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ck72j" event={"ID":"960f0b07-5fd6-4c0b-8808-909939ff95db","Type":"ContainerDied","Data":"ef653b120804ee71403c9fd2db7307d834710dfe6208507b42b396d0e475f2c2"} Feb 17 14:16:49 crc kubenswrapper[4955]: I0217 14:16:49.572710 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ck72j" event={"ID":"960f0b07-5fd6-4c0b-8808-909939ff95db","Type":"ContainerStarted","Data":"a6ca50398debedf57e79bd90f4a2bec5f0ffebf99750eda74a2e4c9db7336d45"} Feb 17 14:16:51 crc kubenswrapper[4955]: I0217 14:16:51.594308 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ck72j" event={"ID":"960f0b07-5fd6-4c0b-8808-909939ff95db","Type":"ContainerStarted","Data":"d6c066a99306707cbb20cbf098ba447c9310a851d17a5a944e3189dd081a5a98"} Feb 17 14:16:55 crc kubenswrapper[4955]: I0217 14:16:55.225144 4955 scope.go:117] "RemoveContainer" containerID="0f74b5105f865cc62e93aedd90435f5aa7c0ed8ad937b4e2a170054c21f3de5a" Feb 17 14:16:55 crc kubenswrapper[4955]: E0217 14:16:55.225936 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 14:16:56 crc kubenswrapper[4955]: I0217 14:16:56.639281 4955 generic.go:334] "Generic (PLEG): container finished" podID="960f0b07-5fd6-4c0b-8808-909939ff95db" containerID="d6c066a99306707cbb20cbf098ba447c9310a851d17a5a944e3189dd081a5a98" exitCode=0 Feb 17 14:16:56 crc kubenswrapper[4955]: I0217 14:16:56.639427 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ck72j" event={"ID":"960f0b07-5fd6-4c0b-8808-909939ff95db","Type":"ContainerDied","Data":"d6c066a99306707cbb20cbf098ba447c9310a851d17a5a944e3189dd081a5a98"} Feb 17 14:16:57 crc kubenswrapper[4955]: I0217 14:16:57.649844 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ck72j" event={"ID":"960f0b07-5fd6-4c0b-8808-909939ff95db","Type":"ContainerStarted","Data":"8b7fe67d502ac7aa40da424c82fffb4a712ae13d313595b752169530e1cd75f8"} Feb 17 14:16:57 crc kubenswrapper[4955]: I0217 14:16:57.683999 4955 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ck72j" podStartSLOduration=3.125174633 podStartE2EDuration="10.683971788s" podCreationTimestamp="2026-02-17 14:16:47 +0000 UTC" firstStartedPulling="2026-02-17 14:16:49.574544808 +0000 UTC m=+4348.097274351" lastFinishedPulling="2026-02-17 14:16:57.133341963 +0000 UTC m=+4355.656071506" observedRunningTime="2026-02-17 14:16:57.672137389 +0000 UTC m=+4356.194866932" watchObservedRunningTime="2026-02-17 14:16:57.683971788 +0000 UTC m=+4356.206701331" Feb 17 14:16:58 crc kubenswrapper[4955]: I0217 14:16:58.261595 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ck72j" Feb 17 14:16:58 crc kubenswrapper[4955]: I0217 14:16:58.261678 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ck72j" Feb 17 14:16:59 crc kubenswrapper[4955]: I0217 14:16:59.314747 4955 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-ck72j" podUID="960f0b07-5fd6-4c0b-8808-909939ff95db" containerName="registry-server" probeResult="failure" output=< Feb 17 14:16:59 crc kubenswrapper[4955]: timeout: failed to connect service ":50051" within 1s Feb 17 14:16:59 crc kubenswrapper[4955]: > Feb 17 14:17:08 crc kubenswrapper[4955]: I0217 14:17:08.302142 4955 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ck72j" Feb 17 14:17:08 crc kubenswrapper[4955]: I0217 14:17:08.351390 4955 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ck72j" Feb 17 14:17:08 crc kubenswrapper[4955]: I0217 14:17:08.548352 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ck72j"] Feb 17 14:17:09 crc kubenswrapper[4955]: I0217 14:17:09.755961 4955 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ck72j" podUID="960f0b07-5fd6-4c0b-8808-909939ff95db" containerName="registry-server" containerID="cri-o://8b7fe67d502ac7aa40da424c82fffb4a712ae13d313595b752169530e1cd75f8" gracePeriod=2 Feb 17 14:17:10 crc kubenswrapper[4955]: I0217 14:17:10.223359 4955 scope.go:117] "RemoveContainer" containerID="0f74b5105f865cc62e93aedd90435f5aa7c0ed8ad937b4e2a170054c21f3de5a" Feb 17 14:17:10 crc kubenswrapper[4955]: E0217 14:17:10.223974 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 14:17:10 crc kubenswrapper[4955]: I0217 14:17:10.257928 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ck72j" Feb 17 14:17:10 crc kubenswrapper[4955]: I0217 14:17:10.351150 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8wkm\" (UniqueName: \"kubernetes.io/projected/960f0b07-5fd6-4c0b-8808-909939ff95db-kube-api-access-h8wkm\") pod \"960f0b07-5fd6-4c0b-8808-909939ff95db\" (UID: \"960f0b07-5fd6-4c0b-8808-909939ff95db\") " Feb 17 14:17:10 crc kubenswrapper[4955]: I0217 14:17:10.351266 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/960f0b07-5fd6-4c0b-8808-909939ff95db-utilities\") pod \"960f0b07-5fd6-4c0b-8808-909939ff95db\" (UID: \"960f0b07-5fd6-4c0b-8808-909939ff95db\") " Feb 17 14:17:10 crc kubenswrapper[4955]: I0217 14:17:10.351333 4955 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/960f0b07-5fd6-4c0b-8808-909939ff95db-catalog-content\") pod \"960f0b07-5fd6-4c0b-8808-909939ff95db\" (UID: \"960f0b07-5fd6-4c0b-8808-909939ff95db\") " Feb 17 14:17:10 crc kubenswrapper[4955]: I0217 14:17:10.353749 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/960f0b07-5fd6-4c0b-8808-909939ff95db-utilities" (OuterVolumeSpecName: "utilities") pod "960f0b07-5fd6-4c0b-8808-909939ff95db" (UID: "960f0b07-5fd6-4c0b-8808-909939ff95db"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 14:17:10 crc kubenswrapper[4955]: I0217 14:17:10.358692 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/960f0b07-5fd6-4c0b-8808-909939ff95db-kube-api-access-h8wkm" (OuterVolumeSpecName: "kube-api-access-h8wkm") pod "960f0b07-5fd6-4c0b-8808-909939ff95db" (UID: "960f0b07-5fd6-4c0b-8808-909939ff95db"). InnerVolumeSpecName "kube-api-access-h8wkm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 17 14:17:10 crc kubenswrapper[4955]: I0217 14:17:10.454306 4955 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8wkm\" (UniqueName: \"kubernetes.io/projected/960f0b07-5fd6-4c0b-8808-909939ff95db-kube-api-access-h8wkm\") on node \"crc\" DevicePath \"\"" Feb 17 14:17:10 crc kubenswrapper[4955]: I0217 14:17:10.454624 4955 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/960f0b07-5fd6-4c0b-8808-909939ff95db-utilities\") on node \"crc\" DevicePath \"\"" Feb 17 14:17:10 crc kubenswrapper[4955]: I0217 14:17:10.478612 4955 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/960f0b07-5fd6-4c0b-8808-909939ff95db-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "960f0b07-5fd6-4c0b-8808-909939ff95db" (UID: "960f0b07-5fd6-4c0b-8808-909939ff95db"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 17 14:17:10 crc kubenswrapper[4955]: I0217 14:17:10.557004 4955 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/960f0b07-5fd6-4c0b-8808-909939ff95db-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 17 14:17:10 crc kubenswrapper[4955]: I0217 14:17:10.767319 4955 generic.go:334] "Generic (PLEG): container finished" podID="960f0b07-5fd6-4c0b-8808-909939ff95db" containerID="8b7fe67d502ac7aa40da424c82fffb4a712ae13d313595b752169530e1cd75f8" exitCode=0 Feb 17 14:17:10 crc kubenswrapper[4955]: I0217 14:17:10.767375 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ck72j" event={"ID":"960f0b07-5fd6-4c0b-8808-909939ff95db","Type":"ContainerDied","Data":"8b7fe67d502ac7aa40da424c82fffb4a712ae13d313595b752169530e1cd75f8"} Feb 17 14:17:10 crc kubenswrapper[4955]: I0217 14:17:10.767412 4955 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ck72j" event={"ID":"960f0b07-5fd6-4c0b-8808-909939ff95db","Type":"ContainerDied","Data":"a6ca50398debedf57e79bd90f4a2bec5f0ffebf99750eda74a2e4c9db7336d45"} Feb 17 14:17:10 crc kubenswrapper[4955]: I0217 14:17:10.767438 4955 scope.go:117] "RemoveContainer" containerID="8b7fe67d502ac7aa40da424c82fffb4a712ae13d313595b752169530e1cd75f8" Feb 17 14:17:10 crc kubenswrapper[4955]: I0217 14:17:10.767433 4955 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ck72j" Feb 17 14:17:10 crc kubenswrapper[4955]: I0217 14:17:10.801311 4955 scope.go:117] "RemoveContainer" containerID="d6c066a99306707cbb20cbf098ba447c9310a851d17a5a944e3189dd081a5a98" Feb 17 14:17:10 crc kubenswrapper[4955]: I0217 14:17:10.811266 4955 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ck72j"] Feb 17 14:17:10 crc kubenswrapper[4955]: I0217 14:17:10.819251 4955 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ck72j"] Feb 17 14:17:10 crc kubenswrapper[4955]: I0217 14:17:10.824236 4955 scope.go:117] "RemoveContainer" containerID="ef653b120804ee71403c9fd2db7307d834710dfe6208507b42b396d0e475f2c2" Feb 17 14:17:10 crc kubenswrapper[4955]: I0217 14:17:10.862186 4955 scope.go:117] "RemoveContainer" containerID="8b7fe67d502ac7aa40da424c82fffb4a712ae13d313595b752169530e1cd75f8" Feb 17 14:17:10 crc kubenswrapper[4955]: E0217 14:17:10.862564 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b7fe67d502ac7aa40da424c82fffb4a712ae13d313595b752169530e1cd75f8\": container with ID starting with 8b7fe67d502ac7aa40da424c82fffb4a712ae13d313595b752169530e1cd75f8 not found: ID does not exist" containerID="8b7fe67d502ac7aa40da424c82fffb4a712ae13d313595b752169530e1cd75f8" Feb 17 14:17:10 crc kubenswrapper[4955]: I0217 14:17:10.862608 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b7fe67d502ac7aa40da424c82fffb4a712ae13d313595b752169530e1cd75f8"} err="failed to get container status \"8b7fe67d502ac7aa40da424c82fffb4a712ae13d313595b752169530e1cd75f8\": rpc error: code = NotFound desc = could not find container \"8b7fe67d502ac7aa40da424c82fffb4a712ae13d313595b752169530e1cd75f8\": container with ID starting with 8b7fe67d502ac7aa40da424c82fffb4a712ae13d313595b752169530e1cd75f8 not found: ID does not exist" Feb 17 14:17:10 crc kubenswrapper[4955]: I0217 14:17:10.862639 4955 scope.go:117] "RemoveContainer" containerID="d6c066a99306707cbb20cbf098ba447c9310a851d17a5a944e3189dd081a5a98" Feb 17 14:17:10 crc kubenswrapper[4955]: E0217 14:17:10.863011 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6c066a99306707cbb20cbf098ba447c9310a851d17a5a944e3189dd081a5a98\": container with ID starting with d6c066a99306707cbb20cbf098ba447c9310a851d17a5a944e3189dd081a5a98 not found: ID does not exist" containerID="d6c066a99306707cbb20cbf098ba447c9310a851d17a5a944e3189dd081a5a98" Feb 17 14:17:10 crc kubenswrapper[4955]: I0217 14:17:10.863045 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6c066a99306707cbb20cbf098ba447c9310a851d17a5a944e3189dd081a5a98"} err="failed to get container status \"d6c066a99306707cbb20cbf098ba447c9310a851d17a5a944e3189dd081a5a98\": rpc error: code = NotFound desc = could not find container \"d6c066a99306707cbb20cbf098ba447c9310a851d17a5a944e3189dd081a5a98\": container with ID starting with d6c066a99306707cbb20cbf098ba447c9310a851d17a5a944e3189dd081a5a98 not found: ID does not exist" Feb 17 14:17:10 crc kubenswrapper[4955]: I0217 14:17:10.863067 4955 scope.go:117] "RemoveContainer" containerID="ef653b120804ee71403c9fd2db7307d834710dfe6208507b42b396d0e475f2c2" Feb 17 14:17:10 crc kubenswrapper[4955]: E0217 14:17:10.863363 4955 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef653b120804ee71403c9fd2db7307d834710dfe6208507b42b396d0e475f2c2\": container with ID starting with ef653b120804ee71403c9fd2db7307d834710dfe6208507b42b396d0e475f2c2 not found: ID does not exist" containerID="ef653b120804ee71403c9fd2db7307d834710dfe6208507b42b396d0e475f2c2" Feb 17 14:17:10 crc kubenswrapper[4955]: I0217 14:17:10.863405 4955 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef653b120804ee71403c9fd2db7307d834710dfe6208507b42b396d0e475f2c2"} err="failed to get container status \"ef653b120804ee71403c9fd2db7307d834710dfe6208507b42b396d0e475f2c2\": rpc error: code = NotFound desc = could not find container \"ef653b120804ee71403c9fd2db7307d834710dfe6208507b42b396d0e475f2c2\": container with ID starting with ef653b120804ee71403c9fd2db7307d834710dfe6208507b42b396d0e475f2c2 not found: ID does not exist" Feb 17 14:17:12 crc kubenswrapper[4955]: I0217 14:17:12.235556 4955 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="960f0b07-5fd6-4c0b-8808-909939ff95db" path="/var/lib/kubelet/pods/960f0b07-5fd6-4c0b-8808-909939ff95db/volumes" Feb 17 14:17:21 crc kubenswrapper[4955]: I0217 14:17:21.223914 4955 scope.go:117] "RemoveContainer" containerID="0f74b5105f865cc62e93aedd90435f5aa7c0ed8ad937b4e2a170054c21f3de5a" Feb 17 14:17:21 crc kubenswrapper[4955]: E0217 14:17:21.225005 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 14:17:33 crc kubenswrapper[4955]: I0217 14:17:33.222719 4955 scope.go:117] "RemoveContainer" containerID="0f74b5105f865cc62e93aedd90435f5aa7c0ed8ad937b4e2a170054c21f3de5a" Feb 17 14:17:33 crc kubenswrapper[4955]: E0217 14:17:33.223548 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 14:17:44 crc kubenswrapper[4955]: I0217 14:17:44.223109 4955 scope.go:117] "RemoveContainer" containerID="0f74b5105f865cc62e93aedd90435f5aa7c0ed8ad937b4e2a170054c21f3de5a" Feb 17 14:17:44 crc kubenswrapper[4955]: E0217 14:17:44.223814 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" Feb 17 14:17:57 crc kubenswrapper[4955]: I0217 14:17:57.223606 4955 scope.go:117] "RemoveContainer" containerID="0f74b5105f865cc62e93aedd90435f5aa7c0ed8ad937b4e2a170054c21f3de5a" Feb 17 14:17:57 crc kubenswrapper[4955]: E0217 14:17:57.225251 4955 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-29qxq_openshift-machine-config-operator(5bf36887-2a2b-4193-abdb-a15beaf46300)\"" pod="openshift-machine-config-operator/machine-config-daemon-29qxq" podUID="5bf36887-2a2b-4193-abdb-a15beaf46300" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515145074237024455 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015145074240017364 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015145063213016505 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015145063214015456 5ustar corecore